As
AI becomes embedded across the research lifecycle from design and sampling to
analysis and insight generation, data quality risks are no longer limited to
poor methodology or execution. Increasingly, they arise from automation itself.
This session examines how shadow AI, AI-assisted workflows, and emerging
agent-like systems quietly undermine data quality by eroding provenance,
auditability, and human accountability.
Drawing on lessons from cybersecurity
and enterprise AI governance, the session reframes data quality as a
systems-level credibility challenge, not simply a technical or methodological
one. Participants will explore how common practices such as unapproved AI
tools, automated synthesis of open-ended responses, and systems that treat data
as implicit instruction introduce invisible quality failures that traditional
QA processes do not detect. These failures often surface only downstream, when
insights must be defended to stakeholders or acted on in high-stakes decisions.
The session focuses on practical, research-relevant governance patterns that
help organizations balance speed, scale, and rigor. Rather than advocating for
less automation, it offers guidance on where human judgment must remain
non-delegable and how research leaders can preserve transparency and trust in
AI-enabled insight pipelines.
Learning
Objectives / Actionable Insights
- Identify where traditional data quality controls fail in AI-driven
research workflows, particularly in the presence of shadow AI and automated
synthesis.
- Understand
when data stops acting as passive input and begins functioning as instruction
and why this matters for insight integrity.
- Apply
governance practices that enable responsible automation while maintaining
defensible, trustworthy research outcomes.
Presented By: Cecilia Dones, Columbia University