| url | https://doi.org/10.1177/10497323251389800 |
|---|---|
| raw | raw/dellafiore-et-al-2025-artificial-intelligence-in-qualitative-research-insights-from-experts-via-reflexive-thematic.pdf |
TL;DR: Fourteen expert qualitative researchers from Italian academic and healthcare settings were interviewed about AI. The study surfaces a culture of concealment around AI use, introduces the “illusion of meaning” as a practitioner concept, and shows that the human–AI divide maps cleanly onto a technical tasks / interpretive tasks split that experts negotiate with ambivalence rather than resolve.
Problem
Most research on AI in qualitative research either tests AI tools against human outputs or surveys researchers’ perceptions quantitatively. What is missing is an in-depth, qualitative account of how expert researchers — people with ten or more years of qualitative practice, doctoral training, and active publication records — actually experience, interpret, and negotiate AI integration within their methodological commitments. This study fills that gap using the most appropriate tool for the job: semi-structured interviews and reflexive thematic analysis applied to the phenomenon itself.
The setting matters. Italian socio-anthropological and healthcare contexts are underexplored in a literature dominated by Anglophone institutions. Participants here include anthropologists, sociologists, physicians, psychologists, and nurses — a disciplinary range that captures real methodological diversity. The healthcare-qualitative crossover context also surfaces concerns about environmental sustainability and data governance that appear more prominently than in methodological debates within social science alone.
Approach
Semi-structured interviews with 14 expert qualitative researchers, conducted April–May 2025 (face-to-face or Microsoft Teams). Purposive sampling targeted researchers with at least two to three completed qualitative projects, peer-reviewed publication records, and active qualitative responsibilities. Mean age 46 years; 7 men, 7 women; near-universal PhD level; all with 10+ years of research experience. Nine specialized in socio-anthropological fields; five in medicine and healthcare.
Notably, “Artificial Intelligence” was deliberately left undefined at the start of interviews — a methodological choice designed to elicit participants’ own conceptualizations rather than constrain responses. The opening question was simply: “What is being said about AI in your qualitative research environment?” Real-time reflective probing clarified what participants meant without imposing definitions.
Data were analyzed using Braun and Clarke’s six-step RTA framework, following RTARGs (Reflexive Thematic Analysis Reporting Guidelines). Two researchers coded independently; discrepancies resolved through discussion; senior researchers consulted for coherence. Member checking with three participants confirmed the thematic structure required no significant amendment.
Four themes emerged:
- Perspectives on AI and Qualitative Research — ambivalent, polarized attitudes including technophobia, prejudice, and openness; a pervasive shame around declaring AI use; skill-related barriers; and a pragmatic call for control
- Anthropological and Philosophical Dimension — the human–machine distinction (capacities vs. skills); AI as “alter ego” or “additional voice”; risks to researcher identity; the “illusion of meaning”
- Pragmatic Dimension of AI Usage — AI as support for technical tasks (transcription, translation, writing); contested role in interpretive phases; limits on fieldwork and ethnographic research
- Ethical and Sustainability Dimension — Western epistemological dominance, data ownership and governance gaps, ecological costs of AI infrastructure
The “Illusion of Meaning”
The most conceptually distinctive contribution is a term that emerged from participants, not from the researchers’ framework: the “illusion of meaning.” One participant articulated it precisely: “AI could, in some way, create the illusion of certain meaning aggregates — it might reconstruct meaning in an interview from the interviewee’s perspective. And that would be something new. I call it an ‘illusion of meaning’ because, clearly, these are algorithmic inferences and combinations.”
This is a practitioner-level articulation of what epistemic-flattening describes at the architectural level. The concern is not that AI is obviously wrong, but that it is subtly misleading — producing outputs that look interpretively meaningful but lack the interpretive grounding that comes from researcher immersion. The participant who coined the term described it with “a firm tone of voice and intense eye contact” — field note evidence of the depth of concern. The illusion is precisely that: it is not obviously wrong, which makes it harder to detect and resist than a hallucination would be.
The Shame Culture Around AI Use
A striking empirical finding is the culture of concealment around AI use. Seven of fourteen participants acknowledged widespread prejudice against AI in their professional circles. Several admitted to a “fear of being judged” that led them to conceal AI use — “in the darkness of our little rooms.” More revealingly: 13 of 14 participants admitted using AI in some form, but several had initially presented themselves as non-users, only clarifying during the interview that they were already employing AI-based tools. One realized mid-interview that NVivo (now Lumivero) itself uses AI.
This has methodological implications: research surveys asking about AI use may systematically underreport it. The Paulus & Marone (paulus-marone-qdas-discourse-2024) finding — that researchers may be absorbing AI epistemologies through platforms without recognizing it — gains empirical texture here. Researchers who don’t know they’re using AI cannot critically evaluate what it’s doing to their analysis.
Technical Tasks vs. Interpretive Tasks
The clearest consensus was a functional division of AI labor:
- Near-universal acceptance: transcription (all 14), scientific English writing and Italian-to-English translation (13/14), document summarization, visual abstracts
- Contested: study design, literature review (hallucination and source fabrication concerns), coding and theme generation
- Strong resistance: interview analysis described as a “black box” by several; ethnographic fieldwork explicitly excluded (“AI can’t do that on its own, because it doesn’t have a body”)
This maps onto the small-q / Big-Q distinction (nicmanis-spurrier-ai-guide-2025) but arrives there through practitioner experience rather than methodological argument. The phrase “it doesn’t have a body” — offered with a smile — captures the phenomenological commitment to fieldwork immersion that computational tools cannot replicate.
The generational divide identified by chatzichristos-ai-positivism-2025 reappears: younger researchers embraced AI more readily, while senior researchers expressed concern about their capacity to guide younger colleagues in developing critical judgment.
Epistemological Stance
Interpretivist and constructionist — consistent with the Braun & Clarke approach the team applies. The paper does not use intercoder reliability metrics, following RTA’s epistemological commitments. Researcher positionality is acknowledged. AI appears as an object of study, not a tool used in the analysis.
The ethical theme surfaces a concern with Western epistemological dominance that connects to structural arguments in the corpus: AI’s training data reflects “prevailing epistemologies, which are predominantly Western,” risking the marginalization of “minority viewpoints from subcultures or underrepresented groups.” This connects directly to epistemic-flattening, dahal-genai-qualitative-nepal-2024, and sakaguchi-chatgpt-japanese-2025, but is articulated here from within a practitioner community — not as an external critique but as a live ethical concern.
Rigor and Trustworthiness
The study follows RTARGs, combines dual independent coding with senior researcher oversight, integrates field notes systematically, conducts member checking, obtains ethical approval (Regional Ethics Committee of Umbria), and complies with GDPR. Consistent with RTA epistemology, quantitative reliability metrics are not reported — explicitly and appropriately.
The limitation acknowledged most honestly: this is a study of perceptions, not practices. The researchers cannot observe what participants actually do with AI; they can only analyze what participants report and how they describe it. The open definition of “AI” was a productive methodological choice, but it means participants may have been describing different tools under the same label.
Limitations
- Geographic and disciplinary scope: All participants from Central and Northern Italy, predominantly healthcare-adjacent disciplines. Transferability to other national or disciplinary contexts is limited.
- Perception–practice gap: Self-reported use is not observed use. The shame culture the paper itself identifies means participants may have underreported AI use even in this relatively safe research context.
- Undefined “AI”: The deliberate non-definition enabled rich conceptual exploration but limits comparability across participants (ChatGPT and NVivo/Lumivero are not the same kind of tool).
- Rapid obsolescence: Perceptions captured April–May 2025 may already be outdated; the paper notes this explicitly.
- No comparative data: No benchmarks against human-AI output differences; this is entirely the qualitative perception literature, not the empirical performance literature.
Connections
- “Illusion of meaning” connects structurally to: epistemic-flattening — the concept is the practitioner-level version of what Brailas describes architecturally; AI outputs that are statistically probable but interpretively hollow
- Shame culture and under-reporting connects to: paulus-marone-qdas-discourse-2024 — the same phenomenon: researchers absorbing and using AI while publicly distancing from it
- Generational gap confirms: chatzichristos-ai-positivism-2025 — younger researchers more receptive, senior researchers more concerned; finding replicated in an Italian healthcare context
- “Black box” / transparency concern connects to: wise-et-al-2026-ai-not-the-enemy — the transparency in analytic decision-making criterion is precisely the governance response to the opacity participants describe
- Technical/interpretive task split maps to: nicmanis-spurrier-ai-guide-2025 (small-q/Big-Q), prescott-ai-thematic-analysis-2024 (speed benefits for bounded tasks)
- Western epistemological dominance connects to: dahal-genai-qualitative-nepal-2024, sakaguchi-chatgpt-japanese-2025 — structural bias in training data; community corroboration
- Environmental concerns connect to: jowsey-et-al-2025-we-reject (environmental justice as pillar of rejection), greenhalgh-2026-beyond-the-binary (governance-not-refusal as response)
- AI as “additional voice” / alter ego connects to: human-ai-collaboration — the alter ego framing is an informal version of the structured frameworks (CAAI, GAITA, AbductivAI) in which AI is a dialogic partner
- Governance gap connects to: ai-research-ethics — participants articulate the same vacuum Davison et al. (davison-ethics-genai-2024) document; no universal ethical code; decisions left to personal moral reflection
- Interprofessional teams proposal connects to: zhang-ai-qualitative-research-2025 — collaborative, expertise-pooling models as practical response
What links here
- AI Research Ethics
- Andrews, Fainshmidt & Gaur (2026) — Progress or Perish: IB and AI Adoption
- Chatzichristos (2025) — Qualitative Research in the Era of AI: A Return to Positivism or a New Paradigm?
- Contested Claims
- Empirical Findings
- Epistemic Flattening
- Epistemology — Stances Across the Literature
- AI in Qualitative Research
- Human-AI Collaboration — Frameworks and Models
- Index
- LLMs for Qualitative Research
- Paulus & Marone (2024) — "In Minutes Instead of Weeks": Discursive Constructions of Generative AI and Qualitative Data Analysis
- Qualitative AI Methods — A Living Taxonomy
- Validity and Trustworthiness