Skip to main content

Detecting AI Responses in Survey Data: NORC’s Next Leap for Data Quality

Innovation Brief
Closeup of two men data science specialists working at office together, analyzing big data on screen. Coders working on new project

Author

Brandon Sepulvado

Senior Research Methodologist 

December 2025

Our AI detector is the newest safeguard for high-quality, affordable data collection.

High-quality survey data are increasingly difficult to collect. Response rates are falling, and organizations that conduct surveys or rely on survey data must contend with falsification. Dealing with these data quality threats across the different modes (e.g., phone, in-person, web), geographies, and cultures surveys are administered can be daunting. They increase the effort and costs of data collection and undermine researchers’ ability to generate timely insights. Now, there is a new challenge: the proliferation of artificial intelligence (AI).

My colleague Joshua Lerner recently described the benefits that AI can bring to survey methodology, and NORC has shown how generative AI might improve the quality of survey interviews. However, AI also has potential downsides. For instance, large language models (LLMs) and AI assistants such as ChatGPT can be used to generate responses to open-ended questions on surveys.



When used to respond to surveys in place of people, AI can pose threats to the quality of survey data collected and any insights derived from the data by reducing—or even completely undermining—the extent to which survey responses represent opinions, beliefs, and behaviors of humans in the sample. Evidence suggests that this is not merely a theoretical possibility but an empirical reality (see Related Presentations below for NORC’s relevant research). How can methodologists and researchers respond to this potential threat to survey data quality?

NORC created an AI detector built for survey science.

To address the data quality threats that AI-generated responses pose to surveys, NORC has developed a tool that predicts whether responses to open-ended survey questions were generated with AI. Our tool is very effective, with more than 99 percent precision and recall when determining whether open-ended responses were written by humans or LLM-powered AI assistants.

Our AI detection tool performs so well because we designed it specifically for surveys, thus avoiding many of the pitfalls that more generalist, one-size-fits-all AI detectors are shown to have. We first created training data by asking AmeriSpeak® panelists and several of the latest LLMs to respond to the same survey questions. We varied the types of open-ended questions to include those that elicit free-form responses and others that tend to get structured lists (such as when asked to name the top issues facing the United States). 

Once we collected the responses from humans and LLMs, we used natural language processing (NLP) and machine learning to train a classifier that predicts whether a given response originated from an LLM or a person.

We evaluated the detector’s performance on surveys in different domains with different study populations and on questions that required very different types of writing styles, and our tool continued to exhibit impressive performance, especially when compared to other currently available AI detection tools, whose accuracy tended to be much worse, hovering around the 50 to 75 percent range. A key benefit of our tool is that it is based on a more general methodological framework that could be easily tailored for new surveys.

Our AI response detector further strengthens the rigorous quality control procedures that NORC puts in place for our data collection efforts. The output from this tool is one additional piece of information included in the holistic decision-making process that NORC methodologists and researchers make about whether survey responses should be retained for analysis.

This human-in-the-loop approach ensures that NORC staff can monitor the performance of the AI detector for any issues that might arise and that the tool is integrated into NORC’s extensive quality assurance processes tailored to survey research. (For example, research has shown that the use of AI detectors that are too general to be fit-for-purpose can actually systematically skew data quality.)

“Our AI response detector further strengthens the rigorous quality control procedures that NORC puts in place for our data collection efforts.”

Senior Research Methodologist, Methodology & Quantitative Social Sciences

“Our AI response detector further strengthens the rigorous quality control procedures that NORC puts in place for our data collection efforts.”

NORC’s AI detector offers data integrity protection for the next era of research.

Having access to reliable data and insights is of paramount importance to decision-makers and the public. With our AI-detection tool, NORC surveys have a quality control solution that can scale from small projects to the largest efforts. Even at the largest of scales, our tool and the quality control processes within which it is used are designed to protect the voices of genuine respondents and the integrity of decisions made with their input. NORC clients can rest assured that their survey data will be of the highest quality.

There is no shortage of challenges to survey data quality, and NORC has—for over 80 years—been at the forefront of developing innovative solutions. By using AI to help combat threats to survey data quality that AI itself has brought, NORC’s AI detection tool for survey responses is the next step in this history of innovation.


Related Presentations & Reports

You may have seen our tool showcased at several recent conferences. if you wish to explore any of the presentations documented below, email Brandon Sepulvado.

  • Sepulvado, Brandon, Lilian Huang, and Joshua Y. Lerner. “Detecting AI-Generated Survey Responses: Algorithm Development and Bias Mitigation.” Joint Statistical Meetings, August 2025.

  • Huang, Lilian, Brandon Sepulvado, Joshua Y. Lerner, and Lilian Huang. “LLMs Do Not Respond like Humans: Experiments in Model Fine Tuning.” American Association of Public Opinion Research annual meeting, May 2025.

  • Lerner, Joshua Y., Brandon Sepulvado, Lilian Huang, Sierra Arnold, Michelle M. Johns, Stefan Vogler, and Erin Fordyce. “From Social Media to Survey Data: Employing AI-Usage Detectors to Identify AI-Generated Responses in the HIRISE+ Survey.” American Association of Public Opinion Research annual meeting, May 2025.

  • Sepulvado, Brandon, Joshua Y. Lerner, Lilian Huang, Leah Christian, and Ipek Bilgen. “Large Language Models do not Respond like Survey Respondents.” Federal Committee on Statistical Methodology Research and Policy Conference, November 2024. 

  • Sepulvado, Brandon, Joshua Y. Lerner, Ipek Bilgen, Leah Christian, and Lilian Huang. “Enhancing Survey Research Quality with LLMs,” Generative AI and Sociology workshop, Yale University, April 2024.


Suggested Citation

Sepulvado, B. (2025, December 22). Detecting AI Responses in Survey Data: NORC’s Next Leap for Data Quality. [Web blog post]. NORC at the University of Chicago. Retrieved from www.norc.org.


Tags

Research Divisions

Departments, Centers & Programs



Explore NORC Research Science Projects

Analyzing Parent Narratives to Create Parent Gauge™

Helping Head Start build a tool to assess parent, family, and community engagement

Client:

National Head Start Association, Ford Foundation, Rainin Foundation, Region V Head Start Association

America in One Room

A “deliberative polling” experiment to bridge American partisanship

Client:

Stanford University