Treffer: Leveraging AI-Generated Synthetic Data to Train Natural Language Processing Models for Qualitative Feedback Analysis

Title:
Leveraging AI-Generated Synthetic Data to Train Natural Language Processing Models for Qualitative Feedback Analysis
Language:
English
Authors:
Stephanie Fuchs (ORCID 0000-0001-9264-3094), Alexandra Werth (ORCID 0000-0003-0310-2654), Cristóbal Méndez (ORCID 0000-0002-1257-6707), Jonathan Butcher (ORCID 0000-0002-9309-6296)
Source:
Journal of Engineering Education. 2025 114(4).
Availability:
Wiley. Available from: John Wiley & Sons, Inc. 111 River Street, Hoboken, NJ 07030. Tel: 800-835-6770; e-mail: cs-journals@wiley.com; Web site: https://www.wiley.com/en-us
Peer Reviewed:
Y
Page Count:
30
Publication Date:
2025
Contract Number:
EF2222434
Document Type:
Fachzeitschrift Journal Articles<br />Reports - Research
DOI:
10.1002/jee.70033
ISSN:
1069-4730
2168-9830
Entry Date:
2025
Accession Number:
EJ1487745
Database:
ERIC

Weitere Informationen

Background: High-quality feedback is crucial for academic success, driving student motivation and engagement while research explores effective delivery and student interactions. Advances in artificial intelligence (AI), particularly natural language processing (NLP), offer innovative methods for analyzing complex qualitative data such as feedback interactions. Purpose: We developed a framework to train sentence transformers using generative AI--created synthetic data to categorize student-feedback interactions in engineering studios. We compared traditional thematic analysis with modern methods to evaluate the realism of synthetic datasets and their effectiveness in training NLP models by exploring how generative AI can aid qualitative coding. Methods: We deidentified and transcribed eight audio recordings from engineering studios. Synthetic feedback transcripts were generated using three locally hosted large language models: Llama 3.1, Gemma 2.0, and Mistral NeMo, adjusting parameters to produce datasets mimicking the real transcripts. We assessed the quality of synthetic transcripts using our framework and used a sentence transformer model (trained on both real and synthetic data) to compare changes in the model's percent accuracy when qualitatively coding feedback interactions. Results: Synthetic data improved the NLP model's performance in classifying feedback interactions, boosting the average accuracy from 68.4% to 81% with Llama 3.1. Although incorporating synthetic data improved classification, all models produced transcripts that occasionally included extraneous details and failed to capture instructor-dominant discourse. Conclusions: Synthetic data offers an opportunity to expand qualitative research, particularly in contexts where real data for NLP training is limited or hard to obtain; however, transparency in its use is paramount to maintain research integrity.

As Provided