Publication Date

1-23-2026

Document Type

Article

Publication Title

Advances in Physiology Education

Volume

50

Issue

1

DOI

10.1152/advan.00252.2024

First Page

244

Last Page

248

Abstract

This study evaluates the effectiveness of large language models (LLMs), specifically Claude Sonnet 4.0 and ChatGPT 4.1, for analyzing formative feedback to support student-centered learning (SCL). In large courses, instructors often struggle to promptly review and synthesize student input. We used a low-stakes task: analyzing 63 anonymous student responses to a Muddiest Point prompt in a human physiology class after a lecture on respiratory physiology. Across 20 runs, both LLMs consistently identified "Ventilation and Lung Mechanics" as the most frequent area of confusion, aligning with human analysis. LLMs completed the task significantly faster than a human reviewer (average: 19.6 s/31.0 s vs. 32 min), with thematic reliability. This suggests LLMs can efficiently generate information from student input, enabling instructors to adapt their instruction in real time. The approach supports SCL and educational equity through the inclusion of all student voices. While promising for formative feedback, observed variability indicates that further refinement is needed before LLMs are utilized for high-stakes summative assessment.NEW & NOTEWORTHY We demonstrate the effectiveness of large language models in a low-stakes learning activity, providing the instructor with feedback on student learning based on theme analysis of free responses to a Muddiest Point prompt, thereby facilitating student-centered learning.

Keywords

formative assessment, large language models, Muddiest Point, qualitative analysis, student-centered learning

Creative Commons License

Creative Commons License
This work is licensed under a Creative Commons Attribution-Noncommercial 4.0 License

Department

Biological Sciences

Share

COinS