TLDR: This research paper investigates an LLM-powered course assistant deployed to 2,000 computer science students. It finds the system effectively addresses temporal support gaps and novice learner needs, with high usage in evenings and for homework. However, challenges include LLM errors, low student engagement with AI-generated follow-up questions, and limitations in promoting higher-order cognitive thinking. The study emphasizes the need for improved system design, enhanced user agency, and greater educator involvement to optimize human-LLM educational interactions.
Universities often struggle to provide students with timely and flexible academic support, especially outside of regular hours. This gap can leave many students without the help they need. Large Language Models (LLMs) offer a promising solution to bridge this divide, but how students interact with these AI assistants, and their educational impact, has been a subject of ongoing study.
A recent research paper, “Investigating Student Interaction Patterns with Large Language Model-Powered Course Assistants in Computer Science Courses,” explores the real-world use and pedagogical implications of an LLM-powered course assistant. Developed by researchers including Chang Liu, Loc Hoang, Andrew Stolman, Rene F. Kizilcec, and Bo Wu, this system was deployed across multiple computer science courses to understand how students engage with it.
The LLM Course Assistant: Design and Deployment
The course assistant is built on Retrieval-Augmented Generation (RAG) technology, which allows it to retrieve relevant information from an educator-curated database to generate accurate and context-specific responses. It includes several key features designed to provide a structured learning experience. These include a multi-route question dispatching system that processes user queries in different modes: a general mode for broad questions, a homework mode that provides hints rather than direct answers, and a practice mode for generating exercises. The system also offers course-level customization, allowing educators to define rules and instructional styles tailored to specific courses, such as adapting responses for introductory students with limited prior knowledge.
By Spring 2024, the system had been deployed to approximately 2,000 students across six courses at three different institutions. The analysis focused on interaction data from three undergraduate Computer Science courses at the Colorado School of Mines: Computer Science for STEM (introductory), Computer Organization (intermediate), and Operating Systems (advanced). This broad deployment allowed for a comprehensive study of how students at different academic levels utilize the AI assistant.
Key Findings on Student Interaction
The study revealed several important interaction patterns. Usage of the course assistant remained strong in the evenings and nights, indicating that it effectively addresses temporal support gaps when traditional help might not be available. Introductory courses showed higher usage, suggesting the system is particularly helpful for novice learners. Most conversations were relatively short, with 64.58% concluding within 10 minutes and 85.88% within three dialogue rounds. Interestingly, about 20.94% of sessions ended without any questions being posed, suggesting students might be exploring or struggling to formulate questions.
When it came to specific modes, homework mode accounted for 53.79% of conversations, significantly surpassing the general mode. This highlights a strong student reliance on the system for homework-related queries, especially in the Computer Science for STEM course. Weekly usage patterns showed an increase as the semester progressed, peaking before midterms and final exams, further emphasizing its role as a supplementary study tool.
Challenges and Opportunities for Improvement
Despite its benefits, the study also identified several challenges. LLMs, while powerful, are not infallible. The analysis found that 2.17% of responses contained erroneous content, including contextual and computational errors. Another 5.5% of responses were deemed unhelpful or irrelevant. These inaccuracies can be particularly problematic for students who may lack the expertise to identify incorrect information, potentially leading to misinformation and frustration.
Furthermore, the study examined the depth of student engagement. While the system attempts to foster inquiry-based learning by posing follow-up questions, approximately 65.27% of these LLM-generated questions were ignored by students, particularly in higher-level courses. This suggests that relying solely on student motivation for deeper engagement might not be sufficient. The cognitive level of LLM-generated questions was also found to be predominantly at the “Apply” level, often not inspiring higher-order cognitive thinking, which is crucial for deeper learning.
Linguistic analysis of student questions revealed that 15.5% contained grammatical errors, and only about 4.5% were polite. Students in lower-level courses were more prone to copying and pasting content without self-organized questions, potentially indicating an over-reliance or difficulty in articulating their queries. This contrasts with human teaching assistants who can interpret multimodal cues and proactively guide students.
Also Read:
- Navigating the AI Partnership: Understanding and Preventing Overreliance on Large Language Models
- Generative AI Streamlines Classification of Tutor Dialogue in Educational Settings
Moving Forward: Enhancing Human-LLM Educational Interactions
The research paper concludes with discussions on improving the system, user agency, and educator engagement. It suggests that while LLMs can mimic human-like intelligence, their fundamental differences from human educators mean that users interact with a distinct teaching agent without fully recognizing these distinctions. Future improvements could involve designing interaction systems that bridge this gap, perhaps by providing more explicit guidance on how to effectively interact with LLMs or by implementing backend features that process raw user input more intelligently.
Enhancing user agency and initiative is also crucial. While the system allows students to select learning scenarios, the high rate of ignored follow-up questions indicates a need for strategies to encourage deeper engagement. The paper also stresses the importance of greater educator involvement. Currently, educators primarily act as supervisors, but their pedagogical expertise could be better leveraged to guide LLM-powered systems in implementing appropriate educational theories and generating more meaningful responses. This could involve educators contributing more course materials, defining common questions, or setting conditions for interactive dialogues.
This study provides valuable insights into the practical application of LLM-powered course assistants in computer science education. It highlights their potential to fill support gaps and enhance learning experiences, while also underscoring the critical need for careful design, continuous evaluation, and thoughtful integration of human pedagogical expertise to overcome current limitations. For more details, you can read the full research paper here.


