TLDR: RAG-PRISM is an adaptive tutoring framework that uses generative AI and Retrieval-Augmented Generation (RAG) with sentiment analysis to provide personalized, rapid, and immersive skill mastery. It addresses workforce skill gaps, especially in 4IR fields like cybersecurity, by tailoring learning content based on individual needs and emotional states. Evaluations showed high retrieval accuracy and strong performance from GPT-4 in generating faithful and relevant responses.
The Fourth Industrial Revolution (4IR) is rapidly changing the world of work, creating new demands for skills, particularly in areas like robotics, artificial intelligence (AI), and cybersecurity. This shift has led to significant skill gaps, especially for the experienced workforce. Traditional training methods often struggle to keep up, failing to offer the personalized, adaptive learning experiences needed to effectively re-skill and up-skill a diverse global workforce. This is where the RAG-PRISM framework steps in, offering a novel approach to tackle these challenges.
RAG-PRISM, which stands for a Personalized, Rapid, and Immersive Skill Mastery Framework with Adaptive Retrieval-Augmented Tutoring, is designed to provide highly personalized training. It leverages the power of generative AI combined with Retrieval-Augmented Generation (RAG) to create learning experiences tailored to each student’s unique needs and learning style. Imagine a tutor that not only understands what you need to learn but also how you feel about it, adjusting its approach in real-time – that’s the essence of RAG-PRISM.
The framework’s core innovation lies in its ability to adapt. It uses metrics like document hit rate and Mean Reciprocal Rank (MRR) to optimize training content. Crucially, it integrates sentiment analysis, meaning it can interpret a learner’s emotional state and engagement from their interactions. This “sentiment-aware” approach allows the system to maintain motivation and focus, much like a human tutor would. For instance, if a student seems frustrated, the system can adjust the content or pace to better suit their emotional and cognitive state.
RAG-PRISM builds upon the foundational PRISM framework by adding Retrieval-Augmented Generation (RAG). RAG is a powerful technique that enhances large language models (LLMs) by allowing them to retrieve relevant information from external knowledge bases before generating a response. This means the AI tutor doesn’t just rely on its pre-trained knowledge but can pull in the most current and specific information, ensuring accuracy and contextual relevance. This is particularly vital in fast-evolving fields like cybersecurity, where information quickly becomes outdated.
The system’s design is a hybrid pipeline that brings together student sentiment analysis, LLM-powered guidance, and knowledge retrieval. The student sentiment analysis module monitors a learner’s state within a VR-based Digital Twin environment, capturing dialogue content and task completion patterns. These observations are then processed by an LLM using zero-shot sentiment analysis to generate a “sentiment vector” – a numerical representation of the student’s emotional tone, confidence, and engagement. This vector then informs the personalized learning module.
The LLM-Powered Personalized Learning Module takes these sentiment-infused queries and processes them using LlamaIndex, a flexible data framework that connects LLMs with external knowledge bases. LlamaIndex simplifies the RAG process, handling everything from document ingestion and chunking to semantic search and prompt construction. The retrieved content is then embedded into a “knowledge vector,” providing highly relevant educational material personalized to the learner’s context and emotion.
The research applied this framework to 4IR cybersecurity learning, creating a synthetic question-answer dataset to emulate trainee behavior. The RAG components were optimized using a curated corpus of cybersecurity learning materials. The framework’s ability to generate new training content was evaluated by comparing its responses to manually curated queries, representing realistic student interactions. The responses were generated using various large language models, including GPT-3.5 and GPT-4 variants.
Evaluation of RAG-PRISM focused on two key areas: the retriever’s ability to find relevant content and the LLM’s ability to generate accurate and relevant responses. Metrics like Hit Rate and Mean Reciprocal Rank (MRR) were used for retrieval performance, while Faithfulness and Relevancy assessed the LLM’s output. Faithfulness ensures the answers are grounded in the retrieved context and not “hallucinated,” while Relevancy measures how well the response directly addresses the user’s query.
The results were promising. The framework achieved perfect document hit rate and MRR scores of 1.00 in its test case, demonstrating its effectiveness in retrieving relevant information. Among the LLMs tested, GPT-4 stood out with a perfect faithfulness score of 1.00 and a high relevancy score of 0.93. This indicates that GPT-4 consistently generated responses that were both accurate and well-aligned with the retrieved context, outperforming other models like GPT-3.5 variants and GPT-4 Turbo in terms of relevancy.
In conclusion, RAG-PRISM offers a significant step forward in adaptive tutoring systems. By integrating vector-based retrieval with LLMs and incorporating real-time sentiment analysis, it provides accurate, grounded, and contextually aligned feedback. This approach not only helps reduce misinformation but also enables truly adaptive, learner-centric instruction, bridging the gap between static digital platforms and intelligent, responsive learning environments. For more details, you can read the full research paper here.
Also Read:
- REFRAG: Boosting LLM Speed and Context for RAG Applications
- AnchorRAG: A Multi-Agent Framework for Enhanced Open-World Question Answering with Knowledge Graphs
Future work aims to expand the framework’s application to real-world educational settings, incorporating inquiries from actual students and measuring their learning outcomes. Researchers also plan to diversify content domains, explore alternative RAG implementations, and integrate human feedback mechanisms to further enhance the system’s generalizability and performance.


