Generative AI-Based Virtual Assistant using Retrieval-Augmented Generation: An evaluation study for bachelor projects

arXiv cs.AI / 4/30/2026

💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The study explores how large language models can power virtual assistants while mitigating issues like hallucinations, missing information, and inaccurate context in specialized domains.
  • It presents a Retrieval-Augmented Generation (RAG)-based assistant tailored to Maastricht University students, aimed at helping them navigate project-specific regulations.
  • The proposed approach improves response accuracy and reliability by combining LLM generation with up-to-date, domain-specific retrieval.
  • The paper evaluates the system using a structured evaluation framework and real-life testing to show it meets students’ needs in an educational setting.
  • The authors position the results as contributing evidence to improve LLM systems for application-specific uses and outline directions for further research.

Abstract

Large Language Models have been increasingly employed in the creation of Virtual Assistants due to their ability to generate human-like text and handle complex inquiries. While these models hold great promise, challenges such as hallucinations, missing information, and the difficulty of providing accurate and context-specific responses persist, particularly when applied to highly specialized content domains. In this paper, we focus on addressing these challenges by developing a virtual assistant designed to support students at Maastricht University in navigating project-specific regulations. We propose a virtual assistant based on a Retrieval-Augmented Generation system that enhances the accuracy and reliability of responses by integrating up-to-date, domain-specific knowledge. Through a robust evaluation framework and real-life testing, we demonstrate that our virtual assistant can effectively meet the needs of students while addressing the inherent challenges of applying Large Language Models to a specialized educational context. This work contributes to the ongoing discourse on improving LLM-based systems for specific applications and highlights areas for further research.