AI Navigate

Evaluation of LLMs in retrieving food and nutritional context for RAG systems

arXiv cs.CL / 3/11/2026

Ideas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The article evaluates four Large Language Models (LLMs) for their capability to retrieve food and nutrition data in Retrieval-Augmented Generation (RAG) systems using a detailed food composition database.
  • The evaluation focuses on the LLMs' ability to convert natural language queries into structured metadata filters to enable efficient data retrieval through a Chroma vector database.
  • Results show that LLMs can achieve high accuracy for easy and moderately complex queries, significantly reducing the need for manual effort and technical expertise by domain experts such as nutritionists.
  • However, the models face challenges with difficult queries that involve non-expressible constraints, indicating limitations when queries extend beyond the scope of available metadata representations.
  • The study highlights the potential for LLM-driven metadata filtering as a high-performance tool for domain-specific retrieval tasks while identifying key areas for improvement in handling complex, less explicitly defined queries.

Computer Science > Computation and Language

arXiv:2603.09704 (cs)
[Submitted on 10 Mar 2026]

Title:Evaluation of LLMs in retrieving food and nutritional context for RAG systems

View a PDF of the paper titled Evaluation of LLMs in retrieving food and nutritional context for RAG systems, by Maks Po\v{z}arnik Vavken and 3 other authors
View PDF HTML (experimental)
Abstract:In this article, we evaluate four Large Language Models (LLMs) and their effectiveness at retrieving data within a specialized Retrieval-Augmented Generation (RAG) system, using a comprehensive food composition database. Our method is focused on the LLMs ability to translate natural language queries into structured metadata filters, enabling efficient retrieval via a Chroma vector database. By achieving high accuracy in this critical retrieval step, we demonstrate that LLMs can serve as an accessible, high-performance tool, drastically reducing the manual effort and technical expertise previously required for domain experts, such as food compilers and nutritionists, to leverage complex food and nutrition data. However, despite the high performance on easy and moderately complex queries, our analysis of difficult questions reveals that reliable retrieval remains challenging when queries involve non-expressible constraints. These findings demonstrate that LLM-driven metadata filtering excels when constraints can be explicitly expressed, but struggles when queries exceed the representational scope of the metadata format.
Comments:
Subjects: Computation and Language (cs.CL)
Cite as: arXiv:2603.09704 [cs.CL]
  (or arXiv:2603.09704v1 [cs.CL] for this version)
  https://doi.org/10.48550/arXiv.2603.09704
Focus to learn more
arXiv-issued DOI via DataCite
Related DOI: https://doi.org/10.1109/BigData66926.2025.11401545
Focus to learn more
DOI(s) linking to related resources

Submission history

From: Matevž Ogrinc [view email]
[v1] Tue, 10 Mar 2026 14:15:35 UTC (173 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled Evaluation of LLMs in retrieving food and nutritional context for RAG systems, by Maks Po\v{z}arnik Vavken and 3 other authors
  • View PDF
  • HTML (experimental)
  • TeX Source
Current browse context:
cs.CL
< prev   |   next >
Change to browse by:
cs

References & Citations

export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo
Bibliographic Tools

Bibliographic and Citation Tools

Bibliographic Explorer Toggle
Bibliographic Explorer (What is the Explorer?)
Connected Papers Toggle
Connected Papers (What is Connected Papers?)
Litmaps Toggle
Litmaps (What is Litmaps?)
scite.ai Toggle
scite Smart Citations (What are Smart Citations?)
Code, Data, Media

Code, Data and Media Associated with this Article

alphaXiv Toggle
alphaXiv (What is alphaXiv?)
Links to Code Toggle
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub Toggle
DagsHub (What is DagsHub?)
GotitPub Toggle
Gotit.pub (What is GotitPub?)
Huggingface Toggle
Hugging Face (What is Huggingface?)
Links to Code Toggle
Papers with Code (What is Papers with Code?)
ScienceCast Toggle
ScienceCast (What is ScienceCast?)
Demos

Demos

Replicate Toggle
Replicate (What is Replicate?)
Spaces Toggle
Hugging Face Spaces (What is Spaces?)
Spaces Toggle
TXYZ.AI (What is TXYZ.AI?)
Related Papers

Recommenders and Search Tools

Link to Influence Flower
Influence Flower (What are Influence Flowers?)
Core recommender toggle
CORE Recommender (What is CORE?)
About arXivLabs

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.