ScholarWorksIndianapolis
  • Communities & Collections
  • Browse ScholarWorks
  • English
  • Català
  • Čeština
  • Deutsch
  • Español
  • Français
  • Gàidhlig
  • Italiano
  • Latviešu
  • Magyar
  • Nederlands
  • Polski
  • Português
  • Português do Brasil
  • Suomi
  • Svenska
  • Türkçe
  • Tiếng Việt
  • Қазақ
  • বাংলা
  • हिंदी
  • Ελληνικά
  • Yкраї́нська
  • Log In
    or
    New user? Click here to register.Have you forgotten your password?
  1. Home
  2. Browse by Subject

Browsing by Subject "Personalized nutrition"

Now showing 1 - 1 of 1
Results Per Page
Sort Options
  • Loading...
    Thumbnail Image
    Item
    Improving Personalized Meal Planning with Large Language Models: Identifying and Decomposing Compound Ingredients
    (MDPI, 2025-04-29) Kopitar, Leon; Bedrač, Leon; Strath, Larissa J.; Bian, Jiang; Stiglic, Gregor; Biostatistics and Health Data Science, Richard M. Fairbanks School of Public Health
    Background/objectives: Identifying and decomposing compound ingredients within meal plans presents meal customization and nutritional analysis challenges. It is essential for accurately identifying and replacing problematic ingredients linked to allergies or intolerances and helping nutritional evaluation. Methods: This study explored the effectiveness of three large language models (LLMs)-GPT-4o, Llama-3 (70B), and Mixtral (8x7B), in decomposing compound ingredients into basic ingredients within meal plans. GPT-4o was used to generate 15 structured meal plans, each containing compound ingredients. Each LLM then identified and decomposed these compound items into basic ingredients. The decomposed ingredients were matched to entries in a subset of the USDA FoodData Central repository using API-based search and mapping techniques. Nutritional values were retrieved and aggregated to evaluate accuracy of decomposition. Performance was assessed through manual review by nutritionists and quantified using accuracy and F1-score. Statistical significance was tested using paired t-tests or Wilcoxon signed-rank tests based on normality. Results: Results showed that large models-both Llama-3 (70B) and GPT-4o-outperformed Mixtral (8x7B), achieving average F1-scores of 0.894 (95% CI: 0.84-0.95) and 0.842 (95% CI: 0.79-0.89), respectively, compared to an F1-score of 0.690 (95% CI: 0.62-0.76) from Mixtral (8x7B). Conclusions: The open-source Llama-3 (70B) model achieved the best performance, outperforming the commercial GPT-4o model, showing its superior ability to consistently break down compound ingredients into precise quantities within meal plans and illustrating its potential to enhance meal customization and nutritional analysis. These findings underscore the potential role of advanced LLMs in precision nutrition and their application in promoting healthier dietary practices tailored to individual preferences and needs.
About IU Indianapolis ScholarWorks
  • Accessibility
  • Privacy Notice
  • Copyright © 2025 The Trustees of Indiana University