使用Player FM应用程序离线!
Use Long Context or RAG?
Manage episode 460390358 series 3496315
In this one, I discuss the dilemma between using retrieval-based generation and the newer "long context models".
Long context models, like the Gemini suite of models, allow us to send up to millions of tokens (thousands of text pages), whereas retrieval (RAG)-based systems allow us to search through as much (if not more) content and retrieve only the necessary bits to send the LLM for improved answers.
Both have advantages and disadvantages. This short episode will help you better understand when to use each.
Build Your First Scalable Product with LLMs: https://academy.towardsai.net/courses/beginner-to-advanced-llm-dev?ref=1f9b29
Master LLMs and Get Industry-ready Now: https://academy.towardsai.net/?ref=1f9b29
Our ebook: https://academy.towardsai.net/courses/buildingllmsforproduction?ref=1f9b29
36集单集
Manage episode 460390358 series 3496315
In this one, I discuss the dilemma between using retrieval-based generation and the newer "long context models".
Long context models, like the Gemini suite of models, allow us to send up to millions of tokens (thousands of text pages), whereas retrieval (RAG)-based systems allow us to search through as much (if not more) content and retrieve only the necessary bits to send the LLM for improved answers.
Both have advantages and disadvantages. This short episode will help you better understand when to use each.
Build Your First Scalable Product with LLMs: https://academy.towardsai.net/courses/beginner-to-advanced-llm-dev?ref=1f9b29
Master LLMs and Get Industry-ready Now: https://academy.towardsai.net/?ref=1f9b29
Our ebook: https://academy.towardsai.net/courses/buildingllmsforproduction?ref=1f9b29
36集单集
Все серии
×欢迎使用Player FM
Player FM正在网上搜索高质量的播客,以便您现在享受。它是最好的播客应用程序,适用于安卓、iPhone和网络。注册以跨设备同步订阅。