Artwork

内容由Zeta Alpha提供。所有播客内容(包括剧集、图形和播客描述)均由 Zeta Alpha 或其播客平台合作伙伴直接上传和提供。如果您认为有人在未经您许可的情况下使用您的受版权保护的作品,您可以按照此处概述的流程进行操作https://zh.player.fm/legal
Player FM -播客应用
使用Player FM应用程序离线!

Zeta-Alpha-E5-Mistral: Finetuning LLMs for Retrieval (with Arthur Câmara)

19:35
 
分享
 

Manage episode 450164769 series 3446693
内容由Zeta Alpha提供。所有播客内容(包括剧集、图形和播客描述)均由 Zeta Alpha 或其播客平台合作伙伴直接上传和提供。如果您认为有人在未经您许可的情况下使用您的受版权保护的作品,您可以按照此处概述的流程进行操作https://zh.player.fm/legal

In the 30th episode of Neural Search Talks, we have our very own Arthur Câmara, Senior Research Engineer at Zeta Alpha, presenting a 20-minute guide on how we fine-tune Large Language Models for effective text retrieval. Arthur discusses the common issues with embedding models in a general-purpose RAG pipeline, how to tackle the lack of retrieval-oriented data for fine-tuning with InPars, and how we adapted E5-Mistral to rank in the top 10 on the BEIR benchmark.
## Sources

InPars

Zeta-Alpha-E5-Mistral

NanoBEIR

  continue reading

21集单集

Artwork
icon分享
 
Manage episode 450164769 series 3446693
内容由Zeta Alpha提供。所有播客内容(包括剧集、图形和播客描述)均由 Zeta Alpha 或其播客平台合作伙伴直接上传和提供。如果您认为有人在未经您许可的情况下使用您的受版权保护的作品,您可以按照此处概述的流程进行操作https://zh.player.fm/legal

In the 30th episode of Neural Search Talks, we have our very own Arthur Câmara, Senior Research Engineer at Zeta Alpha, presenting a 20-minute guide on how we fine-tune Large Language Models for effective text retrieval. Arthur discusses the common issues with embedding models in a general-purpose RAG pipeline, how to tackle the lack of retrieval-oriented data for fine-tuning with InPars, and how we adapted E5-Mistral to rank in the top 10 on the BEIR benchmark.
## Sources

InPars

Zeta-Alpha-E5-Mistral

NanoBEIR

  continue reading

21集单集

所有剧集

×
 
Loading …

欢迎使用Player FM

Player FM正在网上搜索高质量的播客,以便您现在享受。它是最好的播客应用程序,适用于安卓、iPhone和网络。注册以跨设备同步订阅。

 

快速参考指南

边探索边听这个节目
播放