Fine Tuning Embedding Models
This workshop explores the importance of fine-tuning Language and Embedding Models (LLMs)
Fine Tuning Embedding Models
This workshop explores the importance of fine-tuning Language and Embedding Models (LLMs). It highlights how embedding models are used to map natural language to vectors, crucial for pipelines with multiple models to adapt to specific data nuances. An example demonstrates fine-tuning an embedding model for legal text. The notebook discusses existing solutions and hardware considerations, emphasizing GPU usage for large data.
The practical part of the notebook shows the fine-tuning process of the "distilroberta-base" model from the SentenceTransformer library. It utilizes the QQP_triplets dataset from Quora for training, designed around semantic meaning. The notebook prepares the data, sets up a DataLoader, and employs Triplet Loss to encourage the model to map similar data points closely while distancing dissimilar ones. It concludes by mentioning the training duration and resources needed for further improvements.
Tutorial Topics:
This course consists of an on-demand recording, course notebook, and course exercises
Before accessing the course code notebooks it is advisable to review the course prerequisites here.
Mary Grace Moesta
Welcome to the tutorial!
What You'll Learn in This Tutorial
How to use this tutorial
Tutorial Prerequisites
Fine Tuning Part I: Embedding Models
Lesson Notebook: Fine Tuning Embedding Models
Fine Tuning Quiz
A Hands-on Tutorial