Skip to content

Latest commit

 

History

History

EmbeddingModels

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 
 
 
 
 

Promo banner for

Dear learner,

We’re excited to introduce Embedding Models: From Architecture to Implementation, a short course built in collaboration with Vectara, and taught by Ofer Mendelevitch, Vectara’s Head of Developer Relations.

Many LLM apps use a single embedding model for both questions and answers. This leads to retrieval issues, such as getting responses similar to the question itself. A dual encoder architecture uses seperate embedding models for questions and for answers, allowing retrieval to be more meaningful and contextual.

This course goes into the details of the architecture and capabilities of embedding models, which are used in many AI applications to capture the meaning of words and sentences.

You will learn about the evolution of embedding models, from word to sentence embeddings, and build and train a simple dual encoder model. This hands-on approach will help you understand the technical concepts behind embedding models and how to use them effectively.

Enroll Today

Launch email GIFs (33)

In detail, you’ll:

  • Learn about word embedding, sentence embedding, and cross-encoder models; and how they can be used in RAG
  • Understand how transformer models, specifically BERT (Bi-directional Encoder Representations from Transformers), are trained and used in semantic search systems
  • Gain knowledge of the evolution of sentence embedding and understand how the dual encoder architecture was formed
  • Use a contrastive loss to train a dual encoder model, with one encoder trained for questions and another for the responses
  • Utilize separate encoders for question and answer in a RAG pipeline and see how it affects the retrieval compared to using a single encoder model.

By the end of this course, you will understand word, sentence, and cross-encoder embedding models, and how transformer-based models like BERT are trained and used in semantic search. You will also learn how to train dual encoder models with contrastive loss and evaluate their impact on retrieval in a RAG pipeline.

Details

  • Gain an in-depth understanding of the architecture behind embedding models; and learn how to train and use them.

  • Learn how to use different embedding models such as Word2Vec and BERT in various semantic search systems.

  • Learn how to build and train dual encoder models using contrastive loss, enhancing the accuracy of question-answer retrieval applications.

Lesson Video Code
Introduction video
Introduction to embedding models video
Contextualized token embeddings video code
Token vs. sentence embedding video code
Training a dual encoder video code
Using embeddings in RAG video code
Conclusion video
Appendix Tips and Help code