π This video introduces the use of embeddings and ChromaDB for the multi doc retriever.
βοΈ Having a GPU is recommended for faster processing, but it can also be run on a CPU.
π The video demonstrates how to work with multiple PDF files instead of text files.
π There are two ways of doing embeddings: using Hugging Face embeddings or using instructor embeddings.
π The instructor embeddings are custom embeddings that can be used for specific purposes.
π» LangChain is used to locally run the embeddings.
π₯ The model and necessary files are downloaded for usage.
π» The embeddings are set up for vector storage.
π ChromaDB is used to set up the vector store.
π The video introduces the use of instructor embeddings in a retriever for LangChain retrieval QA.
π The retriever utilizes the instructor embeddings to find contexts that match a given query.
π The top documents selected by the embeddings in the retriever provide relevant information for specific queries.
π‘ LangChain Retrieval QA is able to find answers from the same paper and can provide information about ToolFormer.
π By asking questions about ToolFormer, we can learn about its functionalities and the tools that can be used with it.
π LangChain Retrieval QA is useful for extracting specific information from papers and can even provide insights from related survey papers.
π Using embedding system for instructing better without relying on OpenAI for language models.
π The system is able to retrieve information and answer specific questions about retrieval augmentation and differences between REALM and RAG models.
π More privacy in data processing by not sending all data to the large language model for embeddings.
π‘ Using an actual language model for replying and embedding.
π» Deleting and bringing back the ChromaDB database.
π§ Exploring custom models for various tasks.