Published On Sep 26, 2024
LLaMA3.2 has released a new set of compact models designed for on-device use cases, such as locally running assistants. Here, we show how LangGraph can enable these types of local assistant by building a multi-step RAG agent - this combines ideas from 3 advanced RAG papers (Adaptive RAG, Corrective RAG, and Self-RAG) into a single control flow using LangGraph. But we show LangGraph makes it possible to run a complex agent locally.
Code:
https://langchain-ai.github.io/langgr...
Llama3.2:
https://huggingface.co/blog/llama32#w...
Full course on LangGraph:
https://academy.langchain.com/courses...
show more