LLM – Retrieval Augmented Generation (RAG) with Ollama Embeddings
The work on the Large Language Model (LLM) bot so far has seen the running of LLM locally using Ollama, a switch in models (from tinyllama to gemma) whilst introducing LangChain and then the switch to LangChain templates. Note: If you skipped the previous blog entry posts, I’m following along with Real Pythons “Build an LLM […]