A simple Retrieval-Augmented Generation (RAG) Chatbot built using LangChain, Ollama, ChromaDB, and Streamlit.
This chatbot answers user queries based on document context using vector search + LLM reasoning.
- 🔍 Semantic search using vector embeddings
- 🧠 Context-aware answers using RAG pipeline
- 💬 Conversational interface with chat history
- ⚡ Local LLM using Ollama (no API cost)
- 📊 Fast retrieval using Chroma vector database
- 🎯 Accurate answers grounded in provided documents
- Backend Framework: LangChain
- LLM: Ollama (LLaMA 3.2)
- Embeddings: Ollama Embeddings (
nomic-embed-text) - Vector Database: ChromaDB
- Frontend: Streamlit
- Language: Python
- User enters a question
- Query is converted into embeddings
- ChromaDB retrieves relevant document chunks
- Context is combined with the question
- LLM (Ollama) generates final answer
git clone https://github.com/Sunanda01/Basic_RAG_Chatbot.git
cd Basic_RAG_Chatbotpython -m venv venv
source venv/bin/activate # Mac/Linux
venv\Scripts\activate # Windowspip install -r requirements.txtRun required models:
ollama pull llama3.2
ollama pull nomic-embed-text# Document Injestion
python injest.py
# Run UI
streamlit run home.pyOpen browser at http://localhost:8501
Ask questions based on your documents
Chatbot will respond using retrieved context