Retrieval Augmented Generation (RAG) combines the power of LLMs with external knowledge base.
- Install Ollama and Mistral
curl -fsSL https://ollama.com/install.sh | sh
ollama run mistral
- Run Ollama
make llm
- Start Milvus (Docker should be running)
make start-milvus
- Create a virtual env and install requirements
conda create -n llm python==3.9.16
conda activate llm
pip install -r "requirements.txt"
- Run app
make streamlit-new # For first time
make streamlit # After vector db has been created