-
Install Ollama
-
Download the model - we are getting tinyllama which is a very compact model. And it should run on most machines.
ollama pull tinyllama:latest
-
Download model for creating embeddings. The all-minilm aims to train sentence embedding models on very large sentence level datasets using a self-supervised contrastive learning objective.
ollama pull all-minilm:latest
-
Start, and make sure ollama is available at http://localhost:11434/
-
Install the packages in requirements.ttx
-
Run indexing
OLLAMA_HOST='0.0.0.0' python index_content.py
-
Test searching
QUERY="what is five plus five" USE_EMBEDDINGS=yes OLLAMA_HOST='0.0.0.0' python search.py
-
Play around by changing the models and other parameters
-
for a web interface - run the flask app:
FLASK_HOST=0.0.0.0 FLASK_PORT=8080 python flask_server.py
- then go to http://localhost:8080 on your browser.