-
Install Ollama
-
Download the model - we are getting tinyllama which is a very compact model. And it should run on most machines.
ollama pull tinyllama:latest
-
Download model for creating embeddings. The all-minilm aims to train sentence embedding models on very large sentence level datasets using a self-supervised contrastive learning objective.
ollama pull all-minilm:latest
-
Start, and make sure ollama is available at http://localhost:11434/
-
Install the packages in requirements.ttx
-
Run indexing
OLLAMA_HOST='0.0.0.0' python index_content.py
-
Test searching
QUERY="what is five plus five" USE_EMBEDDINGS=yes OLLAMA_HOST='0.0.0.0' python search.py
-
Play around by changing the models and other parameters
-
for a web interface - run the flask app:
FLASK_HOST=0.0.0.0 FLASK_PORT=8080 python flask_server.py
- then go to http://localhost:8080 on your browser.
-
Notifications
You must be signed in to change notification settings - Fork 0
License
study-iitm/llm-rishav
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
About
No description, website, or topics provided.
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published