-
-
Notifications
You must be signed in to change notification settings - Fork 764
Ollama
By default, ShellGPT leverages OpenAI's large language models. However, it also possible to use locally hosted models, which can be a cost-effective alternative. To use local models, you will need to run your own LLM backend server Ollama. Ollama provides an easy way to download and run Llama 2, Mistral, and other large language models locally.
Note
ShellGPT is not optimized for local models and may not work as expected.
Download and launch Ollama app.
curl https://ollama.ai/install.sh | sh
We can have multiple large language models installed in Ollama like Llama2, Mistral and others. It is recommended to use mistral:7b-instruct
for the best results. To install the model, run the following command:
ollama pull mistral:7b-instruct
This will take some time to download the model and install it. Once the model is installed, you can start API server:
ollama serve
Now when we have Ollama backend running we need to configure ShellGPT to use it. Check if Ollama backend is running and accessible:
sgpt --model ollama/mistral:7b-instruct "Hello World"
# -> I'm ShellGPT, your OS and shell assistant...
If you are running ShellGPT for the first time, you will be prompted for OpenAI API key. Just press Enter
to skip this step. If you got an error you can ask ShellGPT community for help.
Now we need to change few settings in ~/.config/shell_gpt/.sgptrc
. Open the file in your editor and change DEFAULT_MODEL
to ollama/mistral:7b-instruct
. Also make sure that OPENAI_USE_FUNCTIONS
is set to false
. And that's it! Now you can use ShellGPT with Ollama backend.
ShellGPT command-line tool for consulting with robots.