Skip to content

Latest commit

 

History

History
26 lines (24 loc) · 800 Bytes

Ollama.md

File metadata and controls

26 lines (24 loc) · 800 Bytes

Ollama

You can use Ollama to run your own model locally. Following the instruction of Ollama to install Ollama.

Start Ollama serving

Start Ollama serving:

ollama serve

Then pull the model you want to use:

ollama pull <your-ollama-model>

You can also run Ollama in other ways by the instruction of Ollama.

Modify the yaml of llms

Yaml is almost same as OpenaiGPT, just change the endpoint to the ollama url.

name: OpenaiGPTLLM
model_id: <your-ollama-model>
api_key: ${env| custom_openai_key, abcd} # api_key is not needed
endpoint: ${env| custom_openai_endpoint, http://<your-ollama-endpoint-domain>:11434/v1}
temperature: 0
vision: true

Then you can use your local Ollama serving.