You can use Ollama to run your own model locally. Following the instruction of Ollama to install Ollama.
Start Ollama serving:
ollama serve
Then pull the model you want to use:
ollama pull <your-ollama-model>
You can also run Ollama in other ways by the instruction of Ollama.
Yaml is almost same as OpenaiGPT, just change the endpoint to the ollama url.
name: OpenaiGPTLLM
model_id: <your-ollama-model>
api_key: ${env| custom_openai_key, abcd} # api_key is not needed
endpoint: ${env| custom_openai_endpoint, http://<your-ollama-endpoint-domain>:11434/v1}
temperature: 0
vision: true
Then you can use your local Ollama serving.