Replies: 7 comments
-
@abdinal1 are you runing your ollama and crewai in the same virtual env? have you tried to connect to the http://localhost:11434/ url to see if it will respond? |
Beta Was this translation helpful? Give feedback.
-
maybe crewAI community can help you https://discord.com/channels/1192246288507474000/1192247377705320479 the link direct to questions channel |
Beta Was this translation helpful? Give feedback.
-
The discord link doesn't work anybody have one that does. |
Beta Was this translation helpful? Give feedback.
-
Im running the ollama server on kaggle rescources as shown in the notebook: https://www.kaggle.com/code/aliabdin1/ollama-server Everything works fine unless my model size increases, I dont get out of memory it does do computations but as soon as 5 minutes are reached as a request it times out with the above mentioned error HTTPSConnectionPool When I access ollama without going over crewai's agents it does not timeout and let me request longer times but as soon as my request goes over crewai's framework I get a 5 minute timeout once I'm above that time. I havent found where I could set the timeout or where it might set a 5 minute treshold for a response (as default) |
Beta Was this translation helpful? Give feedback.
-
@Biancamazzi Hello all, i have created a gradio endpoint using hugging face model " CodeQwen1.5-7B" the gradio endpoint generate successfully but when I try to infrencing my crew agent with gradio endpoint its giving me error : but at first try its give me error but when I did same prompt second time its show this error. even my gradio endpoint working fine. also I retest my gradio endpoint its working fine. |
Beta Was this translation helpful? Give feedback.
-
Is your max_execution_time set to none? https://docs.crewai.com/core-concepts/Agents/#agent-attributes |
Beta Was this translation helpful? Give feedback.
-
I get
HTTPSConnectionPool(host='foobar.com', port=443): Max retries exceeded with url: /foo/bar
With certain models which take more than 5 minutes for an inference. If I swap to a tinyer model it works well but using bigger models results to this error when the request takes longer than 5 minutes. I have played around with a lot of different parameters like Ollama(timeout) etc. all of them had no effect so I came to the conclusion after testing barely on Ollama LLM like:
llm = Ollama(model="agent", base_url="/foo/bar")
that it has to be crewai because above line executed with 13 minutes of needed time.
Any idea how this can be solved @joaomdmoura ?
Beta Was this translation helpful? Give feedback.
All reactions