Hi i am trying to make a RAG pipeline for pdf, im using llama 3.1 as my llm which has been dowloaded locally on my server with L4 gpu how do i use that model with crew ai
PS: i want everything to run locally
Did you manage to get this working?
I suggest you look at Ollama and LMStudio. These should work out of the box with litellm which is used by crewAI for LLMs
iamsp
July 6, 2025, 3:14am
3
Hi @zinyando I am unable to use llm using LM Studio. Could you please guide if you have any idea?
I’ve posted my query here:
I’ve been working on a small project which needs multiple LLM calls. I’ve been using Together API but credit runs out within a day and I want a solution for this.
I have tried OpenRouter with deepseek-r1 from CrewAi docs, does not work. Hugging face the same way, no luck. I’ve been trying to use LLM Studio for local llm, that too is not working.
Please lmk if any of you’ve any solution.
This is how I’ve been using llm to provide llm=llm in agent definition.
lm = LLM(
model
="openrouter/deep…