Given that Phi3 appears to have one of the best ‘reasoning’ scores amongst open-source LLM’s I am looking at using Phi3 as the LLM for the manager within a hierarchical process.
I have ollama running on my Ubuntu/Linux box./ I have pulled ‘phi3:14b’
I use this static class to test various LLMs with various agents, etc.:
When I use Phi3 as the LLM for the manager nothing seems to work, I get errors like: " [2024-09-11 10:19:51][DEBUG]: == [Crew Manager] Task output: Agent stopped due to iteration limit or time limit." here’s my crew set-up:
UPDATE:
If I set all Agent & Crew LLMs → GTP4omini all is OK
if I set Phi3.5 as manager LLM in crew then it starts, but then I get errors unable to find other agents in the crew, or externam tools (SuperDev) stops working with what look like mising arg errors.
So it is something to do with Phi3. I’ve tried Phi3-mini, medium-128K & instruct. All with basically the same types of issues as described above.
Interesting idea. I have always found its the llm when iterations max although for simple tasks I always set the limit lower.
Nexusraven v2 always seems to work for function calling with strict sys instruction for json output on tool requests. @Dabnis is it getting stuck on tool calling? On the superdev issue, do you give it sys prompt to make requests in valid json? Do you use separate llm’s for general agent inference vs tool calling? Is it timing out when you use sequential?
Hi Moto,
re " do you give it sys prompt to make requests in valid json?" This is new to me, could you please explain?
As for tool calling: Only Phi3 models show this issue, when I switch back to GTP4omini, tooling works OK.
FYI: I’ve tried Phi3:mini, Phi3:medium-128k & Phi3-instruct.
I will try it in sequencial mdoe & update here.
@matt ‘agentops’ is yet another new area a CrewAI for me, I’ll investigate.
Pretty sure those are just llms being dumb problems.
On the system prompt, I just add, you MUST use valid JSON format when you use tools or call functions. Some llm endpoints allow you to specify json output schema e.g. gemini flash, etc. In the previous version of LM Studio you could force an llm to output json.