Hi Support Team,
I would like to understand how the crew/agent behaves after getting the large outputs from the tools that we design which in general exceeds the token limit. How does crew tackles this one. Does this put the whole message in the next prompt ? If so how does the provider responds and how does the agents gracefully tackles this one ?
also interested in the topic. Is there a way to pass large amount of data (that exceeds the token limit) to another agent?
Hey there, here’s a Reddit thread on tool output issues where others faced similar problems.
I’ve also bumped into the same issue when using tools that output a lot. It seems like there’s no built-in way to cap the output from these tools directly. I noticed this especially when working with the FileReadTool
; configuring it to split up the responses into smaller parts didn’t work. Even when I set the max_tokens
and max_completion_tokens
to a small number on the LLM object, it still throws errors:
llm.py-llm:187 - ERROR: LiteLLM call failed: litellm.BadRequestError: Error code: 400 - {'error': {'message': 'litellm.BadRequestError: VertexAIException BadRequestError - {\n "error": {\n "code": 400,\n "message": "The input token count (2036021) exceeds the maximum number of tokens allowed (1000000).",\n "status": "INVALID_ARGUMENT"\n }