Well, with over 200k tokens, you might also want to consider an intermediate step where the scraping results feed into a RAG system, which would then be used during your data analysis phase.
@zinyando’s suggestion to use an LLM with a huge context window is great. It’s (almost) a form of Cache-Augmented Generation (CAG), just without the ‘cache’. Alternatively, using RAG is another way to reduce the dependency on those massive context windows.
I dont have access to gemini models, but after scraping a website the token length exceeded my llm context (>128k) , is there any variable that i can declare so that the length truncates after certain length or how to overcome this problem?
I am trying to scrap somewebsites based on the some topic where the websites are collected by serperdev tool and the collected website links will be scraped by firecrawl and i have to generate a report on the scraped content , this is exactly the usecase
I was trying to set-up a rag after scrapping then I understood I can pass the first task of scrapping because of the text lenght. It seems that at the end of the scrapping the agent trying to feed the llm right away while he should just write the result and nothing more.
Any idea how I can make the agent stop and just write the result?
I think the latest version of crewai this has been resolved.
Now if the this particular is recieved, crewai internally summarize all the past message and then hit the llm again, do update the crewai version and try this again.