Error - Request too large for gpt-4o in organization org-e4yW1j9yPVVoLVlKMvWB0s6b on tokens per min (TPM): Limit 30000, Requested 36289. The input or output tokens must be reduced in order to run successfully. Visit https://platform.openai.com/account/rate-limits to learn more.
i have built a conversational agent by parsing a few blog posts. I am using Supabase Vector Store and OpenAI Embeddings. Though I am sending all the embeddings to the Agent, which could be the reason for the error above.
But how to build an agent that looks at 1000s of embeddings in the vector database ?
Hmm the point of the vector store is that you don’t have to send so many documents to the LLM. You want the vector store to pick out the most relevant to your user’s query so top 4-5 and not 1000.
If your goal is to recommend a huge number of blog post urls back to the user, you don’t necessarily need the LLM. Here’s an example where as long as your last node has an “output” json, it’ll reply to the chat.