You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Regular enterprise use cases require support for spiky workloads without violating rate limits from LLM Providers. The Multi Agent Orchestrator should be able to:
Store the incoming messages in a queue
Check the classifier to identify top LLMs for the request
Check the rate limits and current consumption
Request the appropriate LLM endpoint
Receive the response and send it back to the requestor via Websockets
Solution/User Experience
The Multi Agent Orchestrator should be able to:
Store the incoming messages in a queue
Check the classifier to identify top LLMs for the request
Check the rate limits and current consumption
Request the appropriate LLM endpoint
Receive the response and send it back to the requestor via Websockets
Alternative solutions
Frameworks such as LLamaIndex supports async mode - but that targeted towards async IO non blocking http calls only. For a comprehensive enterprise grade Multi Agent Orchestrator solution, we need a custom solution
The text was updated successfully, but these errors were encountered:
kaustavbecs
changed the title
Feature request: Support for async Multi Agent Orchestrator with support for massive scale
Feature request: Add Asynchronous Message Queue Support with Rate Limiting
Dec 11, 2024
To have a async behaviour I think the best way is to handle this per agent (not at orchestrator level), because each agent is not necessary an LLM and because each agent could have different limits.
To implement this you could create a custom agent and handle the asynchronous communication with the LLM inside.
I could see this being implemented in a LambdaAgent for example and this lambda will have the asynchronous communication with the LLM by using queues or any other mechanism.
It would be interesting to create such agent to be part of built in agents.
Happy to review a solution before you can start the implementation if you want to contribute to this repostitory.
Use case
Regular enterprise use cases require support for spiky workloads without violating rate limits from LLM Providers. The Multi Agent Orchestrator should be able to:
Solution/User Experience
The Multi Agent Orchestrator should be able to:
Alternative solutions
The text was updated successfully, but these errors were encountered: