Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Stream LLM Output #46

Open
hyperdriveguy opened this issue Aug 1, 2024 · 0 comments
Open

Stream LLM Output #46

hyperdriveguy opened this issue Aug 1, 2024 · 0 comments
Labels
enhancement New feature or request

Comments

@hyperdriveguy
Copy link
Contributor

wirthlin commented Jun 26, 2024
Can you provide a more detailed summary of why this is difficult? It is something we can look at in the future but we will hold off on it for this summer.

hyperdriveguy commented Jun 26, 2024
Streaming uses (potentially asynchronous) generators. At the moment, LLM generation is abstracted behind several classes so that things can be properly logged and the right "chat branch" for a question can be used. Streaming will require generators at every level of abstraction, from the LLM generation to the web server. The frontend will also need to accept these generators as some kind a stream.
Basically, implementing it means changing the chat interfaces, the chat session manager (including the logging system), the flask web server, and the frontend JavaScript.

The pros outweigh the cons, and at this point it should be more feasible to implement.

@hyperdriveguy hyperdriveguy added the enhancement New feature or request label Aug 1, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant