You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In the example provided at Llama Deploy Python Fullstack, the final output of the workflow is non-streaming. It only produces results once all generated tokens are complete.
As a result, I had to create my own FastAPI service.
Question:
Is there a more suitable modification to the workflow or llama-deploy that allows for direct and true asynchronous output without having to write my own FastAPI service?
Objective:
My main goal is to deliver the results of the workflow to users as quickly as possible.
The text was updated successfully, but these errors were encountered:
In the example provided at Llama Deploy Python Fullstack, the final output of the workflow is non-streaming. It only produces results once all generated tokens are complete.
As a result, I had to create my own FastAPI service.
Question:
Is there a more suitable modification to the workflow or llama-deploy that allows for direct and true asynchronous output without having to write my own FastAPI service?
Objective:
My main goal is to deliver the results of the workflow to users as quickly as possible.
The text was updated successfully, but these errors were encountered: