You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
With the introduction of newer reasoning models in v1.7.0, we've seen that while some tweaks have been made, streaming support for models like o1 was initially unavailable. This support varies based on the provider:
OpenAI: Recently added streaming support.
Azure OpenAI: Still lacks streaming support for o1.
Problem
The current implementation assumes that all results are streamed, which is incompatible with some o1 models. They require complete chat completion before results can be read, creating a gap in configuration and logic, especially for Azure OpenAI users.
Proposal
Add a new configuration value:
stream: false (default to true).
Introduce a CLI flag for manual override:
Use --stream false or --no-stream.
Seem reasonable? If so I'll see if I can knock that out.
Background
With the introduction of newer reasoning models in v1.7.0, we've seen that while some tweaks have been made, streaming support for models like
o1
was initially unavailable. This support varies based on the provider:o1
.Problem
The current implementation assumes that all results are streamed, which is incompatible with some
o1
models. They require complete chat completion before results can be read, creating a gap in configuration and logic, especially for Azure OpenAI users.Proposal
stream: false
(default totrue
).--stream false
or--no-stream
.Seem reasonable? If so I'll see if I can knock that out.
ref: current limitations of o1 in azure openai
The text was updated successfully, but these errors were encountered: