Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

O1 model doesn'ts support streaming completions #430

Open
sheldonhull opened this issue Jan 30, 2025 · 0 comments
Open

O1 model doesn'ts support streaming completions #430

sheldonhull opened this issue Jan 30, 2025 · 0 comments

Comments

@sheldonhull
Copy link
Contributor

sheldonhull commented Jan 30, 2025

Background

With the introduction of newer reasoning models in v1.7.0, we've seen that while some tweaks have been made, streaming support for models like o1 was initially unavailable. This support varies based on the provider:

  • OpenAI: Recently added streaming support.
  • Azure OpenAI: Still lacks streaming support for o1.

Problem

The current implementation assumes that all results are streamed, which is incompatible with some o1 models. They require complete chat completion before results can be read, creating a gap in configuration and logic, especially for Azure OpenAI users.

Proposal

  1. Add a new configuration value:
    • stream: false (default to true).
  2. Introduce a CLI flag for manual override:
    • Use --stream false or --no-stream.

Seem reasonable? If so I'll see if I can knock that out.

ref: current limitations of o1 in azure openai

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant