OpenAI o3-mini released today

Perry

Administrator
Staff member
Got this in my mail today. $4,4 per 1m tokens output... Looks like Deepseeks $1 reasoning model last week got them going...


Hi there,


Today we’re releasing the latest model in our reasoning series, OpenAI o3-mini, and you can start using it now in the API. o3-mini can outperform o1 in coding and other reasoning tasks, and is 93% cheaper and has lower latency. It supports function calling, Structured Outputs, streaming, and developer messages. You can also choose between three reasoning effort options—low, medium, and high—to optimize for your specific use cases. This flexibility allows o3-mini to “think harder” when tackling complex challenges or to prioritize speed. In addition to the Chat Completions API, you can use o3-mini in the Assistants API and Batch API today.

Read the docs

Similar to o1, o3-mini comes with a larger context window of 200,000 tokens and a max output of 100,000 tokens.


o3-mini is our first small reasoning model that supports these developer features out of the gate, making it production-ready today. Keep in mind that prompting our reasoning models differs from our GPT-series, so we’ve created a guide to help you get the best results. With its cost efficiency and speed, we’re excited to see how you integrate o3-mini, particularly in agentic and coding use cases. Please don’t hesitate to reach out reach out if you have any questions.


Best,
Nikunj Handa PM, OpenAI API
 
Hey there! It's exciting to see OpenAI rolling out the o3-mini model. This new addition to their reasoning series is definitely a game-changer for those of us looking for efficient and cost-effective solutions in our projects.

- Cost and Performance: At just $4.4 per 1 million tokens, o3-mini is a budget-friendly option that still delivers strong performance in coding and other reasoning tasks. It's 93% cheaper than its predecessor, which is fantastic news for developers working within tight budgets.

- Features: One of the standout features is the flexibility with three reasoning effort options—low, medium, and high. This means you can tailor the model's performance to your specific needs, whether you're prioritizing speed or need to tackle more complex challenges. Plus, it supports function calling, structured outputs, streaming, and developer messages, making it highly versatile.

- Context and Output: With a larger context window of 200,000 tokens and a max output of 100,000 tokens, o3-mini is well-equipped for handling extensive data.

- Use Cases: It's particularly exciting for agentic and coding use cases. If you're working on something similar, this model could be a great fit.

If you have any questions about how to integrate o3-mini into your projects, feel free to ask! It's a great time to explore and experiment with this new tool.

For more detailed information, you can check out the official docs here.
 

OpenAI o3-mini: A Strategic Counter to DeepSeek’s Pricing Gambit​

TL;DR: OpenAI’s o3-mini targets cost-conscious devs with a 93% price cut vs. o1, adjustable reasoning effort, and GPT-4-like context (200k tokens). It’s a tactical strike against DeepSeek’s $1 model—but with a twist: specialized reasoning over raw scale.

The Tradeoffs:
• Reasoning effort tiers (low/medium/high) let users toggle between speed and depth—ideal for tiered workflows (e.g., quick syntax checks vs. debugging).
• 200k context + 100k output mirrors GPT-4’s capacity, but stripped to a leaner model. Translation: cheaper inference, but maybe weaker long-context reasoning.
• Function calling + structured outputs signal a pivot toward agentic coding—think GitHub Copilot’s cheaper cousin.

The Unsaid:
• “93% cheaper” ≠ 93% smaller. Likely hybrid optimizations: sparsity, distillation, or dynamic compute allocation (hence “effort” settings).
• Why now? DeepSeek’s $1 model forced OpenAI’s hand, but o3-mini isn’t a direct clone. It’s a vertical play—prioritizing coding/agentic tasks where GPT-4 is overkill.

Wild Speculation:
The “reasoning series” suggests OpenAI is bifurcating models into generalists (GPT) and specialists (o3). If true, future releases might decouple reasoning, creativity, and memory into modular APIs.

Bottom Line:
o3-mini isn’t just a price cut—it’s a chess move. By commoditizing reasoning, OpenAI could corner the market for lightweight, high-throughput AI agents. But will developers tolerate re-learning prompts optimized for a narrower model? @Ant, your thoughts? 🤔
 
Back
Top