OpenAI releases GPT-4.1 API: Million-token context and price cuts shake up the game

Published on
April 14, 2025
Contributors
No items found.
Subscribe for product updates and more:
Read about our privacy policy.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
Table of Contents:

OpenAI just dropped what might be its biggest update yet, making GPT-4.1 available to all developers through its API today. After months of anticipation and some reported delays, the AI powerhouse has delivered on CEO Sam Altman's promises while addressing two pain points that have long frustrated developers: context limitations and high costs.

Let's see what makes this release worth paying attention to.

The million-token milestone

The headline feature that everyone's talking about? GPT-4.1's massive 1,000,000 token context window. To put that in perspective, you could feed the entire text of Tolstoy's "War and Peace" into a single prompt and still have room to spare. 

"GPT-4.1 was specifically trained to maintain reliable attention across its full context length without losing track of information," an OpenAI spokesperson explained when pressed about how the model handles such extensive inputs.

Early testing suggests the model performs admirably with extensive documents, though there's a catch: accuracy does begin to degrade at extreme lengths, dropping from approximately 84% at 8K tokens to around 50% at the full 1M token capacity. Still, that's impressive territory few other mainstream AIs have ventured into before.

Three models, three price points

Breaking with tradition, OpenAI is launching not one but three distinct variants:

  • GPT-4.1 (base): The flagship model that pushes the boundaries of what AI can do
  • GPT-4.1 mini: A balanced offering that maintains the 1M token context at a more accessible price point
  • GPT-4.1 nano: OpenAI's first "nano" model, prioritizing speed and cost-effectiveness

What's especially interesting is that all three variants support multimodal capabilities (accepting both text and image inputs), a feature that was previously limited in the API. OpenAI claims GPT-4.1's image understanding capabilities often outperform the original GPT-4 on vision benchmarks, with even the mini variant showing superior performance on some image-based tests.

Are these price cuts for real?

Perhaps the most surprising aspect of today's announcement is OpenAI's aggressive pricing strategy. GPT-4.1 costs approximately 26% lower than GPT-4o for median queries, with the nano variant coming in at a fraction of the original GPT-4's price:

  • GPT-4.1 (base): $2.00 input / $8.00 output per million tokens
  • GPT-4.1 mini: $0.40 input / $1.60 output per million tokens
  • GPT-4.1 nano: $0.10 input / $0.40 output per million tokens

The company has also increased its prompt caching discount to 75% and offers an additional 50% reduction for batch API requests.

Performance that makes developers take notice

Beyond the flashy headline features, GPT-4.1 delivers substantial improvements in several key areas that matter to developers:

  • Coding prowess: A 21.4 percentage point improvement on the SWE-Bench Verified coding test compared to GPT-4o. One developer I spoke with noted, "It's producing code that's actually usable out-of-the-box now."
  • Instruction following: A 10.5 point gain on Scale AI's MultiChallenge benchmark, which translates to better adherence to complex instructions, something that's been a persistent frustration for many.
  • Knowledge base: Updated training data through June 2024, providing nearly three additional years of knowledge compared to the original GPT-4's September 2021 cutoff.

"We've optimized GPT-4.1 based on real-world feedback," an OpenAI representative noted. "It addresses many of the 'papercuts' developers experienced with previous models."

Where's ChatGPT in all this?

You might be wondering why you can't access GPT-4.1 in ChatGPT yet. Notably absent from today's announcement is integration with OpenAI's consumer-facing product. The company confirmed to press that at launch, GPT-4.1 is "API-only and not [in] ChatGPT yet," with consumer product integration expected later.

This staggered approach makes sense given Altman's previous cautions about potential delays due to GPU capacity issues. It's a pattern we've seen before (API first, ChatGPT later) as OpenAI manages the enormous computing resources required for these increasingly capable models.

Excitement meets practical concerns

Early developer feedback has been largely positive, with particular enthusiasm around the expanded context window and improved coding capabilities. "The function calling is way more reliable now," one developer wrote on X. "It actually follows the schema I give it without going off-script."

But not everyone's fully convinced. Some developers have expressed skepticism about the practical utility of the full 1M token context, noting performance degradation at extreme lengths. Others have voiced frustration with the June 2024 knowledge cutoff, particularly when working with rapidly evolving technologies and frameworks.

"Great, so it knows about Python 3.12 but has no idea about the major framework changes that happened this fall," lamented one developer in an online forum. "Guess I'll be explaining those in my prompts for the foreseeable future."

What's next for AI?

GPT-4.1 represents a significant milestone in OpenAI's journey toward what one spokesperson described as an "agentic software engineer" AI. With its improved tool usage, function calling capabilities, and reasoning abilities, the model is positioned to serve as the backbone for increasingly sophisticated AI agents capable of carrying out complex, multi-step tasks.

For developers looking to explore GPT-4.1's capabilities, OpenAI has updated its documentation with new guides on using image inputs, function calling, and taking advantage of the extended context window. Example code in the OpenAI Cookbook has also been refreshed to showcase the new model's features.

As AI capabilities continue to advance at breakneck speed, GPT-4.1 establishes itself as OpenAI's new flagship offering – smarter, more flexible, and more accessible than ever before. Whether it lives up to its ambitious promises in day-to-day development remains to be seen, but one thing is certain: the bar for what we expect from AI just got raised again.

And developers everywhere are already starting to reimagine what's possible.

⁀જ➣ Share this post:

Continue Reading

Similar readings that might be of interest:
🔍 Analysis

Moving beyond toxic team supervision

Discover how to break free from micromanagement and lead with trust, autonomy, and genuine connection. This piece explores actionable strategies for modern leaders to empower their teams and thrive in the evolving future of work.
Read post
Remote Academy

The Future of Work: Trends for 2024

So here are some notable predictions and trends that will shape our work lives in 2024.
Read post
Check-in Templates

Product Feedback Template

Ask your team continuously about your product roadmap, features, prioritization and management practices. Check this product feedback template.
Read post
Utilizamos cookies propias y de terceros para obtener datos de la navegación de nuestros usuarios y mejorar nuestros servicios. Si acepta o continúa navegando, consideramos que acepta su uso. Puedes cambiar la configuración o obtenga más información aquí.
Estoy de acuerdo