As the battle for large language model (LLM) supremacy intensifies, OpenAI finds itself under increasing pressure to maintain its dominant position. While official usage statistics remain undisclosed, the enthusiastic online reception to Anthropic’s Claude 3.5 release in mid-2024 suggests OpenAI is losing ground to its rival. As one user put it, “If Claude 3.5 outperforms GPT-4, why would we keep subscribing to OpenAI?”
Amid months of criticism over its pace of innovation, OpenAI appeared to be in decline. Besides the small-scale GPT-4 mini model, the company had few notable releases. In a symbolic blow, GPT-4 was recently bested by an open-source large model in head-to-head performance tests.
Financial concerns are also mounting. Last week, tech media outlet The Information estimated that OpenAI could face a staggering $5 billion funding gap this year as it burns through cash to train its data-hungry models. For a company that has raised over $11 billion to date, this is an ominous sign.
OpenAI Counterattacks with Voice Mode and Long Output Breakthroughs
Facing an existential threat, OpenAI has come out swinging. Overnight, the company announced two major updates designed to recapture its slipping LLM crown:
- Limited alpha testing of GPT-4 Voice Mode
- Launch of the GPT-4 Long Output model supporting 16X longer responses
GPT-4 Long Output Enables 200-Page Outputs
In a surprise move, OpenAI quietly launched an alpha test of its new GPT-4 Long Output model on its official website this week. Dubbed GPT-4-64k-Output-Alpha, the model will undergo testing for several weeks before a wider release.
The GPT-4 Long Output is a significant leap beyond OpenAI’s flagship GPT-4 model, supporting outputs up to a staggering 64,000 tokens. This is equivalent to around 200 pages of a novel – a 16-fold increase over the original GPT-4’s 4,000 token limit. To accommodate these expansive outputs, the input limit has been reduced to 64,000 tokens.
Pricing reflects the immense computational resources required, with OpenAI charging $6 per million input tokens and $18 per million output tokens – a 3X premium over standard GPT-4 output pricing.
Key Features and Tradeoffs of GPT-4 Long Output
- Extreme long-form output: Supports up to 64,000 tokens, enabling generation of novel-length text content.
- Input-output tradeoff: Users can input a maximum of 64,000 tokens to obtain the maximum 64,000 token output. Longer inputs mean shorter maximum outputs.
- Premium pricing: High costs reflect the immense computational expense of generating very long outputs.
- New use case exploration: OpenAI believes this model will unlock innovative applications like long-form scriptwriting, book authoring, and beyond.
- Unchanged maximum context: Despite radically enhanced output length, the maximum context window remains 128,000 tokens, the same as the base GPT-4 model.
When asked about the strategic impetus behind the Long Output model, an OpenAI spokesperson said:
“We’ve heard consistent feedback from customers that they need longer output contexts to achieve their application goals. This model is our answer – an experiment in pushing the boundaries of what’s possible with current LLM architectures. We’re excited to see all the creative ways developers leverage this new capability.”
Notably absent from OpenAI’s announcement were any claims of major improvements to the model’s base capabilities beyond output length. This suggests Long Output’s generative quality and fundamental skills are on par with the original GPT-4.
GPT-4 Voice Mode Wows Alpha Testers
The Long Output surprise follows OpenAI’s splashy May 2024 unveiling of GPT-4 Voice Mode at its “Spring Into the Future” launch event. After months of refinement, OpenAI has now quietly launched an alpha test of the feature with a small cohort of ChatGPT Plus subscribers.
Testers report being awestruck by the “surreal” experience of conversing naturally with GPT-4 through the ChatGPT interface. OpenAI is using the alpha to stress-test Voice Mode’s safety systems and generative quality before expanding availability.
“Since our initial Voice Mode demo, we’ve been maniacally focused on ensuring this paradigm-shifting technology is safe, stable, and delightful before we bring it to millions of users,” an OpenAI blog post read. “We can’t wait to get it into everyone’s hands (and ears).” The company expects to roll out Voice Mode to all ChatGPT Plus subscribers by fall 2024.
Alpha testers have taken to social media in droves to share their experiences, with reactions ranging from amazement to hilarity. A common thread has been good-natured ribbing of the voice model’s distinctive accent and occasional mispronunciations. But overall, testers rave about the model’s quick wit, broad knowledge, and engaging personality.
Holding Back Innovations for Long-Term Advantage
With these twin breakthroughs, OpenAI is determined to defend its pole position in the LLM race. But some industry watchers believe the company is still holding back its most significant innovations.
By parsing out new features over many months, the theory goes, OpenAI aims to keep customers subscribed for the long haul. The occasional headline-grabbing release that leapfrogs competitors is all part of the plan.
As the LLM wars escalate and funding pressures mount, OpenAI will need every trick in its playbook to stay ahead of hard-charging rivals like Anthropic, Google, and Meta. But if the rapturous reception to Voice Mode and Long Output is any indication, OpenAI’s innovation engine is still firing on all cylinders. The AI giant may be bloodied, but it’s definitely not beaten.
All eyes are now on the horizon for GPT-5 as the next salvo in the battle for LLM supremacy. One thing is certain: in the world of AI, there’s never a dull moment. And for OpenAI, the game is only beginning.