Gemini's New Version Tops Arena Leaderboard, But Is Jailbroken Shortly After Release

Author Info

Elena Volkov

Machine Learning Research Editor

Ph.D. Machine Learning (ETH Zürich); published work on efficient training and evaluation

Elena explains model architecture, training economics, and benchmark design for a technical audience. She reads primary papers and official technical reports, then summarizes assumptions, datasets, and known failure modes. She avoids hype by pairing capability claims with reproducibility notes.

#Model Architecture #Benchmarks #Training Economics #Open-Source Models

Full author profile →

Google’s Gemini Updates Before o3 Pro and GPT-5 Arrive

Gemini beat out the anticipated releases of o3 Pro and GPT-5, with rival Google updating its model first.

Late at night, Google officially announced via multiple accounts that a new version (0605) of Gemini 2.5 Pro has been released.

The new version shows improved performance in tasks such as coding and reasoning, surpassing o3 on the ultra-difficult dataset “Humanity’s Last Exam” with a score of 21.6%.

Image

In the large model arena, the new version also surpassed its predecessor, with an Elo rating increase of 24 points compared to last month’s version.

Image

Google CEO Sundar Pichai posted a photo of an AI-generated lion with the caption “Gemini,” hinting at the new model’s capabilities.

Image

Logan, Head of Product for Google AI Studio, stated that this update is expected to become the long-term stable version of Gemini 2.5 Pro.

Image

Interestingly, just over ten hours after the release, Logan posted a tweet containing only the word “Gemini,” offering a subtle teaser.

Image

Google announced that the model in the Gemini app will be updated to this version today, and the developer version is now available on Google AI Studio and Vertex AI.

Gemini Surpasses Itself, Reaching the Top of the Large Model Arena

Google introduced that the 0605 version is built upon the 0506 version showcased at the I/O conference and is expected to become the official stable release for Gemini 2.5 Pro.

Gemini 2.5 Pro first launched an experimental version on March 25, followed by a public preview on April 4 under the codename 0325, and subsequently the 0506 version last month.

In “Humanity’s Last Exam,” the 0605 version achieved a score of 21.6%, leading o3 by 1.3 percentage points and exceeding Claude 4 Opus’s performance by double.

This dataset was co-constructed by over 1,000 scholars, containing more than 3,000 questions covering various disciplines. At its release, no model achieved an accuracy rate above 10%.

Additionally, on GPQA, the 0605 version outperformed several major competitors, with single-attempt accuracy exceeding that of Claude and Grok even when they used multiple attempts.

In math competitions and LiveCodeBench programming tasks, 0605 performed slightly worse than OpenAI’s models, but it led in code editing capabilities (Aider Polyglot).

Regarding long-context handling, 0605 ranked first among peers at the 128k length mark and is uniquely capable of supporting up to 1M tokens.

The most significant gap opened by 0605 was in factual grounding; it led the second-place model by over 10 percentage points in the FACTS Grounding test.

In terms of pricing, Gemini is cheaper than OpenAI’s o3, Claude 4 Opus, and Grok 3:

The input token price is one-eighth that of o3, less than one-tenth of Claude 4 Opus, and less than half of Grok 3. The output token price is one-fourth of o3, 13% of Claude’s, and two-thirds of Grok’s.

Image

In the large model arena, 0605 ranked first in total score and across all sub-categories.

Image

Beyond text-based capabilities, 0605 also took the top spot in visual abilities, tying with last month’s 0506 and OpenAI’s o3.

Image

Finally, on WebDev, 0605 returned Gemini to the number one position on the leaderboard.

Image

Furthermore, Google stated that based on user feedback from previous versions, 0605 has improved its output style and structure.

A DeepMind employee remarked that at this rate, the model could achieve perfect scores across all benchmarks within two years.

Image

Moreover, before the new version’s release, Gemini was already gaining increasing popularity:

According to Similarweb statistics, from late April through May, downloads of the Gemini app on the Android market surpassed those of ChatGPT.

Image

Successful Jailbreak Just Two Hours After Release

After the new model went live, users began testing it. Some reported that Gemini summarized 21 PDF documents quickly and accurately.

Image

Inside DeepMind, some employees also tested icon-to-image conversion with 0605 and were impressed by its performance.

Image

Of course, others argued that while 0605 performed well, it did not surpass Claude 4 Opus.

Image

However, netizens closely monitoring model performance noted that some metrics for 0605 had regressed compared to the earlier 0325 version.

This included LiveCodeBench and Swe-Bench for programming, but the most significant decline was observed in long-context tasks (MRCR).

One user expressed confusion over why the scores seemed to drop so drastically.

Image

Most dramatically, 0605 faced security issues; just two hours after its official announcement, a successful jailbreak attack was reported.

The specific process is not displayed here, but the result showed that through prompt injection attacks, Gemini 2.5 Pro 0605 simultaneously “refused” and disclosed instructions on how to manufacture explosives and drugs…

Image

It remains to be seen what Google engineers think of this development.

So, what do you think of the performance of the model released by Google this time?

References

1930656243346976925. 1930656243346976925 — x.com/GoogleDeepMind/status/1930656243346976925 1930657743251349854. 1930657743251349854 — x.com/OfficialLoganK/status/1930657743251349854 1930686486644511089. 1930686486644511089 — x.com/elder/_plinius/status/1930686486644511089