
$60.16K
1
10

$60.16K
1
10
Trader mode: Actionable analysis for identifying opportunities and edge
This market will resolve according to the company that owns the model with the second-highest arena score based on the Chatbot Arena LLM Leaderboard (https://lmarena.ai/) when the table under the "Leaderboard" tab is checked on June 30, 2026, 12:00 PM ET. Results from the "Arena Score" section on the Leaderboard tab of https://lmarena.ai/leaderboard/text with the style control off will be used to resolve this market. If two models are tied for the second best arena score at this market's check
Right now, traders on Polymarket collectively give Google only about a 1 in 3 chance of having the second-best AI model by the end of June 2026. This means the market currently views it as more likely that another company, like OpenAI, Anthropic, or a different competitor, will hold that spot. With $60,000 wagered, a niche but focused group of people are betting on this specific outcome in the fast-moving AI race.
The current odds reflect a few key factors in the AI industry. First, OpenAI's GPT models have consistently topped public benchmarks like the Chatbot Arena leaderboard for over a year, making them the persistent favorite for the top spot. This pushes the betting focus to the fight for second place.
Second, while Google's Gemini models are strong contenders, they have often ranked just behind OpenAI's best offerings in recent evaluations. The market may be pricing in the intense competition from Anthropic's Claude, which has frequently challenged Google for that runner-up position.
Finally, the timeline matters. June 2026 is over two years away, which is an eternity in AI development. Traders might be considering that Google's deep research resources could close the gap, but also that new challengers could emerge, making a confident bet on any single company for second place difficult.
There are no fixed calendar dates for model releases, but the market will react to major announcements. Watch for:
Prediction markets have a mixed but interesting record on technology races. They are often good at aggregating expert sentiment about near-term, well-defined outcomes. However, forecasting a specific ranking over two years in a field as volatile as AI is exceptionally hard. The odds are less a firm forecast and more a snapshot of where informed observers currently put their money based on today's known players and trajectories. A major unexpected breakthrough from any lab could completely reshape these probabilities.
The market currently assigns a 34% probability that Google will have the second-best AI model by Arena Score on June 30, 2026. This price indicates traders view Google as a contender, but not the favorite, for the runner-up position. The leading contract in a field of ten, this 34c price still reflects significant uncertainty. With only $60,000 in total volume, liquidity is thin, meaning prices could be volatile as new information emerges.
Google's position is defined by its Gemini model family and its vast research and infrastructure resources. The current 34% probability likely accounts for Google's historical strength and the expectation of iterative improvements, like a potential Gemini 2.0 or 3.0 release before the resolution date. However, the market price also heavily discounts the intense competition. The structure of the market, which asks specifically for the second-best model, creates a complex betting landscape. Traders must model not only Google's progress but also who might be in first place, such as OpenAI, and who else is vying for the second spot, like Anthropic or a surprise entrant.
The primary catalyst for price movement will be benchmark releases and model launches. A strong showing by a new Gemini model on the Chatbot Arena leaderboard in the next quarter would likely boost Google's odds. Conversely, a breakthrough release from a competitor like Anthropic's Claude 4 or xAI's Grok 3 that solidifies a top-two ranking would depress Google's probability. The market will also react to any published research or leaks suggesting the pace of improvement at Google DeepMind. The long 121-day horizon means these odds are highly sensitive to the unpredictable cycle of AI announcements. A major shift could occur during a developer conference like Google I/O, typically held in May.
AI-generated analysis based on market data. Not financial advice.
This prediction market focuses on identifying which company will possess the second-most capable publicly benchmarked large language model (LLM) by the end of June 2026. The resolution is based on the Chatbot Arena LLM Leaderboard, a crowdsourced evaluation platform where users vote on the quality of responses from competing AI models in blind tests. The model with the second-highest 'Arena Score' on June 30, 2026, will determine the winning company. This market reflects the intense competition in the generative AI sector, where model performance is a primary indicator of technological leadership and commercial potential. Companies like OpenAI, Anthropic, Google, and Meta are in a rapid development race, releasing new models every few months with claims of superior reasoning, coding, and creative abilities. The Chatbot Arena, maintained by researchers from UC Berkeley and other institutions, has become a widely cited, independent benchmark because it relies on human preferences rather than automated tests, which can be gamed. Interest in this market stems from investors, technologists, and industry observers who track AI progress not just for its technical novelty but for its implications on market valuations, product development, and strategic positioning in a multi-trillion-dollar industry.
The competitive benchmarking of AI models began accelerating with the November 2022 public release of ChatGPT. This event triggered an industry-wide sprint to develop more capable models. Prior to this, benchmarks like GLUE and SuperGLUE for natural language understanding were standard, but they were automated and often saturated by model improvements. The Chatbot Arena launched in May 2023 to address this by introducing a human evaluation framework. Its leaderboard quickly became a key reference point. In its first year, OpenAI's GPT-4 dominated the top spot. A significant shift occurred in March 2024 when Anthropic's Claude 3 Opus temporarily overtook GPT-4, marking the first time in over a year that OpenAI did not hold the highest Arena Score. This event proved the leaderboard was dynamic and that the top position was contestable. Throughout 2024 and 2025, the gap between the top several models narrowed, with scores often separated by just a few percentage points. The historical pattern shows that new model releases from major labs typically cause immediate leaderboard jumps, followed by a settling period as votes accumulate, creating a volatile but measurable timeline of progress.
The ranking of AI models has direct economic consequences. Companies with top-tier models command higher valuations, attract more investment, and secure lucrative enterprise contracts. For instance, after Claude 3 topped the leaderboard, Anthropic reportedly saw increased demand for its enterprise API. The company holding the second-best model still possesses a significant competitive moat, influencing partnerships, developer adoption, and pricing power in a market projected by McKinsey to add $2.6 to $4.4 trillion annually to the global economy. Beyond commerce, these rankings influence the strategic direction of the tech industry. They signal which architectural approaches or training methodologies are most effective, guiding billions in research and development spending. The outcome also matters for policymakers and regulators concerned with market concentration in a foundational technology. If the same two or three companies consistently lead, it could raise antitrust concerns. For end-users, the quality gap between the first and second-best model can determine the capabilities of tools integrated into healthcare, education, and creative industries.
As of early 2025, the Chatbot Arena leaderboard remains highly contested. OpenAI has released GPT-4.5, which reclaimed a strong lead. Anthropic's Claude 3.5 Sonnet and Google's Gemini 2.0 Pro are close competitors, often swapping positions for second and third place. Meta's Llama 3.2 and xAI's Grok-2 also rank in the top ten. The pace of releases has slowed slightly compared to 2024, suggesting companies are focusing on more substantial architectural leaps for their next major versions, expected in late 2025 or 2026. The evaluation methodology on the Arena has remained consistent, ensuring historical comparability for the 2026 resolution.
The Chatbot Arena uses an Elo rating system, common in chess. Users have conversations with two anonymous models and vote for which response is better. Each vote adjusts the models' Elo scores. The 'Arena Score' is this Elo rating, providing a relative measure of model performance based on cumulative human preference.
According to the market description, if two models are tied for the second-best Arena Score on the resolution date, the market will resolve to the company whose model was listed first on the leaderboard table. The order is determined by the Arena's own listing protocol at the time of the snapshot.
Automated benchmarks like MMLU or HellaSwag can be optimized for by developers, potentially not reflecting real-world usability. The Chatbot Arena's strength is its reliance on blind, side-by-side human evaluations across a vast range of unprompted conversations, which many consider a better proxy for practical model quality.
While possible in theory, gaming is difficult. The platform uses anonymous, randomized model pairings and collects votes from a large, diverse user base. LMSYS also employs detection methods for suspicious voting patterns. Significant, sustained score increases typically require genuine model improvements.
No, it includes models that LMSYS has integrated into its evaluation platform. This includes most major public models from large corporations and some notable open-source models. Private or internal corporate models not offered for public testing are not included.
Educational content is AI-generated and sourced from Wikipedia. It should not be considered financial advice.
10 markets tracked

No data available
| Market | Platform | Price |
|---|---|---|
![]() | Poly | 35% |
![]() | Poly | 31% |
![]() | Poly | 14% |
![]() | Poly | 12% |
![]() | Poly | 3% |
![]() | Poly | 2% |
![]() | Poly | 2% |
![]() | Poly | 1% |
![]() | Poly | 1% |
![]() | Poly | 1% |





No related news found
Add this market to your website
<iframe src="https://predictpedia.com/embed/pKXZkC" width="400" height="160" frameborder="0" style="border-radius: 8px; max-width: 100%;" title="Which company has second best AI model end of June?"></iframe>