This event has ended. Showing historical data.

$682.27K
1
11

$682.27K
1
11
Trader mode: Actionable analysis for identifying opportunities and edge
This market will resolve according to the company that owns the model with the third-highest arena score based on the Chatbot Arena LLM Leaderboard (https://lmarena.ai/) when the table under the "Leaderboard" tab is checked on January 31, 2026, 12:00 PM ET. Results from the "Arena Score" section on the Leaderboard tab of https://lmarena.ai/leaderboard/text with the style control off will be used to resolve this market. If two models are tied for the third best arena score at this market's chec
Prediction markets currently show near-certainty that Google will own the third-best AI model at the end of February 2026. Traders are assigning this a 100% probability, which means they see it as virtually guaranteed. This forecast is based on the upcoming snapshot of the Chatbot Arena LLM Leaderboard, a public ranking where AI models are evaluated by thousands of user votes in head-to-head comparisons.
The extreme confidence stems from the current leaderboard standings and Google's consistent position. As of now, Google's Gemini models, like Gemini Ultra or its successors, have held a firm spot near the top of these rankings for an extended period. The leaderboard is not a static test but a live, crowd-sourced evaluation, making sudden shifts less common.
Another factor is the competitive landscape. The top two spots are widely expected to be held by OpenAI and Anthropic, given the strong performance of their flagship models like GPT and Claude. The race for third place appears less contested. While companies like Meta or xAI release capable models, they have not consistently surpassed Google's offerings in the Arena's public voting system over the long term. The market reflects a belief that Google's research investment and iterative improvements will maintain this gap through the snapshot date.
The only definitive date is February 28, 2026, at 12:00 PM ET, when the leaderboard will be checked to resolve the market. However, the odds could theoretically shift if a competitor releases a surprise model that rapidly climbs the rankings before that date. Major AI conferences, like Google's own I/O or developer events from Meta or other rivals, could be moments to watch for announcements of new models that might change the competitive order. A model would need to gain enough user votes in the Arena to overtake Google's entry in a short timeframe, which traders currently view as very unlikely.
Prediction markets are generally reliable for forecasting outcomes based on clear, publicly available data like a leaderboard score. They aggregate the collective judgment of many participants who have financial incentive to be correct. For this specific question, the reliability is very high because the resolution source is objective and transparent. The main limitation is time. The snapshot is over a year away, and the field of AI is unpredictable. A breakthrough from another lab could change things, but the market's 100% probability indicates traders see that risk as minimal.
Polymarket traders are pricing Google's "Gemini" model at a 100% probability to hold the third position on the LMarena.ai Chatbot Arena leaderboard for February 2026. This price indicates near-total certainty in the outcome. The market has seen over $233,000 in volume, suggesting high trader conviction rather than thin, speculative liquidity. The resolution is based on a single, public benchmark score at a specific future date, removing subjective interpretation.
The market's extreme confidence stems from the current competitive landscape and Google's historical performance. As of early 2024, the Chatbot Arena leaderboard has been consistently dominated by OpenAI's GPT-4 series and Anthropic's Claude 3 Opus in the top two positions. Google's Gemini Ultra model has firmly occupied the third-place spot for an extended period. Traders are betting this hierarchy will persist for two more years, viewing Google's vast research budget and infrastructure as a moat that competitors like Meta's Llama or open-source models cannot breach for a podium finish. The pricing reflects a belief that while the gap between first and second may shift, the third-place company is effectively locked.
The primary risk to this consensus is an unexpected technological leap from a well-funded challenger. Companies like xAI, with its Grok model, or a significantly advanced version of Meta's Llama could disrupt the standings if they release a model that outperforms Gemini on the specific Arena benchmark before the February 2026 checkpoint. A major strategic misstep or research delay at Google DeepMind could also open the door. However, the 100% price shows the market discounts these possibilities as highly improbable within the given timeframe. The most likely catalyst for a price shift would be a published paper or leak in late 2025 showing another model decisively outperforming Gemini Ultra in preliminary evaluations.
AI-generated analysis based on market data. Not financial advice.
This prediction market focuses on determining which company will own the third-ranked artificial intelligence model by the end of February 2026, as measured by the Chatbot Arena LLM Leaderboard. The market resolves based on the 'Arena Score' published on the leaderboard's website at a specific date and time. The Chatbot Arena, created by researchers at UC Berkeley and LMSYS Org, is a crowdsourced platform where users vote on the outputs of competing AI models in blind tests, generating an Elo-style rating system. This leaderboard has become a widely cited benchmark for comparing the conversational capabilities of large language models from different organizations. The specific interest in the third-place position reflects the competitive dynamics of the AI industry, where leadership is often contested between a few frontrunners, making the race for the next tier highly significant. Companies invest billions in developing these models, and their ranking influences investor confidence, developer adoption, and market perception. The outcome of this market provides a snapshot of the competitive hierarchy at a precise moment, offering insight into which organizations are successfully translating research into performant, publicly accessible AI systems.
The Chatbot Arena leaderboard was launched in May 2023 by researchers from UC Berkeley and LMSYS Org. It introduced a novel evaluation method based on anonymous, side-by-side human comparisons of model outputs, moving beyond static benchmark testing. In its first year, the leaderboard tracked the rapid ascent of models like OpenAI's GPT-4, which dominated the top position for an extended period. The competitive landscape shifted notably with the release of Anthropic's Claude 3 family in March 2024, whose Opus variant briefly challenged GPT-4's lead according to some arena scores. Throughout 2024, the rankings below the very top were highly volatile, with companies like Google, Meta, and Mistral AI frequently trading positions. This volatility established a precedent where the third-place ranking, in particular, could change hands multiple times within a quarter based on new model releases or updates. The historical data shows that a gap often exists between the top one or two models and the next tier, making the competition for third place a distinct and fiercely contested bracket. Past movements in these rankings have correlated with shifts in developer community interest, venture capital attention, and strategic partnerships for the companies involved.
The ranking of AI models has substantial economic implications. Companies with highly-ranked models attract more enterprise customers, developer mindshare, and investment. A top-three position can validate a company's research direction and justify continued high spending on compute resources and talent. For the broader technology sector, the hierarchy of AI capabilities influences which platforms businesses build upon, potentially creating ecosystem lock-in effects. The outcome also matters for AI safety and governance discussions. The competitive pressure to climb leaderboards can sometimes conflict with careful safety testing, making the composition of the top tier relevant for policymakers assessing industry self-regulation. Different companies have different approaches to model access, with some offering fully open-source models and others maintaining closed APIs, so the ranking affects the availability of advanced AI capabilities. Investors use these benchmarks as proxies for technological maturity when evaluating both public companies and private startups in the AI space.
As of late 2024, the Chatbot Arena leaderboard shows a tightly contested top group. OpenAI's GPT-4 series and Anthropic's Claude 3 Opus have been vying for the highest scores. The positions immediately below them, including the third-place rank, have seen rotation between models from Google (Gemini), Meta (Llama), and other specialized AI labs. The competitive environment is dynamic, with new model releases from these companies and others like xAI and Mistral AI announced regularly. Each release has the potential to reshuffle the rankings within weeks. The methodology of the Arena, relying on ongoing user votes, means scores can drift between major releases based on subtle changes in user demographics or voting patterns.
The Chatbot Arena uses an Elo rating system adapted from chess. Models gain or lose points based on the outcomes of anonymous head-to-head comparisons voted on by users. A win against a higher-rated model yields more points than a win against a lower-rated one, and vice versa for losses.
Unlike standardized tests like MMLU or HellaSwag that measure specific knowledge or reasoning, the Arena score reflects subjective human preference for conversational quality, creativity, and helpfulness in open-ended chats. It measures real-world user satisfaction more than narrow technical capability.
Historically, OpenAI, Anthropic, and Google have most frequently occupied the top three positions. However, companies like Meta and Mistral AI have also appeared in the top tier following major model releases, demonstrating the volatility of the rankings.
Yes, a single company can have several models on the leaderboard. For this market, resolution depends on the single highest-scoring model from each company. The company owning the model with the third-highest score overall wins the market, regardless of its other models' positions.
The leaderboard updates continuously as new votes are processed, but the public-facing table is typically refreshed multiple times per day. The market uses a specific snapshot taken on February 28, 2026, at 12:00 PM ET for resolution.
Educational content is AI-generated and sourced from Wikipedia. It should not be considered financial advice.
11 markets tracked

No data available
| Market | Platform | Price |
|---|---|---|
![]() | Poly | 0% |
![]() | Poly | 0% |
![]() | Poly | 0% |
![]() | Poly | 0% |
![]() | Poly | 0% |
![]() | Poly | 0% |
![]() | Poly | 0% |
![]() | Poly | 0% |
![]() | Poly | 0% |
![]() | Poly | 0% |
![]() | Poly | 0% |





No related news found
Add this market to your website
<iframe src="https://predictpedia.com/embed/nH2pV6" width="400" height="160" frameborder="0" style="border-radius: 8px; max-width: 100%;" title="Which company has the third best AI model end of January?"></iframe>