Performative rankings and technological innovations intensify the fierce competition among artificial intelligence models. The supremacy of Gemini-2.5-Pro asserts itself as an unexpected phenomenon, shaking up established hierarchies. Besides this performance, the AI landscape reveals unexpected players that disrupt the conventional order, such as Grok and Claude. The stakes of this dynamic are not limited to simple quantifiable advances but also generate debates about the future of machine learning technologies. Understanding these new frameworks is essential for grasping the transformation of the sector.
AI models at the top of the July 2025 ranking
The model Gemini-2.5-Pro stands out as the undisputed leader of the ranking established by LMArena. Its ascent knows no respite and reflects dominance across all assessed categories. Analyses show that it surpasses its competitors, a sign of continuous innovation in the field.
Among other notable models, OpenAI distinguishes itself with the models o3 and ChatGPT-4o. Although they are on the podium, they remain considerably distanced by the Gemini-2.5-Pro.
Performance ranking by category
Text
In the Text category, Gemini-2.5-Pro maintains a slight lead over its rivals, some of whom are dangerously close. OpenAI is also contending in this category, with positions being very tight.
Web development
In terms of Web development, Gemini, accompanied by DeepSeek and Claude models, forms the leading trio. OpenAI finds itself relegated to a disappointing 7th place, marking a decline in this sector.
Image analysis
In the field of Image analysis, Gemini maintains its dominant position. However, the gaps with OpenAI models are narrowing, suggesting increased competition.
Online search
Regarding Online search, the Gemini models are vying for first place with those from Perplexity, which are very present in this category. OpenAI, on the other hand, remains more discreet, displaying behind results.
Code assistants
The Code assistants category is dominated by DeepSeek, but Claude remains vigilant. Mistral finally succeeds in climbing to the top of the ranking, thus broadening the competition.
Image generation
For Image generation, GPT-Image-1 stays at the top with an impressive score. The competition, led by Flux, attempts to close the gap but struggles to compete with the leader.
Ranking criteria by LMArena
The ranking process established by the Large Model Systems Organization relies on an anonymous duel system. Each pair of models is evaluated, and users choose the one providing the most relevant answer. The results influence the Elo score, thus establishing a fair and precise ranking.
This algorithm creates various rankings based on performances recorded for each task or area. Thus, an overall ranking emerges, summarizing the performances of the competing models.
The list of the 10 AI models in July 2025
Here are the ten most performing artificial intelligence models, according to data collected by LMArena in July 2025:
- Gemini-2.5 Pro
- OpenAI o3
- ChatGPT 4o
- Grok-4
- GPT-4.5
- Kimi-k2
- Claude Opus 4 (version “thinking”)
- Claude Opus 4 (standard version)
- DeepSeek R1
- Gemini-2.5 Flash
Frequently asked questions
What are the top-performing AI models in July 2025?
The 10 top-performing AI models in July 2025 are: Gemini-2.5 Pro, OpenAI o3, ChatGPT 4o, Grok-4, GPT-4.5, Kimi-k2, Claude Opus 4 (version “thinking”), Claude Opus 4 (standard version), DeepSeek R1, and Gemini-2.5 Flash.
How does LMArena evaluate AI model performances?
LMArena uses an anonymous duel system where two models are compared blindly. Users choose the model that provided the best answer, thus influencing an Elo score that determines the rankings.
Why is Gemini-2.5 Pro the leader of the ranking in July 2025?
Gemini-2.5 Pro is the leader due to its exceptional performance across all observed categories, reaching the top in text, web development, and image analysis rankings.
What categories are included in the LMArena ranking?
LMArena provides rankings by categories such as text, web development, image analysis, online search, code assistants, and image generation.
Which OpenAI models are present in the top 10?
The OpenAI models present in the top 10 are OpenAI o3 and ChatGPT 4o, which rank second and third respectively.
How were the models ranked in the code assistant category?
In the code assistants category, DeepSeek holds the first place, closely followed by Claude with its 3.5 Sonnet models, while Mistral makes a strong entry into the ranking.
What trends emerge from the July 2025 rankings?
The trends show a dominance of Gemini in almost all categories, a tightening of the podium among the models, and a rise of new entrants like Claude and Grok.
Has Claude made a comeback in the rankings in July 2025?
Yes, Claude makes a comeback in the rankings by placing two of its models in the top 10, thus showing resilience in the face of increased competition.
What criteria can influence the Elo score of AI models?
The Elo score of models can be influenced by the results of duels against other models: defeating a higher-ranked model improves the score, while a loss against a lower-ranked model decreases it.
What are the main user concerns regarding AI models in 2025?
User concerns include the accuracy of responses, execution speed, the quality of content generation, and the ethics of artificial intelligence.