AndroidWorld

AndroidWorld evaluates an agent's ability to operate in real Android GUI environments, completing multi-step tasks by perceiving screen content and executing touch/type actions.

GLM-5V-Turbo from Zhipu AI currently leads the AndroidWorld leaderboard with a score of 0.757 across 2 evaluated AI models.

Zhipu AIGLM-5V-Turbo leads with 75.7%, followed by Alibaba Cloud / Qwen TeamQwen3.6-27B at 70.3%.

Progress Over Time

Interactive timeline showing model performance evolution on AndroidWorld

State-of-the-art frontier
Open
Proprietary

AndroidWorld Leaderboard

2 models
ContextCostLicense
1
Zhipu AI
Zhipu AI
2
Alibaba Cloud / Qwen Team
Alibaba Cloud / Qwen Team
28B262K$0.60 / $3.60
Notice missing or incorrect data?

FAQ

Common questions about AndroidWorld.

What is the AndroidWorld benchmark?

AndroidWorld evaluates an agent's ability to operate in real Android GUI environments, completing multi-step tasks by perceiving screen content and executing touch/type actions.

What is the AndroidWorld leaderboard?

The AndroidWorld leaderboard ranks 2 AI models based on their performance on this benchmark. Currently, GLM-5V-Turbo by Zhipu AI leads with a score of 0.757. The average score across all models is 0.730.

What is the highest AndroidWorld score?

The highest AndroidWorld score is 0.757, achieved by GLM-5V-Turbo from Zhipu AI.

How many models are evaluated on AndroidWorld?

2 models have been evaluated on the AndroidWorld benchmark, with 0 verified results and 2 self-reported results.

What categories does AndroidWorld cover?

AndroidWorld is categorized under agents and vision. The benchmark evaluates multimodal models.

More evaluations to explore

Related benchmarks in the same category

View all agents
Humanity's Last Exam

Humanity's Last Exam (HLE) is a multi-modal academic benchmark with 2,500 questions across mathematics, humanities, and natural sciences, designed to test LLM capabilities at the frontier of human knowledge with unambiguous, verifiable solutions

visionmultimodal
74 models
MMMU

MMMU (Massive Multi-discipline Multimodal Understanding) is a benchmark designed to evaluate multimodal models on college-level subject knowledge and deliberate reasoning. Contains 11.5K meticulously collected multimodal questions from college exams, quizzes, and textbooks, covering six core disciplines: Art & Design, Business, Science, Health & Medicine, Humanities & Social Science, and Tech & Engineering across 30 subjects and 183 subfields.

visionmultimodal
62 models
MMMU-Pro

A more robust multi-discipline multimodal understanding benchmark that enhances MMMU through a three-step process: filtering text-only answerable questions, augmenting candidate options, and introducing vision-only input settings. Achieves significantly lower model performance (16.8-26.9%) compared to original MMMU, providing more rigorous evaluation that closely mimics real-world scenarios.

visionmultimodal
47 models
BrowseComp

BrowseComp is a benchmark comprising 1,266 questions that challenge AI agents to persistently navigate the internet in search of hard-to-find, entangled information. The benchmark measures agents' ability to exercise persistence in information gathering, demonstrate creativity in web navigation, and find concise, verifiable answers. Despite the difficulty of the questions, BrowseComp is simple and easy-to-use, as predicted answers are short and easily verifiable against reference answers.

agents
45 models
Terminal-Bench 2.0

Terminal-Bench 2.0 is an updated benchmark for testing AI agents' tool use ability to operate a computer via terminal. It evaluates how well models can handle real-world, end-to-end tasks autonomously, including compiling code, training models, setting up servers, system administration, security tasks, data science workflows, and cybersecurity vulnerabilities.

agents
39 models
MathVista

MathVista evaluates mathematical reasoning of foundation models in visual contexts. It consists of 6,141 examples derived from 28 existing multimodal datasets and 3 newly created datasets (IQTest, FunctionQA, and PaperQA), combining challenges from diverse mathematical and visual tasks to assess models' ability to understand complex figures and perform rigorous reasoning.

visionmultimodal
36 models