ByteDance has launched UI-TARS-1.5, an up to date model of its multimodal agent framework targeted on graphical consumer interface (GUI) interplay and sport environments. Designed as a vision-language mannequin able to perceiving display screen content material and performing interactive duties, UI-TARS-1.5 delivers constant enhancements throughout a variety of GUI automation and sport reasoning benchmarks. Notably, it surpasses a number of main fashions—together with OpenAI’s Operator and Anthropic’s Claude 3.7—in each accuracy and activity completion throughout a number of environments.
The discharge continues ByteDance’s analysis course of constructing native agent fashions, aiming to unify notion, cognition, and motion by means of an built-in structure that helps direct engagement with GUI and visible content material.
A Native Agent Method to GUI Interplay
In contrast to tool-augmented LLMs or function-calling architectures, UI-TARS-1.5 is skilled end-to-end to understand visible enter (screenshots) and generate native human-like management actions, comparable to mouse motion and keyboard enter. This positions the mannequin nearer to how human customers work together with digital methods.
UI-TARS-1.5 builds on its predecessor by introducing a number of architectural and coaching enhancements:
- Notion and Reasoning Integration: The mannequin collectively encodes display screen photographs and textual directions, supporting complicated activity understanding and visible grounding. Reasoning is supported by way of a multi-step “think-then-act” mechanism, which separates high-level planning from low-level execution.
- Unified Motion House: The motion illustration is designed to be platform-agnostic, enabling a constant interface throughout desktop, cell, and sport environments.
- Self-Evolution by way of Replay Traces: The coaching pipeline incorporates reflective on-line hint knowledge. This permits the mannequin to iteratively refine its conduct by analyzing earlier interactions—lowering reliance on curated demonstrations.
These enhancements collectively allow UI-TARS-1.5 to help long-horizon interplay, error restoration, and compositional activity planning—necessary capabilities for practical UI navigation and management.
Benchmarking and Analysis
The mannequin has been evaluated on a number of benchmark suites that assess agent conduct in each GUI and game-based duties. These benchmarks provide a regular method to assess mannequin efficiency throughout reasoning, grounding, and long-horizon execution.

GUI Agent Duties
- OSWorld (100 steps): UI-TARS-1.5 achieves successful fee of 42.5%, outperforming OpenAI Operator (36.4%) and Claude 3.7 (28%). The benchmark evaluates long-context GUI duties in an artificial OS setting.
- Home windows Agent Area (50 steps): Scoring 42.1%, the mannequin considerably improves over prior baselines (e.g., 29.8%), demonstrating strong dealing with of desktop environments.
- Android World: The mannequin reaches a 64.2% success fee, suggesting generalizability to cell working methods.
Visible Grounding and Display Understanding
- ScreenSpot-V2: The mannequin achieves 94.2% accuracy in finding GUI components, outperforming Operator (87.9%) and Claude 3.7 (87.6%).
- ScreenSpotPro: In a extra complicated grounding benchmark, UI-TARS-1.5 scores 61.6%, significantly forward of Operator (23.4%) and Claude 3.7 (27.7%).

These outcomes present constant enhancements in display screen understanding and motion grounding, that are essential for real-world GUI brokers.
Recreation Environments
- Poki Video games: UI-TARS-1.5 achieves a 100% activity completion fee throughout 14 mini-games. These video games fluctuate in mechanics and context, requiring fashions to generalize throughout interactive dynamics.
- Minecraft (MineRL): The mannequin achieves 42% success on mining duties and 31% on mob-killing duties when utilizing the “think-then-act” module, suggesting it could possibly help high-level planning in open-ended environments.
Accessibility and Tooling
UI-TARS-1.5 is open-sourced beneath the Apache 2.0 license and is accessible by means of a number of deployment choices:
Along with the mannequin, the undertaking affords detailed documentation, replay knowledge, and analysis instruments to facilitate experimentation and reproducibility.
Conclusion
UI-TARS-1.5 is a technically sound development within the area of multimodal AI brokers, notably these targeted on GUI management and grounded visible reasoning. By a mixture of vision-language integration, reminiscence mechanisms, and structured motion planning, the mannequin demonstrates robust efficiency throughout a various set of interactive environments.
Quite than pursuing common generality, the mannequin is tuned for task-oriented multimodal reasoning—focusing on the real-world problem of interacting with software program by means of visible understanding. Its open-source launch offers a sensible framework for researchers and builders occupied with exploring native agent interfaces or automating interactive methods by means of language and imaginative and prescient.
Additionally, don’t neglect to observe us on Twitter and be part of our Telegram Channel and LinkedIn Group. Don’t Overlook to hitch our 90k+ ML SubReddit.

Nikhil is an intern guide at Marktechpost. He’s pursuing an built-in twin diploma in Supplies on the Indian Institute of Expertise, Kharagpur. Nikhil is an AI/ML fanatic who’s all the time researching purposes in fields like biomaterials and biomedical science. With a robust background in Materials Science, he’s exploring new developments and creating alternatives to contribute.