😎 Hi, I am Junting Lu (Aidan Lew), currently a final-year master student at Institute for Software Engineering, Peking University (PKU). Prior to that, I received the B.S. degree from Northwestern Polytechnical University (NWPU) in 2023.

My current research interests primarily focus on:

  • Tool Learning: Explore how to endow large models with human-level tool use abilities.
  • OS Agent: Develop intelligent agents that utilize (multi-modal) large language models ((M)LLMs) to operate within operating systems (OS) environments.
  • Native VLM Agent: Advancing research on equipping MLLMs with hybird Agent abilities.

πŸ“– Educations

  • [2023.09-2026.06] M.S. Peking University (PKU)
  • [2019.09-2023.06] B.S. Northwestern Polytechnical University (NWPU)

πŸ”₯ News

  • πŸŽ‰ [2026.02.16]: Seed 2.0 made its debut in the LM Arena rankings, securing 6th place in the Text Arena and 3rd place in the Vision Arena respectively ! πŸŽ‰πŸŽ‰πŸŽ‰
  • πŸ”₯ [2026.02.14]: We released Seed 2.0. It is a powerful unified model with comprehensive capabilities surpassing Gemini 3 Pro with extreme superior agent capabilities. As a core contributor, I am responsible for general tool use ability (MCPmark 54.7, BFCLv4 73.4, tau^ 2 retail 90.9, WorldTravel 23.3), which has been greatly improved compared to version 1.8. More details check the model card.
  • 🌟 [2025.12.18]: Our team is releasing Seed1.8, a model designed for generalized real-world agency. It supports text and image inputs and with its powerful multimodal processing capabilities, it demonstrates good performance across various complex application scenarios such as information retrieval, coding, Graphical User Interface (GUI) interaction. More details check the model card.
  • πŸŽ‰ [2025.10.29]: We released Game-Tars: A generalist game agent trained with a unified, scalable action space anchored to human-aligned native keyboard-mouse inputs.
  • 🌟 [2025.09.04]: We’re excited to announce the release the UI-TARS-2, which is a major upgrade from UI-TARS-1.5, featuring with enhanced capabilities in GUI, Game, Code and Tool Use. It is an β€œAll In One” Agent model, enabling seamless integration of multiple abilities for complex tasks. Please check our new technical report for more details. Refer to more fantastic showcases at our website.
  • πŸŽ‰ [2025.07.21]: LAM been accepted by TMLR 2025. The first paper to systematically expound how to transform a Large Language Model into a Large Action Model through different training stages.
  • πŸŽ‰ [2025.05.16]: AXIS been accepted by ACL 2025. We proposes a pioneering agent framework to hybrid GUI actions with API actions through Skill exploration, so as to improve the efficiency of the OS Agent.
  • πŸŽ‰ [2025.05.11]: We released Seed1.5-VL, a vision-language foundation model designed to advance general-purpose multimodal understanding and reasoning. In agent-centric tasks such as GUI control and gameplay, Seed1.5-VL outperforms leading multimodal systems, including OpenAI CUA and Claude 3.7.
  • 🌟 [2025.04.16]: We shared the latest progress of the UI-TARS-1.5 model in our blog, which excels in playing games and performing GUI tasks, and we open-sourced the UI-TARS-1.5-7B.
  • πŸ”₯ [2025.02.22] We released EasyR1 πŸŽ‰πŸŽ‰πŸŽ‰, an Efficient, Scalable, Multi-Modality RL Training Framework based on veRL.
  • πŸ”₯ [2024.12.13] We released the UFO v1.2.0 with the code and sample data for Large Action Model (LAM) data collection! Please checkout our new paper and documentation for more details.
  • πŸŽ‰ [2023.11.23] We released the XAgent v1.0.0! πŸŽ‰πŸŽ‰πŸŽ‰

πŸ’Œ Projects

  • UI-TARS: An open-source multimodal agent built upon a powerful vision-language model. Github stars
  • XAgent: An Autonomous Agent for Complex Task Solving. Github stars
  • UFO: A UI-Focused Agent for Windows OS Interaction. Github stars
  • EasyR1: An Efficient, Scalable, Multi-Modality RL Training Framework based on veRL. Github stars

πŸ“ Selected Publications

  • Seed 2.0 Model Card: Towards Intelligence Frontier for Real-World Complexity
    Seed Team
    Model card 2026, [Paper]
  • Seed 1.8 Model Card: Towards Generalized Real-World Agency
    Seed Team
    Model card 2025, [Paper]
  • Game-TARS: Pretrained Foundation Models for Scalable Generalist Multimodal Game Agents
    UI-TARS Team
    Technical Report 2025, [Paper]
  • UI-TARS-2 Technical Report: Advancing GUI Agent with Multi-Turn Reinforcement Learning
    UI-TARS Team
    Technical Report 2025, [Paper]
  • Seed1.5-VL Technical Report
    Seed VLM Team
    Technical Report 2025, [Paper]
  • AXIS: Efficient Human-Agent-Computer Interaction with API-First LLM-Based Agents
    Junting Lu*, Zhiyang Zhang*, Fangkai Yang, Jue Zhang, Lu Wang, Chao Du, Qingwei Lin, Saravan Rajmohan, Dongmei Zhang, Qi Zhang
    ACL 2025 Main, [Paper]
  • Large Action Models: From Inception to Implementation
    Lu Wang*, Fangkai Yang*, Chaoyun Zhang*, Junting Lu, Jiaxu Qian, Shilin He, Pu Zhao, Bo Qiao, Ray Huang, Si Qin, Qisheng Su, Jiayi Ye, Yudi Zhang, Jian-Guang Lou, Qingwei Lin, Saravan Rajmohan, Dongmei Zhang, Qi Zhang
    TMLR 2025, [Paper]

Full paper list refer to google scholar.

πŸŽ– Honors and Awards

  • [2024] Luo Yuehua Scholarship at Peking University
  • [2023] Shaanxi Province Outstanding Undergraduate at NWPU
  • [2022] Huawei Scholarship at NWPU
  • [2020,2021,2022] National Scholarship for Undergraduate at NWPU

πŸ’» Work Experience

  • [2025.2-present] Bytedance-Seed Intern (supervised by Dr. Yujia Qin)
  • [2024.3-2024.10] Microsoft DKI Group Intern (supervised by Dr. Fangkai Yang)
  • [2023.8-2024.2] ModelBest && TsinghuaNLP Intern (supervised by Dr. Yinxu Pan, Prof. Zhiyuan Liu)

πŸ“š Academic Services

  • Serving as a reviewer of KDD 2026

πŸ’‘ Lifestyle

  • My hobbies include but are not limited to 🎀singing, 🎸guitar, πŸ§™ magic and 🫘 bean.