Scaling Synthetic Task Generation for Agents via Exploration

AutoPlay explores apps first—then synthesizes diverse, verifiable tasks at scale.

Training capable UI agents is bottlenecked by the scarcity of high-quality, grounded downstream tasks. AutoPlay tackles this by explicitly exploring interactive environments to discover capabilities and states before generating tasks. The result: diverse, executable, and verifiable tasks that reflect what the environment can actually do—without heavy human annotation.

The pipeline has two stages:

  • Exploration: An MLLM explorer systematically uncovers novel app states and functionalities.
  • Task Generation: A task generator uses exploration trajectories plus guideline prompts to synthesize tasks with built-in verifiability.

At scale, AutoPlay produces 20k tasks over 20 Android apps and 10k tasks over 13 Ubuntu apps. These tasks enable automated demonstration synthesis via an MLLM executor and verifier. Training on this data boosts agent success rates by up to 20.0% (mobile-use) and 10.9% (computer-use). Adding RL with verifier-based rewards yields a further +5.7% gain—evidence that exploration-grounded tasks are a foundation for scalable post-training.

Why it matters: Grounding tasks in actually reachable states solves a chronic problem in synthetic data generation—plausible but infeasible instructions. AutoPlay’s exploration-first design improves diversity, feasibility, and coverage, making it a powerful engine for building practical UI agents.

Paper: arXiv: AutoPlay
Register: https://www.AiFeta.com

#Agents #TaskGeneration #UIAutomation #MLLM #ReinforcementLearning #Scaling #DataGeneration #HCI

Read more