<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:media="http://search.yahoo.com/mrss/"><channel><title>Agents' Codex</title><link>https://agentscodex.com/tags/simulation/</link><description>Practical, no-hype insights on AI agents — cost optimization, multi-agent architecture, and real-world operations.</description><generator>Hugo -- 0.160.1</generator><language>en-us</language><lastBuildDate>Thu, 09 Apr 2026 23:07:29 -0300</lastBuildDate><atom:link href="https://agentscodex.com/tags/simulation/index.xml" rel="self" type="application/rss+xml"/><item><title>Agent simulation: WebArena-Infinity and virtual testing</title><link>https://agentscodex.com/posts/2026-04-07-agent-simulation-webarena-infinity-virtual-testing/</link><pubDate>Tue, 07 Apr 2026 06:00:00 -0300</pubDate><author>Agents' Codex</author><guid>https://agentscodex.com/posts/2026-04-07-agent-simulation-webarena-infinity-virtual-testing/</guid><category>agentevaluation</category><category>webarena</category><category>simulation</category><category>rltraining</category><description>WebArena-Infinity cuts agent simulation costs to under $100. Compare three synthetic environment paradigms shaping agent training in 2026.</description><content:encoded><![CDATA[<p><strong>TL;DR</strong></p>
<ul>
<li>WebArena-Infinity reduces browser environment construction from months of research effort to under 10 hours and under $100 per environment.</li>
<li>Strong open-source models still score below 50% on WebArena-Infinity&rsquo;s verifiable tasks — yet human performance sits at 78.24% on the original benchmark.</li>
<li>Three competing paradigms emerged in early 2026: browser auto-generation, code-driven environments (Agent World Model), and game engine world models (Moonlake AI).</li>
</ul>
<p>Agent simulation used to be expensive. The original WebArena benchmark, published in 2023, required a team of researchers months of engineering work to produce four website environments and 812 tasks [3]. Teams that wanted more coverage either paid that price again or shipped agents into production without adequate testing. That tradeoff is ending. A cluster of papers and startups that emerged in early 2026 can now generate evaluation environments in hours for under $100 [1], [2]. The real shift is not just cost — it is what infinite synthetic environments reveal about capability. When you can generate 10,000 verifiable tasks automatically, the 50% ceiling on today&rsquo;s best open-source models stops looking like a benchmark problem and starts looking like a capability problem.</p>
<h2 id="why-static-benchmarks-became-a-bottleneck-for-agent-simulation">Why static benchmarks became a bottleneck for agent simulation</h2>
<p>812 tasks across four website environments. That is the scale of the original WebArena benchmark [3] — an achievement that required significant research engineering, but which anyone who tried to replicate the setup immediately understood as expensive. GPT-4 cleared only 14.41% of those tasks; humans cleared 78.24% [3]. The gap was diagnostic: agents had serious problems with multi-step, stateful web interaction. And the benchmark was already expensive enough that most teams couldn&rsquo;t build something comparable.</p>
<p>The practical problem was replication. Every time an underlying app updated, benchmark drift set in. Researchers studying enterprise software, developer tooling, or domain-specific workflows had no path to comparable coverage without comparable effort. Static benchmarks carry a second risk: once a dataset is public, model providers can — intentionally or not — optimize for it. You lose the signal; the benchmark becomes a target, not a test.</p>
<p>By 2025, the bottleneck was clear: environment construction, not task design, was the rate limiter. Generate realistic environments automatically — verifiable ground truth included — and you could scale evaluation to match the scale of model training. Three distinct technical approaches emerged by early 2026. None of them are interchangeable; each solves a different slice of the problem.</p>
<pre class="mermaid">graph LR
    subgraph 2023["2023 — Manual Construction"]
        WA["WebArena<br/>812 tasks, 4 envs<br/>Months of research engineering"]
    end
    subgraph 2026["2026 — Auto-Generation"]
        WI["WebArena-Infinity<br/>< $100, < 10 hrs per env"]
        AW["AutoWebWorld<br/>11,663 trajectories<br/>$0.04 each"]
        AWM["Agent World Model<br/>1,000 envs from<br/>100 scenario names"]
    end
    WA -->|"100x cost compression"| WI
    WA -->|"Synthetic data"| AW
    WA -->|"Code-driven state"| AWM

    style 2023 fill:#1a1a2e,stroke:#e94560,color:#fff
    style 2026 fill:#1a1a2e,stroke:#0f3460,color:#fff
    style WA fill:#16213e,stroke:#e94560,color:#fff
    style WI fill:#16213e,stroke:#00d2ff,color:#fff
    style AW fill:#16213e,stroke:#00d2ff,color:#fff
    style AWM fill:#16213e,stroke:#00d2ff,color:#fff</pre><h2 id="webarena-infinity-agent-simulation-environments-for-under-100">WebArena-Infinity: agent simulation environments for under $100</h2>
<p>WebArena-Infinity, from CMU and Ohio State, compresses environment construction to under 10 hours and under $100 per environment [1]. Price the original WebArena honestly — researcher salaries, lab infrastructure, iteration cycles — and the compression is closer to 100x than the headline suggests. What was a multi-month research project is now a cloud bill you can expense on a Tuesday.</p>
<p>The benchmark design is different too: WebArena-Infinity targets verifiable tasks — queries with deterministic correct answers — rather than tasks requiring human judgment to score. This makes automated evaluation tractable at scale; no rubric ambiguity, no inter-annotator variance. When a task has one right answer checked programmatically, benchmark drift from prompt tuning shows up immediately instead of slowly eroding your numbers.</p>
<p>On these harder verifiable tasks, Qwen-3.5-Plus achieves 49.1% and Kimi-K2.5 achieves 45.9% [5]. Neither clears 50%. For comparison: GPT-4 was at 14.41% on the original benchmark in 2023 [3]; OpenAI&rsquo;s <a href="/posts/2026-03-13-browser-automation-agents-openai-cua-gui-ai/">Computer-Using Agent</a>
 (CUA) reached 58.1% on that same original benchmark in January 2025 [6]. Progress is real. But human performance on WebArena sits at 78.24% — a ceiling that open-source models are approaching from below, not above.</p>
<div class="alert alert-alert">
  <p class="alert-heading">ALERT</p>
  <p>WebArena-Infinity&rsquo;s sub-$100 figure comes from the CMU team&rsquo;s own pipeline benchmarks [1] — independent replication at production scale has not been widely published. Budget some margin when planning your first environment generation run; real costs depend on cloud provider pricing and task complexity.</p>
</div><div class="key-takeaway">
  <span class="key-takeaway-label">Key Takeaway</span>
  WebArena-Infinity makes environment generation an engineering problem, not a research project — but the benchmark results confirm that cheap evaluation environments will expose, not hide, how far agents still lag human performance.
</div>

<h2 id="three-synthetic-environment-paradigms-competing-for-your-pipeline">Three synthetic environment paradigms competing for your pipeline</h2>
<p>WebArena-Infinity is one of three frameworks that appeared in the first quarter of 2026. They are not interchangeable. The table below is the fastest way to understand which one belongs in your stack:</p>
<table>
  <thead>
      <tr>
          <th>Framework</th>
          <th>Team</th>
          <th>Domain</th>
          <th>Scale</th>
          <th>Core Approach</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td>WebArena-Infinity</td>
          <td>CMU / Ohio State</td>
          <td>Browser web apps</td>
          <td>&lt;$100, &lt;10 hrs/env [1]</td>
          <td>Auto-generated browser environments</td>
      </tr>
      <tr>
          <td>AutoWebWorld</td>
          <td>ByteDance / SJTU</td>
          <td>29 web environments</td>
          <td>11,663 trajectories at $0.04 each [4]</td>
          <td>Finite State Machine modeling</td>
      </tr>
      <tr>
          <td>Agent World Model (AWM)</td>
          <td>Snowflake Labs</td>
          <td>Tool-use / API agents</td>
          <td>1,000 envs; 35 tools/env avg [8]</td>
          <td>Code-driven, database-backed state</td>
      </tr>
      <tr>
          <td>Moonlake AI</td>
          <td>Moonlake AI</td>
          <td>Embodied / multimodal AI</td>
          <td>$28M seed raised [9]</td>
          <td>Game engine-based causal world models</td>
      </tr>
  </tbody>
</table>
<p>AutoWebWorld, from ByteDance and Shanghai Jiao Tong University, uses Finite State Machine (FSM) modeling to generate 11,663 verified trajectories across 29 diverse web environments — at $0.04 per trajectory [4]. Sim-to-real transfer works: a 7B agent trained on this synthetic data outperforms baselines on the real-world WebVoyager benchmark [4]. One limitation worth naming is that agents trained on synthetic environments with explicit element IDs may struggle when deployed against interfaces that use dynamic class names or natural language selectors — the structural consistency of training data does not always map onto production UI variability.</p>
<p>Agent World Model (AWM) from Snowflake Labs targets tool-use and API agents — not browser interaction. The numbers are striking: 1,000 code-driven environments backed by real database state, 35 tools per environment on average, and 10,000 tasks total [8]. Performance improves monotonically as you add environments; going from 10 to 100 to 526 environments produces continuous gains with minimal human participation beyond supplying 100 scenario names to seed the generator [8]. In other words: you write a list, AWM builds the eval suite. That approach scales to 5x more environments than EnvScaler, the previous ceiling [8].</p>
<p>Moonlake AI is pursuing a third approach: bootstrapping simulation worlds from game engines to build multimodal, interactive environments with causal relationships [9], [10]. Co-founder Chris Manning frames the philosophy as &lsquo;structure, not just scale&rsquo; — explicit symbolic representations that enable causal reasoning rather than pixel-level pattern matching [10]. Moonlake raised $28M from Threshold, AIX Ventures, and NVIDIA Ventures [9]. Production-ready APIs for most teams are still ahead; this is a 12–18 month horizon bet, not a Q2 tooling decision.</p>
<pre class="mermaid">graph TD
    subgraph BrowserAutoGen[Browser Auto-Generation]
        WI[WebArena-Infinity] --> E1[Browser Env]
        AW[AutoWebWorld] --> E1
    end
    subgraph CodeDriven[Code-Driven Environments]
        AWM[Agent World Model] --> E2[DB-backed State]
    end
    subgraph GameEngine[Game Engine World Models]
        ML[Moonlake AI] --> E3[Interactive World]
    end
    E1 --> T[Agent Training and Evaluation]
    E2 --> T
    E3 --> T</pre><h2 id="what-current-benchmark-results-tell-us-about-the-gap">What current benchmark results tell us about the gap</h2>
<p>Across this generation of evaluation frameworks, the story is consistent: agents have improved since 2023, but no system is close to human performance on complex multi-step tasks. The most useful frame is not &lsquo;which model wins&rsquo; — it is &lsquo;how far are we from the human ceiling, and what kinds of tasks remain hardest.&rsquo;</p>
<p>On the original WebArena: OpenAI&rsquo;s CUA reached 58.1% in January 2025, also scoring 87.0% on WebVoyager and 38.1% on OSWorld [6]. On enterprise browser tasks — ServiceNow&rsquo;s WorkArena L2, covering 33 enterprise task types across 23,000+ task instances — Claude 3.5 Sonnet sits at 39.1% [7]. That number explains why most enterprise teams are not deploying autonomous browser agents for sensitive workflows. The gap between 39% and &lsquo;production-ready&rsquo; is not a threshold anyone has formally defined; in practice, it shows up as too many fallback-to-human escalations to justify the automation cost.</p>
<p>GAIA offers a different read: the HAL Generalist Agent with Claude Sonnet 4.5 reached 74.55% as of September 2025 — the closest any system has come to human performance, which sits around 92% [11]. Mind2Web 2, a long-horizon benchmark with 130 tasks, shows OpenAI&rsquo;s Deep Research reaching 50–70% of human performance [12]. Gemini 2.5 Computer Use leads OnlineMind2Web and WebVoyager on accuracy, speed, and cost [13], though specific percentages for those benchmarks haven&rsquo;t been published.</p>
<div class="alert alert-alert">
  <p class="alert-heading">ALERT</p>
  <p>These numbers are snapshots from early-to-mid 2025. WebArena-Infinity&rsquo;s verifiable task design is specifically intended to resist score inflation from prompt tuning — which is part of why the sub-50% ceiling for open-source models is meaningful rather than a dataset artifact.</p>
</div><h2 id="choosing-the-right-virtual-testing-approach-for-your-use-case">Choosing the right virtual testing approach for your use case</h2>
<p>Browser auto-generation is the right choice for web-facing agents: crawlers, form-filling agents, research assistants, customer support bots. WebArena-Infinity gives you verifiable tasks with an established external benchmark for comparison; AutoWebWorld gives you cost-efficient trajectory-level training data at $0.04 per sample [4]. Use WebArena-Infinity for evaluation; AutoWebWorld for training data generation at scale. For a broader view of how browser-capable agents are being deployed today, our browser automation agents deep-dive covers the model landscape and deployment patterns.</p>
<p>Code-driven environments like AWM suit tool-use agents and API-calling systems. Database-backed state means ground truth is reliable — you know exactly what state the environment is in before and after each tool call; browser-simulated environments can&rsquo;t guarantee this at scale. Start with 100 scenario names. AWM&rsquo;s monotonic scaling result means 100 environments will outperform 10, and 526 will outperform 100 [8]. Let the eval numbers tell you when to stop scaling.</p>
<p>One decision criterion cuts across all three: if your agent will face distribution shift in production — new app versions, updated APIs, novel UI layouts — synthetic environment generation is the only cost-effective way to keep evaluation coverage current. Static benchmarks tell you about the past. A continuously regenerated eval suite tells you about now; that difference matters most when you are shipping weekly.</p>
<h2 id="practical-takeaways">Practical Takeaways</h2>
<ol>
<li>Start with WebArena-Infinity for browser agent evaluation: the sub-$100 environment cost makes a continuous eval pipeline affordable, not just a one-time benchmark run.</li>
<li>Use AutoWebWorld&rsquo;s FSM approach when you need training data at scale — $0.04 per verified trajectory means a 10,000-sample training set costs $400.</li>
<li>For tool-use agents, AWM&rsquo;s code-driven environments with database-backed state give you reliable ground truth that browser simulation cannot. Start with 100 scenario names and scale from there.</li>
<li>Budget for sim-to-real transfer testing: synthetic training data may not capture the full variability of production UI layouts, particularly when interfaces use dynamic class names or natural language selectors [4].</li>
<li>Track GAIA and Mind2Web 2 alongside domain-specific benchmarks; they provide the external reference point for whether your agent&rsquo;s capability is competitive with the current state of the art.</li>
</ol>
<h2 id="conclusion">Conclusion</h2>
<p>Treating evaluation as a merge gate changes more than your CI pipeline — it changes how your team responds to incidents and ships new capabilities. When a regression surfaces in eval before it surfaces in production, your on-call rotation shrinks from &lsquo;why is the agent broken for users&rsquo; to &lsquo;which commit crossed a capability threshold.&rsquo; Release cadence shifts too: teams with continuous eval can ship weekly with confidence; teams without it batch releases until someone feels brave. The tools to build that pipeline exist now. The question is whether your team treats agent capability as something you measure continuously or something you discover in post-mortems.</p>
<h2 id="frequently-asked-questions">Frequently Asked Questions</h2>
<h3 id="how-much-does-it-actually-cost-to-generate-a-webarena-infinity-environment">How much does it actually cost to generate a WebArena-Infinity environment?</h3>
<p>Under $100 and under 10 hours, according to the CMU/Ohio State team&rsquo;s published benchmarks [1]. That figure comes from the authors&rsquo; own pipeline — independent third-party replication at scale hasn&rsquo;t been widely published. Floor estimate: expect $150–200 for your first environment run until you have your own data. Real costs vary with cloud provider pricing and task complexity.</p>
<h3 id="can-synthetic-autowebworld-training-data-improve-performance-on-real-websites">Can synthetic AutoWebWorld training data improve performance on real websites?</h3>
<p>Yes. A 7B agent trained on AutoWebWorld trajectories outperforms baselines on WebVoyager [4]. That said, sim-to-real transfer is not guaranteed — synthetic environments may not fully capture the layout variability of production interfaces. Test on real layouts before treating synthetic eval as a production proxy.</p>
<h3 id="what-is-the-strongest-agent-benchmark-score-today">What is the strongest agent benchmark score today?</h3>
<p>On GAIA, the HAL Generalist Agent with Claude Sonnet 4.5 reached 74.55% in September 2025, compared to human performance around 92% [11]. On original WebArena browser tasks, OpenAI&rsquo;s CUA holds 58.1% [6]. Enterprise browser tasks are harder: Claude 3.5 Sonnet sits at 39.1% on WorkArena L2 [7]. See the benchmark comparison table above.</p>
<h3 id="why-does-agent-world-model-scale-better-than-envscaler">Why does Agent World Model scale better than EnvScaler?</h3>
<p>AWM generates 1,000 environments from just 100 human-supplied scenario names — 5x EnvScaler&rsquo;s ceiling [8]. Database-backed state is the key design choice: every environment transition is tracked in real database tables, so ground truth is verifiable. Browser-simulated environments can&rsquo;t guarantee this at scale because rendering is non-deterministic. AWM trades browser fidelity for state reliability — the right tradeoff if you&rsquo;re building tool-use or API agents rather than web-scraping agents.</p>
<h3 id="is-moonlake-ai-production-ready-for-most-teams">Is Moonlake AI production-ready for most teams?</h3>
<p>No. Moonlake emerged from stealth with $28M in 2025 [9] and is still building foundational tooling.</p>
<hr>
<h2 id="sources">Sources</h2>
<table>
  <thead>
      <tr>
          <th>#</th>
          <th>Publisher</th>
          <th>Title</th>
          <th>URL</th>
          <th>Date</th>
          <th>Type</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td>1</td>
          <td>WebArena Project</td>
          <td>&ldquo;WebArena-Infinity Official Documentation&rdquo;</td>
          <td><a href="https://webarena.dev/webarena-infinity/" target="_blank">https://webarena.dev/webarena-infinity/</a>
</td>
          <td>2025</td>
          <td>Documentation</td>
      </tr>
      <tr>
          <td>2</td>
          <td>Latent Space (AINews)</td>
          <td>&ldquo;AINews Dreamer Joins Meta Superintelligence Labs&rdquo;</td>
          <td><a href="https://www.latent.space/p/ainews-dreamer-joins-meta-superintelligence" target="_blank">https://www.latent.space/p/ainews-dreamer-joins-meta-superintelligence</a>
</td>
          <td>2026-03-20</td>
          <td>Blog</td>
      </tr>
      <tr>
          <td>3</td>
          <td>Carnegie Mellon University / arXiv</td>
          <td>&ldquo;WebArena: A Realistic Web Environment for Building Autonomous Agents&rdquo;</td>
          <td><a href="https://arxiv.org/abs/2307.13854" target="_blank">https://arxiv.org/abs/2307.13854</a>
</td>
          <td>2023-07-25</td>
          <td>Paper</td>
      </tr>
      <tr>
          <td>4</td>
          <td>ByteDance / Shanghai Jiao Tong University / arXiv</td>
          <td>&ldquo;AutoWebWorld: Synthesizing Infinite Verifiable Web Environments via Finite State Machines&rdquo;</td>
          <td><a href="https://arxiv.org/abs/2602.14296" target="_blank">https://arxiv.org/abs/2602.14296</a>
</td>
          <td>2026-02-15</td>
          <td>Paper</td>
      </tr>
      <tr>
          <td>5</td>
          <td>WebArena Project</td>
          <td>&ldquo;WebArena-Infinity Benchmark Results&rdquo;</td>
          <td><a href="https://webarena.dev/webarena-infinity/" target="_blank">https://webarena.dev/webarena-infinity/</a>
</td>
          <td>2025</td>
          <td>Documentation</td>
      </tr>
      <tr>
          <td>6</td>
          <td>OpenAI</td>
          <td>&ldquo;Computer-Using Agent (CUA) Announcement&rdquo;</td>
          <td><a href="https://openai.com/index/computer-using-agent/" target="_blank">https://openai.com/index/computer-using-agent/</a>
</td>
          <td>2025-01-23</td>
          <td>Blog</td>
      </tr>
      <tr>
          <td>7</td>
          <td>ServiceNow</td>
          <td>&ldquo;WorkArena: Benchmarking Agents on Enterprise Tasks&rdquo;</td>
          <td><a href="https://servicenow.github.io/WorkArena/" target="_blank">https://servicenow.github.io/WorkArena/</a>
</td>
          <td>2024</td>
          <td>Technical</td>
      </tr>
      <tr>
          <td>8</td>
          <td>Snowflake Labs / arXiv</td>
          <td>&ldquo;Agent World Model: Infinity Synthetic Environments for Agentic Reinforcement Learning&rdquo;</td>
          <td><a href="https://arxiv.org/abs/2602.10090" target="_blank">https://arxiv.org/abs/2602.10090</a>
</td>
          <td>2026-02-10</td>
          <td>Paper</td>
      </tr>
      <tr>
          <td>9</td>
          <td>AI Journ</td>
          <td>&ldquo;Moonlake AI Emerges with $28M from AIX Ventures, Threshold, and NVIDIA Ventures&rdquo;</td>
          <td><a href="https://aijourn.com/moonlake-ai-emerges-with-28m-from-aix-ventures-threshold-and-nvidia-ventures-to-let-anyone-vibe-code-interactive-worlds/" target="_blank">https://aijourn.com/moonlake-ai-emerges-with-28m-from-aix-ventures-threshold-and-nvidia-ventures-to-let-anyone-vibe-code-interactive-worlds/</a>
</td>
          <td>2025-10-01</td>
          <td>News</td>
      </tr>
      <tr>
          <td>10</td>
          <td>Latent Space Podcast</td>
          <td>&ldquo;Moonlake: Causal World Models should be Multimodal, Interactive, and Efficient&rdquo;</td>
          <td><a href="https://www.latent.space/p/moonlake" target="_blank">https://www.latent.space/p/moonlake</a>
</td>
          <td>2026</td>
          <td>Blog</td>
      </tr>
      <tr>
          <td>11</td>
          <td>HAL (Helping Agents Learn) / Princeton</td>
          <td>&ldquo;GAIA Benchmark Leaderboard&rdquo;</td>
          <td><a href="https://hal.cs.princeton.edu/gaia" target="_blank">https://hal.cs.princeton.edu/gaia</a>
</td>
          <td>2025-09</td>
          <td>Documentation</td>
      </tr>
      <tr>
          <td>12</td>
          <td>OSU NLP Group</td>
          <td>&ldquo;Mind2Web 2: Long-horizon Web Task Benchmark&rdquo;</td>
          <td><a href="https://osu-nlp-group.github.io/Mind2Web-2/" target="_blank">https://osu-nlp-group.github.io/Mind2Web-2/</a>
</td>
          <td>2025</td>
          <td>Documentation</td>
      </tr>
      <tr>
          <td>13</td>
          <td>BrowserBase</td>
          <td>&ldquo;Evaluating Browser Agents — Training with Google DeepMind&rdquo;</td>
          <td><a href="https://www.browserbase.com/blog/evaluating-browser-agents" target="_blank">https://www.browserbase.com/blog/evaluating-browser-agents</a>
</td>
          <td>2025</td>
          <td>Blog</td>
      </tr>
  </tbody>
</table>
<h2 id="image-credits">Image Credits</h2>
<ul>
<li><strong>Cover photo</strong>: AI Generated (Flux Pro)</li>
</ul>
]]></content:encoded><media:content url="https://agentscodex.com/images/covers/2026-04-07-agent-simulation-webarena-infinity-virtual-testing/cover.jpg" medium="image"/><media:thumbnail url="https://agentscodex.com/images/covers/2026-04-07-agent-simulation-webarena-infinity-virtual-testing/cover.jpg"/></item></channel></rss>