• About Us
  • Disclaimer
  • Contact Us
  • Privacy Policy
Tuesday, May 5, 2026
mGrowTech
No Result
View All Result
  • Technology And Software
    • Account Based Marketing
    • Channel Marketing
    • Marketing Automation
      • Al, Analytics and Automation
      • Ad Management
  • Digital Marketing
    • Social Media Management
    • Google Marketing
  • Direct Marketing
    • Brand Management
    • Marketing Attribution and Consulting
  • Mobile Marketing
  • Event Management
  • PR Solutions
  • Technology And Software
    • Account Based Marketing
    • Channel Marketing
    • Marketing Automation
      • Al, Analytics and Automation
      • Ad Management
  • Digital Marketing
    • Social Media Management
    • Google Marketing
  • Direct Marketing
    • Brand Management
    • Marketing Attribution and Consulting
  • Mobile Marketing
  • Event Management
  • PR Solutions
No Result
View All Result
mGrowTech
No Result
View All Result
Home Technology And Software

Terminal-Bench 2.0 launches alongside Harbor, a new framework for testing agents in containers

Josh by Josh
November 8, 2025
in Technology And Software
0
Terminal-Bench 2.0 launches alongside Harbor, a new framework for testing agents in containers



The developers of Terminal-Bench, a benchmark suite for evaluating the performance of autonomous AI agents on real-world terminal-based tasks, have released version 2.0 alongside Harbor, a new framework for testing, improving and optimizing AI agents in containerized environments.

READ ALSO

OpenAI turns its sold-out GPT-5.5 party into a monthlong Codex giveaway for 8,000 developers

Elon Musk Settles With The SEC For $1.5 Million After Years-Long Dispute Over His Twitter Investment

The dual release aims to address long-standing pain points in testing and optimizing AI agents, particularly those built to operate autonomously in realistic developer environments.

With a more difficult and rigorously verified task set, Terminal-Bench 2.0 replaces version 1.0 as the standard for assessing frontier model capabilities.

Harbor, the accompanying runtime framework, enables developers and researchers to scale evaluations across thousands of cloud containers and integrates with both open-source and proprietary agents and training pipelines.

“Harbor is the package we wish we had had while making Terminal-Bench," wrote co-creator Alex Shaw on X. "It’s for agent, model, and benchmark developers and researchers who want to evaluate and improve agents and models."

Higher Bar, Cleaner Data

Terminal-Bench 1.0 saw rapid adoption after its release in May 2025, becoming a default benchmark for evaluating agent performance across the field of AI-powered agents operating in developer-style terminal environments. These agents interact with systems through the command line, mimicking how developers work behind the scenes of the graphical user interface.

However, its broad scope came with inconsistencies. Several tasks were identified by the community as poorly specified or unstable due to external service changes.

Version 2.0 addresses those issues directly. The updated suite includes 89 tasks, each subjected to several hours of manual and LLM-assisted validation. The emphasis is on making tasks solvable, realistic, and clearly specified, raising the difficulty ceiling while improving reliability and reproducibility.

A notable example is the download-youtube task, which was removed or refactored in 2.0 due to its dependence on unstable third-party APIs.

“Astute Terminal-Bench fans may notice that SOTA performance is comparable to TB1.0 despite our claim that TB2.0 is harder,” Shaw noted on X. “We believe this is because task quality is substantially higher in the new benchmark.”

Harbor: Unified Rollouts at Scale

Alongside the benchmark update, the team launched Harbor, a new framework for running and evaluating agents in cloud-deployed containers.

Harbor supports large-scale rollout infrastructure, with compatibility for major providers like Daytona and Modal.

Designed to generalize across agent architectures, Harbor supports:

  • Evaluation of any container-installable agent

  • Scalable supervised fine-tuning (SFT) and reinforcement learning (RL) pipelines

  • Custom benchmark creation and deployment

  • Full integration with Terminal-Bench 2.

Harbor was used internally to run tens of thousands of rollouts during the creation of the new benchmark. It is now publicly available via harborframework.com, with documentation for testing and submitting agents to the public leaderboard.

Early Results: GPT-5 Leads in Task Success

Initial results from the Terminal-Bench 2.0 leaderboard show OpenAI's Codex CLI (command line interface), a GPT-5 powered variant, in the lead, with a 49.6% success rate — the highest among all agents tested so far.

Close behind are other GPT-5 variants and Claude Sonnet 4.5-based agents.

Top 5 Agent Results (Terminal-Bench 2.0):

  1. Codex CLI (GPT-5) — 49.6%

  2. Codex CLI (GPT-5-Codex) — 44.3%

  3. OpenHands (GPT-5) — 43.8%

  4. Terminus 2 (GPT-5-Codex) — 43.4%

  5. Terminus 2 (Claude Sonnet 4.5) — 42.8%

The close clustering among top models indicates active competition across platforms, with no single agent solving more than half the tasks.

Submission and Use

To test or submit an agent, users install Harbor and run the benchmark using simple CLI commands. Submissions to the leaderboard require five benchmark runs, and results can be emailed to the developers along with job directories for validation.

harbor run -d terminal-bench@2.0 -m "<model>" -a "<agent>" –n-attempts 5 –jobs-dir <path/to/output>

Terminal-Bench 2.0 is already being integrated into research workflows focused on agentic reasoning, code generation, and tool use. According to co-creator Mike Merrill, a postdoctoral researcher at Stanford, a detailed preprint is in progress covering the verification process and design methodology behind the benchmark.

Aiming for Standardization

The combined release of Terminal-Bench 2.0 and Harbor marks a step toward more consistent and scalable agent evaluation infrastructure. As LLM agents proliferate in developer and operational environments, the need for controlled, reproducible testing has grown.

These tools offer a potential foundation for a unified evaluation stack — supporting model improvement, environment simulation, and benchmark standardization across the AI ecosystem.



Source_link

Related Posts

OpenAI turns its sold-out GPT-5.5 party into a monthlong Codex giveaway for 8,000 developers
Technology And Software

OpenAI turns its sold-out GPT-5.5 party into a monthlong Codex giveaway for 8,000 developers

May 5, 2026
Elon Musk Settles With The SEC For $1.5 Million After Years-Long Dispute Over His Twitter Investment
Technology And Software

Elon Musk Settles With The SEC For $1.5 Million After Years-Long Dispute Over His Twitter Investment

May 5, 2026
Greg Brockman Defends $30B OpenAI Stake: ‘Blood, Sweat, and Tears’
Technology And Software

Greg Brockman Defends $30B OpenAI Stake: ‘Blood, Sweat, and Tears’

May 5, 2026
Image AI models now drive app growth, beating chatbot upgrades
Technology And Software

Image AI models now drive app growth, beating chatbot upgrades

May 4, 2026
xAI launches Grok 4.3 at an aggressively low price and a new, fast, powerful voice cloning suite
Technology And Software

xAI launches Grok 4.3 at an aggressively low price and a new, fast, powerful voice cloning suite

May 4, 2026
Sony Will Soon Settle A PlayStation Store Class Action Lawsuit For $7.8 Million
Technology And Software

Sony Will Soon Settle A PlayStation Store Class Action Lawsuit For $7.8 Million

May 4, 2026
Next Post
Grow a Garden Buttercup Wiki

Grow a Garden Buttercup Wiki

POPULAR NEWS

Trump ends trade talks with Canada over a digital services tax

Trump ends trade talks with Canada over a digital services tax

June 28, 2025
Communication Effectiveness Skills For Business Leaders

Communication Effectiveness Skills For Business Leaders

June 10, 2025
15 Trending Songs on TikTok in 2025 (+ How to Use Them)

15 Trending Songs on TikTok in 2025 (+ How to Use Them)

June 18, 2025
App Development Cost in Singapore: Pricing Breakdown & Insights

App Development Cost in Singapore: Pricing Breakdown & Insights

June 22, 2025
Comparing the Top 7 Large Language Models LLMs/Systems for Coding in 2025

Comparing the Top 7 Large Language Models LLMs/Systems for Coding in 2025

November 4, 2025

EDITOR'S PICK

The Resurgence Of Brand Loyalty

The Resurgence Of Brand Loyalty

March 17, 2026

NASA Astronauts Install High-Definition Cameras During Spacewalk

April 11, 2025
Abstract or die: Why AI enterprises can't afford rigid vector stacks

Abstract or die: Why AI enterprises can't afford rigid vector stacks

October 19, 2025
Build Your Agency to Thrive Through Transformations

Build Your Agency to Thrive Through Transformations

July 26, 2025

About

We bring you the best Premium WordPress Themes that perfect for news, magazine, personal blog, etc. Check our landing page for details.

Follow us

Categories

  • Account Based Marketing
  • Ad Management
  • Al, Analytics and Automation
  • Brand Management
  • Channel Marketing
  • Digital Marketing
  • Direct Marketing
  • Event Management
  • Google Marketing
  • Marketing Attribution and Consulting
  • Marketing Automation
  • Mobile Marketing
  • PR Solutions
  • Social Media Management
  • Technology And Software
  • Uncategorized

Recent Posts

  • Silverpush Launches SilverScreen™ in EMEA for Broadcast-Quality YouTube Ads
  • Why Gradient Descent Zigzags and How Momentum Fixes It
  • My Take on the 8 Best Employee Experience Software
  • Here’s how we’re helping Belgium’s farmers save water with AI.
  • About Us
  • Disclaimer
  • Contact Us
  • Privacy Policy
No Result
View All Result
  • Technology And Software
    • Account Based Marketing
    • Channel Marketing
    • Marketing Automation
      • Al, Analytics and Automation
      • Ad Management
  • Digital Marketing
    • Social Media Management
    • Google Marketing
  • Direct Marketing
    • Brand Management
    • Marketing Attribution and Consulting
  • Mobile Marketing
  • Event Management
  • PR Solutions