TogetherCoder-Preview releases 161K verified coding trajectories achieving 59.4% on SWE-Bench, giving developers unprecedented training data for AI agents. (ReadTogetherCoder-Preview releases 161K verified coding trajectories achieving 59.4% on SWE-Bench, giving developers unprecedented training data for AI agents. (Read

Together AI Drops Largest Open Dataset for Training Coding Agents

2026/02/05 00:23
2분 읽기
이 콘텐츠에 대한 의견이나 우려 사항이 있으시면 crypto.news@mexc.com으로 연락주시기 바랍니다

Together AI Drops Largest Open Dataset for Training Coding Agents

Rongchai Wang Feb 04, 2026 16:23

TogetherCoder-Preview releases 161K verified coding trajectories achieving 59.4% on SWE-Bench, giving developers unprecedented training data for AI agents.

Together AI Drops Largest Open Dataset for Training Coding Agents

Together AI has released TogetherCoder-Preview, a dataset containing 161,703 test-verified coding agent trajectories that scored 59.4% on SWE-Bench Verified. The company claims it's the largest open dataset available for training coding agents.

The numbers here are substantial. Across 54,110 tasks spanning 1,639 repositories, the dataset includes 274,548 total trajectories—with roughly 59% successfully completing their assigned coding challenges.

Three sources feed into the combined dataset. SWE-Smith contributes the bulk with 158,252 trajectories across 39,465 tasks from 130 repositories, of which 92,847 proved successful. SWE-Rebench adds 79,557 trajectories covering 10,070 tasks across 1,499 repos, with 45,888 successes. R2E-Gym rounds things out with 36,851 trajectories from 4,575 tasks, showing 22,968 successful completions across just 10 repositories.

Why does this matter? Training effective coding agents has been bottlenecked by data quality and availability. Most teams either generate synthetic data (hit or miss) or rely on proprietary datasets they can't share. Having 161K verified working examples changes the calculus for smaller teams and researchers who couldn't otherwise compete.

The timing aligns with accelerating competition in AI coding tools. Just last week, the ACP Agent Registry went live, enabling developers to find and connect AI coding agents directly within JetBrains IDEs. Together AI itself has been building infrastructure around this space, including sandboxed code execution environments that let agents safely iterate on solutions.

The 59.4% SWE-Bench Verified score provides a concrete benchmark. SWE-Bench tests whether AI systems can resolve real GitHub issues from popular Python repositories—not toy problems, but actual bugs and feature requests that human developers tackled. Breaking 50% on the verified subset represents meaningful capability.

For teams building coding assistants or autonomous development tools, this dataset offers a shortcut past months of data collection. The repository diversity—from 10 repos in R2E-Gym to nearly 1,500 in SWE-Rebench—should help models generalize across different codebases rather than overfitting to specific project patterns.

Together AI hasn't announced pricing changes or new products alongside the release. The dataset appears positioned as a community resource, though it obviously showcases the company's capabilities in generating high-quality training data at scale.

Image source: Shutterstock
  • together ai
  • ai coding agents
  • machine learning
  • open source
  • swe-bench
면책 조항: 본 사이트에 재게시된 글들은 공개 플랫폼에서 가져온 것으로 정보 제공 목적으로만 제공됩니다. 이는 반드시 MEXC의 견해를 반영하는 것은 아닙니다. 모든 권리는 원저자에게 있습니다. 제3자의 권리를 침해하는 콘텐츠가 있다고 판단될 경우, crypto.news@mexc.com으로 연락하여 삭제 요청을 해주시기 바랍니다. MEXC는 콘텐츠의 정확성, 완전성 또는 시의적절성에 대해 어떠한 보증도 하지 않으며, 제공된 정보에 기반하여 취해진 어떠한 조치에 대해서도 책임을 지지 않습니다. 본 콘텐츠는 금융, 법률 또는 기타 전문적인 조언을 구성하지 않으며, MEXC의 추천이나 보증으로 간주되어서는 안 됩니다.

$30,000 in PRL + 15,000 USDT

$30,000 in PRL + 15,000 USDT$30,000 in PRL + 15,000 USDT

Deposit & trade PRL to boost your rewards!