SWE-1.5: Cognition AI Launches New Fast Agent Model

SWE-1.5: Cognition AI Launches New Fast Agent Model

Introducing SWE‑1.5: A New Fast Agent Model

Read the full announcement (via)

---

Overview

On the same day that Cursor released Composer‑1, Windsurf announced SWE‑1.5 — its latest frontier-size coding model boasting:

  • Hundreds of billions of parameters
  • Near state-of-the-art (SOTA) coding performance
  • Extreme speed: up to 950 tokens/second, which is
  • 6× faster than Haiku 4.5
  • 13× faster than Sonnet 4.5

> Powered by Cerebras hardware for inference — a strategic move to achieve high-speed serving.

Like Composer‑1, SWE‑1.5 is currently:

  • Accessible only via Windsurf’s editor
  • No public API yet
  • Built on a “leading open-source base model” (details undisclosed)

---

First Impressions

I tested it by asking for an SVG of a pelican riding a bicycle. The output was impressively fast:

image

The responsiveness felt genuinely faster than most coding models I’ve tried.

---

Training Insights

From the official blog post:

> SWE‑1.5 is trained on our state‑of‑the‑art cluster of thousands of GB200 NVL72 chips.

> We believe SWE‑1.5 may be the first public production model trained on the new GB200 generation.

>

> Our RL rollouts require high‑fidelity environments with code execution and web browsing.

> We use our VM hypervisor otterlink to scale Devin to tens of thousands of concurrent machines (see blockdiff).

> This enables high concurrency and aligns training environments with our Devin production environments.

Key similarities to Composer‑1:

  • Heavy use of reinforcement learning (RL)
  • Massive concurrent sandboxed coding environments

---

The Emerging Trend in Agent‑Based Coding Models

If you want to build a truly capable agent‑based coding tool:

  • Fine‑tune with reinforcement learning for integration with custom tools.
  • Use hundreds of thousands of concurrent simulated environments to train deeply specialized agents.
  • Align training infrastructure closely with real deployment settings.

Both Windsurf and Cursor are demonstrating this infrastructure‑heavy, concurrency‑driven approach.

---

Broader Ecosystem Connections

Beyond code generation, similar scaling concepts are emerging in multi‑platform AI content creation.

Open‑source projects like AiToEarn官网 and its GitHub repo enable creators to:

  • Generate, publish, and monetize AI content across platforms:
  • Douyin, Kwai, WeChat, Bilibili, Rednote (Xiaohongshu), Facebook, Instagram, LinkedIn, Threads, YouTube, Pinterest, X (Twitter)
  • Integrate analytics and model ranking tools
  • Build cross‑platform pipelines for scaling content workflows

Just as large‑scale concurrent environments help train SWE‑1.5:

  • AiToEarn’s distributed publishing system scales creative output.
  • Both approaches show how massively parallel infrastructure is becoming key to sustained AI‑driven productivity.

---

In summary: SWE‑1.5 is not only fast and powerful, but it's also emblematic of a wider trend — coupling RL training, custom tooling, and high‑concurrency clusters to push the limits of AI agents. This parallels innovations in other AI domains where infrastructure scalability determines real-world success.

Read more