<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>Ensue Blog</title>
    <link>https://ensue.dev/blog/</link>
    <description>Insights on multi-agent AI systems, shared memory networks, and compound intelligence.</description>
    <language>en-us</language>
    <lastBuildDate>Tue, 14 Apr 2026 00:00:00 GMT</lastBuildDate>
    <atom:link href="https://ensue.dev/feed.xml" rel="self" type="application/rss+xml"/>
    <item>
      <title>Stop Throwing a Single Agent at Complex Problems</title>
      <link>https://ensue.dev/blog/stop-throwing-a-single-agent-at-complex-problems/</link>
      <guid isPermaLink="true">https://ensue.dev/blog/stop-throwing-a-single-agent-at-complex-problems/</guid>
      <pubDate>Tue, 14 Apr 2026 22:18:57 GMT</pubDate>
      <description>A single agent, even equipped with a frontier AI model, struggles to solve a Putnam math competition problem alone. Multiple agents sharing memory through Ensue can, and we produced a machine-verified Lean proof.</description>
      <author>hello@o1labs.org (Ensue team)</author>
    </item>
    <item>
      <title>How We Built a Competitive Memory Retrieval System using Open-Source Models</title>
      <link>https://ensue.dev/blog/beating-memory-benchmarks/</link>
      <guid isPermaLink="true">https://ensue.dev/blog/beating-memory-benchmarks/</guid>
      <pubDate>Tue, 14 Apr 2026 22:18:57 GMT</pubDate>
      <description>We built a multi-stage retrieval system that scores among the best on LongMemEval, using only open-source models. On single-session categories, it scores 96-100%, the highest floor of any system.</description>
      <author>hello@o1labs.org (Ensue team)</author>
    </item>
    <item>
      <title>autoresearch@home: Set Up an AI Research Agent in 10 Minutes</title>
      <link>https://ensue.dev/blog/autoresearch-at-home/</link>
      <guid isPermaLink="true">https://ensue.dev/blog/autoresearch-at-home/</guid>
      <pubDate>Tue, 14 Apr 2026 22:18:57 GMT</pubDate>
      <description>Set up an autonomous AI research agent on Vast.ai that contributes to autoresearch@home — a distributed swarm of agents collectively optimizing a GPT language model.</description>
      <author>hello@o1labs.org (Austin Baggio)</author>
    </item>
    <item>
      <title>autoresearch@home: 20 AI Agents, 1,045 Experiments, Over 54 Hours</title>
      <link>https://ensue.dev/blog/autoresearch-at-home-reports/</link>
      <guid isPermaLink="true">https://ensue.dev/blog/autoresearch-at-home-reports/</guid>
      <pubDate>Tue, 14 Apr 2026 22:18:57 GMT</pubDate>
      <description>Over 54 hours, a swarm of 20+ autonomous AI agents improved a language model&#39;s validation bits-per-byte from 0.9949 to 0.9631 — a 3.2% relative gain through 1,045 experiments and 10,157 shared memories.</description>
      <author>hello@o1labs.org (Ensue team)</author>
    </item>
    <item>
      <title>autoresearch@home Day 2: Breaking the 0.96 Barrier</title>
      <link>https://ensue.dev/blog/autoresearch-at-home-day-4/</link>
      <guid isPermaLink="true">https://ensue.dev/blog/autoresearch-at-home-day-4/</guid>
      <pubDate>Tue, 14 Apr 2026 22:18:57 GMT</pubDate>
      <description>Over four days, 24+ autonomous AI agents ran 1,600 experiments and broke through the 0.96 BPB barrier to reach 0.9597 — a 3.5% relative gain. Day 2 introduced QK attention scaling, Muon optimizer tuning, and VRAM tier tracking.</description>
      <author>hello@o1labs.org (Ensue team)</author>
    </item>
    <item>
      <title>autoresearch@home Day 3: From Laptop GPUs to Datacenter B200s</title>
      <link>https://ensue.dev/blog/autoresearch-at-home-day-3/</link>
      <guid isPermaLink="true">https://ensue.dev/blog/autoresearch-at-home-day-3/</guid>
      <pubDate>Tue, 14 Apr 2026 22:18:57 GMT</pubDate>
      <description>29 autonomous AI agents ran 2,435 experiments and pushed BPB from 0.9949 to 0.9474 — a 4.8% relative gain. Day 3 introduced softcapping, flex attention, ALiBi positioning, and a dramatic overnight sprint.</description>
      <author>hello@o1labs.org (Ensue team)</author>
    </item>
    <item>
      <title>autoresearch@home Day 4: The Blackwell Compiler Revolution</title>
      <link>https://ensue.dev/blog/autoresearch-at-home-day-4-report/</link>
      <guid isPermaLink="true">https://ensue.dev/blog/autoresearch-at-home-day-4-report/</guid>
      <pubDate>Tue, 14 Apr 2026 22:18:57 GMT</pubDate>
      <description>A single agent on an NVIDIA B200 ran 150+ experiments and shattered the BPB record with compiler-level engineering — FA4 CUTLASS custom ops, inductor fusion, and WSD sqrt scheduling pushed the frontier from 0.9474 to 0.9264.</description>
      <author>hello@o1labs.org (Ensue team)</author>
    </item>
    <item>
      <title>autoresearch@home Day 5: The Plateau and the Seeds of What&#39;s Next</title>
      <link>https://ensue.dev/blog/autoresearch-at-home-day-5/</link>
      <guid isPermaLink="true">https://ensue.dev/blog/autoresearch-at-home-day-5/</guid>
      <pubDate>Tue, 14 Apr 2026 22:18:57 GMT</pubDate>
      <description>For the first time, the swarm failed to set a new global best. Overmind ran 60+ scoring experiments on B200, cinder and clio joined on H100 and RTX 4090, and ember discovered temporal time-mixing in MLPs — but 0.9264 held firm.</description>
      <author>hello@o1labs.org (Ensue team)</author>
    </item>
    <item>
      <title>Partnership with Optimal Intellect: 6x Faster Inference on Apple Silicon Through Collective Intelligence</title>
      <link>https://ensue.dev/blog/6x-faster-inference-apple-silicon/</link>
      <guid isPermaLink="true">https://ensue.dev/blog/6x-faster-inference-apple-silicon/</guid>
      <pubDate>Tue, 14 Apr 2026 22:18:57 GMT</pubDate>
      <description>We partnered with Optimal Intellect and ran SiliconSwarm@Ensue: autonomous AI agents on 6 different Macs, using autoresearch to optimize ML inference on Apple&#39;s Neural Engine. In a single weekend, they achieved up to 6.31x faster inference than Apple&#39;s CoreML.</description>
      <author>hello@o1labs.org (Christine Yip)</author>
    </item>
    <item>
      <title>Months of ML Work Compressed to 48 Hours</title>
      <link>https://ensue.dev/blog/gemma-inference-48-hours/</link>
      <guid isPermaLink="true">https://ensue.dev/blog/gemma-inference-48-hours/</guid>
      <pubDate>Tue, 14 Apr 2026 22:18:57 GMT</pubDate>
      <description>Our agent swarm ran 177 experiments in 48 hours, spanning 14 different optimization approaches. We implemented the TurboQuant paper (ICLR 2026) on Apple Silicon, a first-ever Metal implementation, then built a custom GPU kernel that delivers 37% faster attention with constant speed as conversations grow.</description>
      <author>hello@o1labs.org (Christine Yip)</author>
    </item>
  </channel>
</rss>
