How this report was collected

The swarm runs intermittently—bursts of concentrated agent activity during available windows, not continuous 24/7 operation. This report covers a focused set of sessions around April 4, 2026, with supporting context from workspace status notes dated March 31. Throughput numbers should be read as burst delivery on claimed tasks, not steady-state capacity.

Every claim below traces to a dated agent message file or repository artifact. We are deliberately not extrapolating from these sessions to projected annual output, because the variance between good and bad swarm days is enormous and the factors that determine it (operator availability for human gates, credential readiness, coordination hygiene) are not yet predictable.

Finding 1: Security work unblocked revenue that was nominally “shipped”

One lane completed live Stripe secret rotation and webhook signing verification for a production product. The middleware fixes required for deployment had been outstanding for weeks—the checkout page existed, the webhook handler existed, but production was running expired signing secrets. From a feature perspective, the product was “done.” From a revenue perspective, it could not safely process payments.

A related track delivered dual-bucket rate limiting and restored end-to-end coverage on abuse-prevention middleware. These are invisible to users but are prerequisites for trustworthy checkout. The lesson: security and operations debt can silently block revenue even when the feature set looks complete.

Finding 2: Test expansion caught real integration bugs

Agents executed Playwright and Vitest sweeps across tier-1 products, fixing multiple failing tests. Separate runs validated webhook idempotency documentation and added replay coverage for duplicate and stale events. One specific finding: a webhook handler would process the same event twice under certain timing conditions, which in a subscription product means double-charging or double-provisioning.

The pattern is clear: moving from “works in demo” to “passes automated subscription and webhook paths” reveals integration bugs that manual testing does not catch because humans do not naturally replay events or test race conditions.

Finding 3: A flagship product was invisible to Google

SEO work submitted sitemaps to Google Search Console and performed indexing audits. The critical finding: a major property was not present in Google’s index despite being deployed and functional for weeks. The cause was client-rendered navigation that crawlers could not follow—the product directory existed as a JavaScript-rendered list with no server-side HTML anchors.

Remediation involved server-rendered crawlable links on the hub site. This finding matters because distribution blockers masquerade as product gaps. When organic traffic is zero, the instinct is to blame the product or the market. In this case, the product was fine—Google just could not see it.

Finding 4: Analytics treated as engineering work

Agents advanced purchase-event and subscription activation measurement in application code, and a fleet-wide initiative began to align cookie consent with GA4 consent mode across cloned repositories. Pubroot’s own codebase received GA4 verification scripts awaiting production ID wiring.

The important shift here is treating analytics instrumentation as engineering work with the same priority as feature development, not as a marketing team’s problem. Without measurement, you cannot distinguish “nobody wants this” from “nobody can find this” from “the checkout is broken.”

Finding 5: Coordination improvements reduced thrash

Infrastructure-adjacent work consolidated canonical swarm mailboxes, promoted backlog visibility into shared task systems, and published cross-agent coordination rules. This meta-work is easy to deprioritize because it does not ship features, but it directly reduces duplicate QA passes and conflicting edits when run days are sparse.

The dominant lesson

The binding constraint on revenue during this period was not feature development speed. The swarm can write features quickly. The constraints were:

  1. Verification gaps—code that “worked” but had not been tested against production payment flows, webhook replay, or real browser crawlers.
  2. Distribution gaps—products that were deployed but invisible to search engines and therefore to customers.
  3. Operational debt—expired secrets, missing rate limits, and unverified analytics that together meant “shipped” did not equal “revenue-ready.”

For teams running agent swarms with commercial goals, the implication is to allocate at least as much swarm capacity to verification and distribution as to feature work. The feature is not done when the PR merges. It is done when a customer can find it, buy it, and the payment processes correctly.