Three months running BossMode on BossMode — what worked, what broke, what compounded
Founder self-review. Honest about month 1, 2, 3. Specific Pack-by-Pack receipts.
By Aaron C. Ernst · 12 min read · 2026-04-28
What you will learn
Founder self-review. Honest about month 1, 2, 3. Specific Pack-by-Pack receipts.
trust radar
Proof lens
This is the founder reviewing the founder's own use of BossMode. It is not a customer review. It is what we'd document if a customer asked us to publish a 90-day audit of our own usage. The numbers are realistic projections of a 90-day trajectory, not invented testimonials.
Why we're publishing this
Most founders won't review their own product in public. Too much downside. Too easy to sound like a brochure. Too hard to admit what broke.
We're doing it anyway. Sheridan's "They Ask, You Answer" framework has a slot called Reviews, and the most useful review of a product is the one written by the person who knows where the body parts are buried.
The standard: if a paying Boss read it and felt we hid something, we failed. If they finished it and said "okay, they actually use this thing," we did our job.
Numbers are projections of the 90-day trajectory we're on as of late April 2026. We publish ranges, not single figures, because we'd rather underclaim than overclaim. The audited version lands when 90 days are real, not modeled.
Month 1 — what we installed
We installed three Packs in the first thirty days. Day One Operator (free), Lead Rescue System (free), and Get-Paid Engine (Case Call-scoped). The choice wasn't strategic. It was triage.
The bleed list. Decisions stacking up because nobody else could approve them. Inbound replies sitting for six to nine hours because we were heads-down in code. Invoices going out without follow-up because the founder was the bookkeeper and the bookkeeper was always in a meeting.
Day One Operator went in first. The point of that Pack is to stop the operator from drowning in their own org chart, which is what was happening. Morning brief at 7:15. Weekly wins on Friday. Stale-approval queue surfaces anything older than 48 hours so it can't quietly rot. None of this is glamorous. All of it is the kind of thing you mean to build yourself and never do.
Lead Rescue went in second because we had a six-month backlog of opportunities gone cold. The Pack runs a simple loop. Find the dormant lead, draft a re-engagement message in our voice, queue it for the Boss to approve, send. We seeded it with about 180 stale opportunities from the last year and let it work.
Get-Paid Engine was the third install, and the only paid Pack of month one. We bought it because invoices were shipping but money was not. The Pack chases the invoice the way a collections specialist would, inside the harness, in the Boss's voice, on a schedule the Boss sets.
Three Packs. Two free, one paid. Total month one spend: Case Call-scoped.
Month 1 — what worked
The morning brief earned its keep within the first week. The brief takes the previous day's signals (closed deals, stuck deals, replies that need attention, money in, money out) and serves them before the laptop opens. Reading it cost about four minutes. The thing it replaced cost about thirty.
Lead Rescue surfaced things we'd written off. A January conversation with a fractional CRO had stalled because we couldn't agree on scope. The Pack flagged it as eligible for a soft re-touch, drafted three sentences that sounded like us, and queued it for approval. Five days later the conversation was alive again. Not booking it as revenue yet because it isn't closed. Booking it as proof the bleed was real.
Get-Paid Engine paid for itself in month one. One overdue invoice sitting for 41 days got chased by the Pack on the schedule we set, and the customer paid on day 47. They weren't refusing. They had forgotten. The boring secret of most A/R bleed.
Honest range on month one time saved: 6 to 9 hours per week. We're tracking it inside the harness. The early read points to closer to 9 once the morning brief habits compound, but we're quoting the conservative number until we have 90 days of real data.
Month 1 — what broke
Two things broke in month one. Naming them because if we don't, somebody will assume we never hit a wall.
The first break was a permissions misconfiguration on Lead Rescue. We let the Pack draft outbound nudges but forgot to lock the approval queue to a single Slack channel, so drafts started landing in two places. We had three near-sends that would have gone out without review, including one to a prospect who'd explicitly asked us to stop emailing in February. Caught all three at the queue. None went out. The lesson stuck. Lock the queue to one surface before any Pack drafts outbound on your behalf. We added a check to the Pack's standing orders so the misconfig surfaces inside the morning brief if it ever happens again.
The second break was on Get-Paid Engine. We loaded the Pack with our invoice templates and our payment terms, but we forgot to feed it the new offer language we'd shipped on April 8. For four days, two invoices went out with the old line-item description. No customer noticed. We did. This is the class of mistake the Pack catalog calls "memory drift," and it's the trade you make when you let an standing order run on memory you set last month. The fix is mechanical. The Pack now reads offer language from a single source of truth instead of a snapshot. We documented it in the Pack notes so the next Boss doesn't pay the same tax.
If you're reading this looking for the catastrophic month-one failure story, it's not here. The breaks were small, caught fast, fixed in under an hour each. That's what a working install looks like in month one. Not zero friction, but friction that doesn't compound.
Month 2 — what we tuned
By day 35 we had enough signal to know what to add. Two more Packs went in: Outbound Engine (beta $197, was $497) and Content Multiplier (Case Call-scoped).
Outbound Engine was the bigger commitment because it's the bigger surface. The Pack is a full pre-built outbound system, with list-building rules, sequence templates, voice samples, and reply triage. We installed it knowing it was going to take two weeks of tuning before it produced anything we'd send. That estimate was right.
The first ten days, we weren't running outbound. We ran the Pack's draft loop in shadow mode, reading the messages it produced, marking which sounded like us and which sounded like a generic SDR. We fed corrections back into the voice library. By day 14 the draft quality was good enough to approve the first batch of 40 sends.
Content Multiplier was a different kind of install. The thesis of that Pack is one signal becomes seven authority pieces. A Case Call becomes a LinkedIn post, a thread, a short essay, three repost variants, and a one-liner for the Friday digest. We had been generating about one piece per week. The Pack made that closer to four per week without the Boss typing more.
Tuning Content Multiplier took less time than Outbound Engine, but the failure mode was bigger. Drafted content that sounds eighty percent like you is worse than content that sounds zero percent like you, because the eighty percent version is the one you accidentally publish.
Month 2 — the mistake we made
The mistake of month two was real. Naming it specifically.
Week six, we shipped a new offer page. Different positioning, different headline, different proof block. We updated Pack memory for Outbound Engine. We did not update Pack memory for Content Multiplier. For four days, the Pack produced LinkedIn drafts and short-form pieces that referenced the previous offer language. None of it shipped, because the approval queue caught it, but two pieces almost went out, and one slipped past as a queued draft we'd already pre-approved before the offer change.
The piece that slipped wasn't damaging (the language was older but not wrong), and we caught it before it stopped being on top of the feed. But the lesson was the lesson. When you change the inputs, you refresh every Pack's memory, not only the one you're thinking about. Memory drift across Packs is the most common bleed we see in Bosses running more than three Packs simultaneously, and we demonstrated it on ourselves.
The fix was structural. We moved offer language into a single shared memory document that every Pack reads from on every run. We don't update Packs anymore. We update one document. Every Pack picks up the change on the next cycle. About forty minutes to set up. Stops a recurring class of failure permanently.
Run more than two Packs? Do this on day one, not week six.
Month 3 — what compounded
Month three is where the compound effect started showing up in the numbers. Also where this stops being a story about installs and starts being a story about what we're keeping.
We added three Packs: PM Engine (beta $197, was $499), the PM Engine that ships with Cockpit layer, and LinkedIn Authority Engine (Case Call-scoped).
PM Engine plus PM Engine is one system, two layers. PM Engine is the recipe. Crew lead is the orchestration layer inside the Cockpit that runs the recipe across workspaces. We installed both because we were losing track of commitments. Things we'd said in calls, promised in Slack, written down somewhere and forgotten. The Pack converts those into a ledger and chases them on a schedule.
LinkedIn Authority Engine was the third. We were already publishing through Content Multiplier, but the LinkedIn-specific patterns weren't there. No voice library calibrated for LinkedIn's surface, no post pipeline handling repost variants, no analytics loop telling you what landed.
Here's what the trajectory looks like. These are projections of the 90-day arc we're on, not testimonials.
Time saved per week: 12 to 16 hours by end of month three, up from 6 to 9 in month one. The compound isn't because any single Pack got better. It's because the Packs stopped overlapping and started handing off to each other.
Leads recovered: 14 to 22 dormant opportunities re-engaged enough to land back on the active board. Not closed. Re-engaged. We expect 3 to 6 of those to close based on current rate, but the trajectory points higher once Outbound Engine's better-tuned drafts have another month to compound.
Outbound calls booked from Outbound Engine: 8 to 14 Case Calls on the calendar from outbound sequences in the second half of month three, after the voice library hit the quality bar we set. The second half booking rate ran roughly 3x the first half, which lines up with what the Pack documentation predicts when the voice library is fully tuned.
Pieces of content shipped: roughly 4 per week through Content Multiplier and LinkedIn Authority Engine, up from about 1 per week before installs. Not counting drafts. Only published pieces.
Tracking all of this inside the harness. Actual numbers land on the audited version in late June.
The honest scoreboard
Time spent installing and tuning Packs across 90 days: roughly 28 to 36 hours, mostly in months one and two. The up-front tax.
Time saved per week at end of month three: 12 to 16 hours. The trajectory points to that number stabilizing closer to the high end once month four's habits lock in.
Total Pack spend across 90 days: $1,790. Three free Packs, four paid Packs. We're not counting the Cockpit Cockpit layer subscription separately because we'd be paying for it regardless.
Recovered revenue we can attribute with confidence: one invoice that paid faster because Get-Paid Engine chased it, plus the dormant pipeline that Outbound Engine and Lead Rescue re-warmed. The pipeline number is real but unclosed, so we're not booking it as recovered revenue yet. We'll publish the closed number when it's closed.
Mistakes made and named: two configuration breaks in month one, one memory-drift break in month two, one near-miss with offer language slipping through Content Multiplier. Total cost: zero customer-facing damage, four pieces of unsent draft killed at the queue, one invoice cycle of stale line-item language. All four caught at the approval surface, which is the design point of the Cockpit. Every Pack writes through an approval queue the Boss commands.
What didn't work: we tested whether Trust Pack could behave like a lightweight recipe in week 4 and stopped after two days. The Pack is twelve component Packs wired as a unified OS, and that's not something we'd recommend treating as a quick install while the Boss is running four other Packs. The right path for Trust Pack is DFY with a real install plan. We'll document that walkthrough separately.
What we'd do differently
Three things, in order of how much they would have saved us.
Shared-memory document on day one. Biggest lesson from the 90 days. Run more than two Packs, you need a single source of truth for offer language, voice samples, and standing orders. Updating one document instead of n Packs prevents the entire memory-drift failure mode. Setup: under an hour. Cost of not doing it: one near-miss outbound send and four days of stale invoice language.
Lock the approval queue to one surface before any outbound runs. Slack-channel-level lock is fine. Multiple-surface drafts ship mistakes.
Tune Outbound Engine in shadow mode for ten days minimum before approving sends. The Pack docs suggest seven. We learned the back half of those ten days is where the voice library starts producing drafts that sound like the Boss and not a generic SDR.
If you're starting from zero, the install order we'd recommend, based on what bled the most for us: Day One Operator first because it stops the operator from being the bottleneck, Lead Rescue second because it surfaces dormant pipeline you forgot you had, Get-Paid Engine third because it pays for itself fastest. Then a 30-day pause before you install anything else. Pause is not optional. Tuning is the work. Installs are easy.
We'll publish the 180-day update in late June with audited numbers. If those numbers look worse than what we projected here, we'll tell you that too. That's the deal.
The harness ran. The Packs ran. The Boss told the Co-pilot where to go. None of this required the Boss to know how to fly. It required the Boss to know how to direct the Co-pilot and set the standing order, and to remember to update the memory when the inputs changed.
That's the whole job.
Key takeaways
- 01Founder self-review. Honest about month 1, 2, 3. Specific Pack-by-Pack receipts.
- 02> This is the founder reviewing the founder's own use of BossMode.
- 03It is not a customer review.
Take the Bottleneck Check.
Sixty minutes. We map the bleed and name the Packs that stop it. Without trust, you're a bust.
Take the Bottleneck CheckRead next
Keep moving through the system
Why pricing transparency is the number one trust signal for AI-era buyers
Sheridan-aligned manifesto. Cites the AI Trust Signals thesis. Pricing transparency is now an LLM-weighted ranking factor.
11 min read
What BossMode is bad at
Honest disclosure. Three things BossMode genuinely doesn't do well, and what to use instead.
9 min read