DIY Outbound Audit
The $50K Outbound Audit You Can Run Yourself in 3 Hours
By Peter Korpak · 18 min read · April 2026
Direct answer: This is the same six-dimension audit we sell as the Outbound Autopsy ($697), written out so you can run it yourself in three hours. You'll need terminal access for dig, accounts at MXToolbox and NeverBounce, and the discipline to score yourself honestly — most teams overrate themselves by 20–30 points. Score each of the 18 checks on a 0–3 scale. Total the six dimensions (two dimensions weighted 1.2x), and compare against the ranges below. At the end, you'll know exactly which fixes move the most pipeline. If three hours sounds longer than you have — which is the honest reason most people pay for this — the Autopsy is at the bottom of this page.
We wrote this because gatekeeping a framework that teams need is not a business strategy. It is a fear strategy. The audit itself is not the value — the implementation is. If you run this, find your problems, and fix them yourself, good. If you run this, find your problems, and realize you don't have three hours to go deeper, we're here.
Either way, the framework is yours.
Why Give This Away
Trust compounds faster than gatekeeping. Teams that find their problems through a framework remember where the framework came from. When they need help implementing — or when they need someone else to do it — they come back to the people who helped them see it clearly.
There's also a more practical reason. The people who need an outbound audit most — the teams spending $60,000 on AI SDRs with a broken infrastructure layer, or sending 200 emails per mailbox per day wondering why their reply rates are collapsing — need to see the problem concretely before they'll invest in fixing it. (For context on how widespread the broken-infrastructure problem was with AI SDR buyers specifically, see The AI SDR Post-Mortem.) The DIY audit does that. It makes the invisible visible.
Run this audit. Score yourself honestly. Then decide what to do about it.
How to Score Yourself
Six dimensions. Three checks per dimension. 18 total checks. Each check is scored 0–3:
- 0 — Not done, not set up, or actively harmful
- 1 — Minimal / partial / inconsistent
- 2 — Solid but not optimized
- 3 — Best practice, fully implemented
Two dimensions are weighted 1.2x because infrastructure failures are invisible (your email never reaches anyone if authentication fails) and signal saturation is the fastest-moving variable in 2026. Those two are Infrastructure and Timing and Signals.
| Dimension | Weight | Max points | Time |
|---|---|---|---|
| 1. Sender Infrastructure | 1.2× | 10.8 | 30 min |
| 2. List Quality | 1.0× | 9.0 | 30 min |
| 3. Messaging Quality | 1.0× | 9.0 | 30 min |
| 4. Timing and Signals | 1.2× | 10.8 | 30 min |
| 5. Channel Strategy | 0.9× | 8.1 | 30 min |
| 6. Measurement Maturity | 0.7× | 6.3 | 30 min |
Maximum total: 54 points. Score ranges at the end of this page.
Dimension 1 — Sender Infrastructure (30 min)
Infrastructure is the gate. Every other dimension depends on it. If your email never reaches the inbox, nothing else matters.
Check 1.1: Domain configuration
Are you sending cold outreach from your primary company domain? If yes, score 0. Your cold email program is putting your entire company's email reputation at risk. A deliverability problem on your cold outreach domain will eventually affect your support emails, your invoices, and your communications with existing customers.
The correct configuration: a separate domain (or subdomain) dedicated to outreach, registered at least 4–6 weeks before first send, with a different registrar than your primary domain. Score: 0 (primary domain), 1 (separate but recently registered), 2 (separate, 4+ weeks old), 3 (separate, 6+ weeks, dedicated to cold outreach only).
Check 1.2: Authentication (SPF + DKIM + DMARC)
Run these commands in your terminal against your sending domain:
dig TXT yourdomain.com +short | grep spf
dig TXT mail._domainkey.yourdomain.com +short
dig TXT _dmarc.yourdomain.com +short
The first should return an SPF record (a line starting with v=spf1). The second should return your DKIM public key. The third should return your DMARC policy.
Gotcha: having all three records is not the same as having them configured correctly. Check that your DMARC policy is p=quarantine or p=reject, not just p=none. Check that the domain in your DKIM d= tag matches your From: header domain. Google's enforcement error code 4.7.32 fires specifically when these domains don't align — even if both records exist.
Score: 0 (none configured), 1 (SPF only), 2 (SPF + DKIM, no DMARC or p=none), 3 (SPF + DKIM + DMARC at p=quarantine or p=reject, properly aligned).
Check 1.3: Per-mailbox send volume
How many cold emails are you sending per mailbox per day? Check your sending platform's per-account sending stats for the last 30 days.
The practical limit for cold outreach in 2026 is 30–50 emails per day per mailbox on Google Workspace and 30–40 on Microsoft 365. These are not hard walls — they are soft reputation cliffs. A mailbox sending 150/day for three weeks accumulates reputation damage before any bounce code appears.
Gotcha: "under 50" counts per mailbox, not per domain. If you have 3 mailboxes on one domain sending 50 each, that is 150 per domain. Domain-level volume matters to Google's bulk-sender classification as well.
Score: 0 (200+ per mailbox), 1 (100–200), 2 (50–100), 3 (under 50, with warmup completed).
Dimension 1 subtotal: _____ × 1.2 = _____
Dimension 2 — List Quality (30 min)
A clean list is not a nice-to-have. A 10% bounce rate actively damages sender reputation. ESPs throttle or block senders above 2% hard bounces. Most B2B email lists decay at 22.5–30% per year — a list not cleaned in 12 months has a predictable problem built in.
Check 2.1: ICP definition and fit scoring
Pull the last 50 meetings booked from cold outbound. What percentage of those prospects match your ICP on at least 3 of 5 criteria (industry, company size, seniority, tech stack, geography)?
Gotcha: most teams define their ICP broadly enough that nearly everything qualifies. The test is whether you would take a meeting with anyone on your current list right now. If the honest answer is "well, most of them," the ICP definition is too wide and the list is generating activity without generating pipeline.
Score: 0 (no defined ICP, mass prospecting), 1 (ICP defined but loosely applied), 2 (ICP applied consistently, 60–70% fit on recent meetings), 3 (ICP tight, 80%+ fit on recent meetings, regular list review).
Check 2.2: Email verification before sending
Are you running list verification before each campaign? Take a sample of 500 addresses from your current active list and run them through NeverBounce or ZeroBounce. The result will show valid, invalid, catch-all, and unknown.
Gotcha: catch-all domains accept all incoming mail at the server level — which means NeverBounce shows them as "accept-all" rather than invalid. They look clean. In practice, many catch-all domains forward all mail to a spam folder or report it without delivering. Including a high percentage of catch-all domains in your cold sequences silently damages your sender reputation without generating bounce codes. Filter them to a separate, lower-frequency sequence or remove them entirely.
Score: 0 (no verification, or haven't verified in 6+ months), 1 (verified once at list build, not ongoing), 2 (verify each campaign, but include catch-alls), 3 (verify each campaign, filter catch-alls, remove invalids same day).
Check 2.3: Bounce rate monitoring
Check your sending platform for bounce rate on campaigns in the last 60 days. Hard bounce rate (addresses that don't exist) should be under 2%. If you don't know your current bounce rate, that is a 0.
Score: 0 (don't track, or over 10%), 1 (5–10%), 2 (2–5%), 3 (under 2%, monitored weekly).
Dimension 2 subtotal: _____ × 1.0 = _____
Dimension 3 — Messaging Quality (30 min)
This is the dimension teams overrate most consistently. Read your actual sequence copy as if you are the recipient, not the sender. You have never heard of this company. You did not request this email. You have 8 seconds.
Check 3.1: Subject line quality
Count the words in your current subject lines. Pull the subject lines from your three most recent campaigns.
The benchmark: 4–7 words. Subject lines in this range consistently outperform longer ones in cold outreach datasets. Under 4 words risks being cryptic. Over 7 words gets truncated on mobile.
Gotcha: AI-generated personalized first lines have become a pattern that spam filters recognize. A subject line that looks like "Quick note for [Company]'s team" or "Loved your [LinkedIn post topic]" may feel personalized but now matches the behavioral fingerprint of high-volume AI-assisted outreach. Specificity works; generic-sounding personalization tokens do not.
Run your subject lines through the free subject line tester for a scored analysis.
Score: 0 (no subject line consistency, or all over 10 words), 1 (7–10 words, some variation), 2 (4–7 words, moderate specificity), 3 (4–7 words, trigger-based or highly specific to a company event).
Check 3.2: Personalization tier
What is the personalization tier of your email body? Token personalization (first name, company name) is table stakes — it lifts open rates by roughly 9–22% but is now so common it is no longer a differentiator. The jump happens at trigger-based personalization: referencing a specific event (funding announcement, job change, product launch, recent hire) that is observable and verifiable.
Read your last 10 sent emails. How many reference a specific, verifiable event at the prospect's company?
Score: 0 (same template for everyone, first-name token only), 1 (industry or role segmentation), 2 (company-specific first line, no trigger), 3 (trigger-event reference — funding, hiring, launch, news item).
Check 3.3: Email length
Count the words in your standard email body (excluding signature). Not your ideal email — your actual median send.
The benchmark: 50–100 words. Emails over 125 words see meaningfully lower reply rates in cold outreach datasets. The constraint is not arbitrary — a stranger's inbox is not the place for a long pitch. Short emails require the recipient to invest less to read them. Less investment means more reads, more replies.
Score: 0 (200+ words), 1 (125–200 words), 2 (100–125 words), 3 (50–100 words with a clear single ask).
Dimension 3 subtotal: _____ × 1.0 = _____
Dimension 4 — Timing and Signals (30 min)
At any given time, roughly 3–5% of your total addressable market is actively in-market for what you sell. Without signals, you are emailing the 95% who are not ready and hoping to catch the 5% by accident.
Check 4.1: Signal library
List every trigger your team uses to time outreach. Count how many are in active use (firing sequences today, not just configured in theory).
Common signals: funding announcement, job change (relevant seniority), technology adoption (new tool in stack), hiring surge (adding headcount in a relevant function), product launch or press coverage.
Gotcha: funding round signals are the most saturated trigger in B2B sales right now. A funded prospect's inbox receives roughly 47 identical congratulations emails within 48 hours of the announcement (It's Just Revenue, February 2026). If funding is your only signal, you are in the most crowded lane. Diversify to signals with lower adoption (hiring pattern, technology change, press coverage outside major tech media).
Score: 0 (no signals, time-based sequences only), 1 (one signal type, usually funding), 2 (2–3 signal types with different saturation levels), 3 (4+ signals including at least one low-adoption trigger, reviewed monthly for saturation).
Check 4.2: Sequence structure
How many touches are in your standard cold sequence, and what is the spacing between them?
The benchmark: 4–7 touches, with 3–5 day spacing. First follow-up adds approximately 49% more replies. After the 4th follow-up, diminishing returns are sharp and spam complaint rates rise. Sequences with 8+ touches show meaningfully higher opt-out rates.
Score: 0 (1 email, no follow-up), 1 (2–3 emails, spacing under 2 days), 2 (4–7 emails, but spacing is irregular), 3 (4–7 emails, 3–5 day spacing, last touch is a break-up email).
Check 4.3: Signal freshness
What is the median time between a trigger firing and the first outreach touch? Check your CRM or sequencing platform for time-to-first-touch after enrollment.
For funding signals: the window is 48 hours. After 30 days, the signal is effectively cold. For job changes: 14–30 days is the high-intent window. For technology adoption: the window is wider (30–90 days).
Score: 0 (signals are reviewed weekly or less, often stale), 1 (reviewed a few times per week, 3–5 day lag), 2 (daily review, 24–48 hour lag), 3 (real-time or near-real-time enrollment, first touch within signal window).
Dimension 4 subtotal: _____ × 1.2 = _____
Dimension 5 — Channel Strategy (30 min)
Email-only outbound is increasingly filtered. The teams achieving 5–12% response rates in 2026 are not using better email subject lines. They are making contact through multiple channels before the cold email lands — which means the cold email is no longer arriving from a stranger.
Check 5.1: Multi-channel coordination
Does your outreach sequence coordinate across at least two channels? Does your platform support LinkedIn step insertion (connection request, reaction, comment) as part of the same sequence?
The pattern that works: LinkedIn visibility before the email. Engage with a prospect's content, connect, or comment on their posts before the first email touch. When the email arrives, the name is familiar. Unfamiliar-name emails are deleted or ignored. Familiar-name emails are read.
Score: 0 (email only), 1 (email + LinkedIn connection requests, no pre-email engagement), 2 (email + LinkedIn with some pre-email touch), 3 (systematic pre-email LinkedIn engagement before first email touch, tracked per prospect).
Check 5.2: LinkedIn presence
Does the person leading outreach (founder, sales lead, or rep sending the emails) post on LinkedIn at least weekly? Does the content address problems your ICP cares about?
This is the recognition engine. Prospects who see your name in their feed multiple times before your email arrives are not treating your email as cold outreach. They are responding to someone they have seen around. The reply rate difference is significant — teams with active LinkedIn presence from the sender report 5–12× higher cold email response rates versus sending from a presence-free profile.
Score: 0 (no LinkedIn posting), 1 (occasional posts, no consistent cadence), 2 (weekly posts, relevant to ICP), 3 (2+ posts per week, specific to ICP problems, sender is recognizable in the prospect's feed).
Check 5.3: Before-email warmup
For your highest-priority accounts, is there a systematic pre-email touchpoint before the first cold email? This could be a LinkedIn comment on their content, a reaction to a company post, a mention of their work in your own content, or an introduction through a mutual connection.
Score: 0 (cold email is always the first contact), 1 (LinkedIn connection sent before email, no other engagement), 2 (LinkedIn engagement before email on some accounts), 3 (systematic pre-email engagement documented for all priority accounts before first touch).
Dimension 5 subtotal: _____ × 0.9 = _____
Dimension 6 — Measurement Maturity (30 min)
Open rate is broken. Apple Mail Privacy Protection inflates opens by 15–25% across iOS inboxes. If open rate is your primary metric, you are measuring a number that partly reflects iOS pre-fetching, not human engagement. Positive reply rate is the metric. Everything else supports it.
Check 6.1: Funnel tracking
Can you report, today, the following metrics from your cold outbound program for the last 90 days: emails sent → replies → positive replies → meetings booked → opportunities created → deals closed?
If you can report only emails sent and open rate, score 0. If you can report replies and meetings booked but not further, score 1. Full funnel tracking from email to closed deal is the standard required to make investment decisions about outbound.
Score: 0 (track emails sent and open rate only), 1 (track reply rate and meetings booked), 2 (track full funnel to opportunities), 3 (track full funnel through close, with attribution per sequence).
Check 6.2: Cost per meeting
Do you know your current cost per meeting booked from cold outbound? The formula: (platform costs + list costs + operator time in hours × hourly rate) ÷ meetings booked in the same period.
Most teams that have not run this calculation are surprised by the result. Cold outbound costs more per meeting than it appears to from the tool subscription cost alone. When you factor in operator time (list building, sequence management, reply handling), the all-in cost per meeting is often 3–5× the visible platform cost.
Gotcha: Apple MPP inflates opens 15–25%, so any metric derived from opens (open-to-reply conversion, click-through from opener) is partially inflated. Use positive reply count as your conversion denominator, not opens.
Score: 0 (don't know), 1 (rough estimate, not calculated), 2 (calculated monthly, platform costs included), 3 (calculated monthly, all costs included — platform, list, operator time).
Check 6.3: CAC:ACV ratio
Divide your cost to acquire a customer from cold outbound by your average contract value (ACV). This is your CAC:ACV ratio.
Under 0.33x is healthy — you are spending less than one-third of a year's contract value to close the contract. Above 0.5x means cold email is likely unprofitable relative to deal size. Above 1.0x means you are spending more to acquire customers than they are worth in the first year.
Use the free ROI calculator to run this math against your actual numbers.
Score: 0 (no CAC tracking, or ratio above 1.0x), 1 (rough CAC, ratio 0.5–1.0x), 2 (calculated CAC, ratio 0.33–0.5x), 3 (calculated CAC compared monthly to ACV, ratio under 0.33x).
Dimension 6 subtotal: _____ × 0.7 = _____
How to Read Your Score
Add your six weighted dimension subtotals. The maximum is 54.
| Score range | Rating | What it means |
|---|---|---|
| Below 27 | Critical | One or more dimensions has a fundamental failure. Stop sending. Fix infrastructure first. |
| 27–40 | Leaky | Basics are in place but structural gaps are costing you 40–60% of potential pipeline. Fix highest-weighted dimensions first. |
| 41–48 | Solid | Well-configured system. Marginal improvements available. Look for the specific checks where you scored 0 or 1. |
| 49+ | Competitive | Top-quartile outbound operation. The constraint is probably recognition and ICP quality, not mechanics. |
Note: a high total score with a 0 on any single check is a false positive. A 0 on Infrastructure Check 1.2 (authentication) means your emails may not be reaching recipients at all — regardless of what every other dimension shows.
The Three Fixes That Move the Most Pipeline
If you scored below 48 and need to prioritize, the three fixes with the highest pipeline impact across every team we have audited are:
Fix 1: Authentication alignment (Infrastructure)
If your DMARC is at p=none or missing entirely, this is the first fix. Configure DMARC at p=quarantine on your sending domain, verify that your DKIM signing domain aligns with your From: header domain, and set up Google Postmaster Tools for daily reputation monitoring. This is a one-time configuration change with permanent benefit. Most teams can complete it in under two hours with their DNS provider's documentation.
Fix 2: Diversify your signal library (Timing and Signals)
If your only trigger is funding announcements, you are in the most saturated signal lane in B2B sales. Add two low-adoption signals: technology adoption changes (newly added tool to stack detectable via BuiltWith or Wappalyzer), and press coverage in niche publications outside major tech media. These signals have lower adoption because they require more work to monitor. That is why they work better.
Fix 3: Pre-email LinkedIn engagement (Channel Strategy)
Before your first email touch to any priority account, spend 2 minutes engaging with something the prospect has publicly posted on LinkedIn — a thoughtful comment on their content, a reaction, a mention of their work in your own post. This single change consistently moves cold email response rates from 2–4% to 5–12% among teams that implement it systematically. It is not magic. It is the difference between landing in a stranger's inbox and landing in the inbox of someone who has seen your name.
Or Pay $697 to Skip the 3 Hours
The self-audit gives you a directional score and a prioritized fix list. It does not tell you what your actual domain reputation looks like in Google Postmaster Tools. It does not pull your real bounce rate from a live sample. It does not read your actual sequence copy as an outsider who has never heard of your company.
The Outbound Autopsy does those things. We pull your sending domain reputation directly. We run NeverBounce against a sample of your active list. We read your sequences with fresh eyes — not your description of them, but the actual emails. We deliver a specific fix stack ranked by pipeline impact, a full infrastructure audit with remediation steps, and a messaging review with rewrite recommendations. In 5 business days.
The honest reason most teams pay for this instead of running it themselves: three hours of senior operator time costs more than $697 at most companies. The ROI math on the Autopsy works out to approximately $1 in investment for every $4–8 in recovered pipeline, based on the fixes most commonly surfaced.
If you ran the audit above and found problems — and most teams do — the next question is whether you want to fix them yourself or hand it off.
Get Monthly Field Notes
One email per month. What's working in outreach, what's not, and why. No fluff, no funnel.
No webinars. No launch countdowns. Unsubscribe anytime.
Frequently Asked Questions
Why give this away free?
Two reasons. First: trust compounds faster than gatekeeping. Teams that run this audit and find real problems remember where they got the framework. Second: the people who need it most — the ones spending $60K on AI SDRs with a broken infrastructure layer — need to see the problem clearly before they'll invest in fixing it properly. The audit is the diagnostic. The Autopsy is the prescription.
How accurate is self-scoring?
Self-scoring on infrastructure is reliable — either your DMARC record exists or it doesn't, either your SPF passes or it fails. Self-scoring on messaging and measurement tends to run 15–30 points high because it's easy to convince yourself your personalization is better than it reads to a stranger. The audit is most useful when you score what you actually do, not what you intend to do.
Which dimension matters most?
Infrastructure and Timing and Signals are weighted 1.2x in our scoring because infrastructure failures are invisible (your copy never reaches the inbox if authentication fails) and signal fatigue is the fastest-moving risk in 2026. But if you score below 1 out of 3 on any single dimension, that dimension is your ceiling regardless of how well you score on the others.
What do I do if my score is below 27?
Stop sending and fix infrastructure first. A score below 27 almost always has a critical infrastructure problem — your primary domain is your sending domain, or DMARC is missing, or per-mailbox volume is too high. Every email you send on broken infrastructure compounds the reputation damage. Fix the foundation before optimizing anything else.
What's actually different about the paid Autopsy?
Three things: time, an outsider's eyes, and depth the self-audit can't reach. Time: this takes you three hours; the Autopsy takes us 5 business days. Outsider's eyes: you cannot see your own blind spots. Depth: the Autopsy pulls your actual sending domain reputation from Postmaster Tools, runs NeverBounce against a sample of your list, reads your actual sequence copy — not your description of it. The self-audit gives you a directional score. The Autopsy gives you a specific fix stack ranked by pipeline impact.
3 Hours of Your Time. Or 5 Business Days of Ours.
The Outbound Autopsy covers all six dimensions with your actual data — live domain reputation, real bounce rate, fresh eyes on your copy. Delivered in 5 business days.