How to Know If That $50K Training Was Worth It

This before/after analysis in Excel helps you confirm what’s working—before spending more.

Good Morning, It’s Wednesday, March 26.

  • Topic: Before/After Analysis | Paired Sample T-Test | Excel Tutorial

  • For: B2B and B2C Managers.

  • Subject: Statistics → Practical Application

    • Concept: Paired hypothesis testing

    • Application: Using paired t-test in Excel to validate business investments

Don’t keep us a secret. Share this newsletter with friends (copy URL here).

Introduction

We've all been there—you invest in a new business initiative and believe it will work.

Your company spends significant resources on employee training, a new sales process, or customer experience improvements.

But early results are confusing. Some metrics improve while others show no change. You can't tell if your investment is working or if you're just seeing normal business fluctuations.

Most businesses lack the statistical expertise to validate their results properly. This leads to continuing investments in initiatives that don't work—or abandoning programs that are quietly delivering tremendous value.

Paired t-tests solve this problem. Unlike regular t-tests (which compare different groups), paired t-tests measure changes in the same people or units before and after an intervention—revealing whether your initiative truly made a difference, with statistical confidence.

Real-World Example

Imagine you're the VP of Sales for a B2B company. You've invested $50,000 in a new sales training program for your team of 24 sales representatives.

You want to know: Did this expensive training actually improve performance?

Step 1: Define the goal.

Your goal is to determine if the training program significantly increased sales performance. You're measuring monthly sales figures per rep, before and after training.

Step 2: Collect data.

You track each sales rep's performance for the month before training and the month after training.

Initial results look promising:

  • Average monthly sales before training: $42,500 per rep

  • Average monthly sales after training: $45,700 per rep

That's a $3,200 increase per rep. With 24 reps, that's potentially $76,800 in additional monthly revenue.

But is this improvement real, or just random variation? Sales figures naturally fluctuate month to month, even without training.

Step 3: Perform a paired t-test.

A paired t-test is perfect here because it examines each rep's individual change in performance, accounting for the fact that some reps naturally sell more than others.

The paired approach asks: "Did each rep improve relative to their own baseline?"

Step 4: Interpret the results.

The paired t-test gives you a p-value of 0.08above the standard threshold of 0.05. This means the improvement might just be random variation, not a real effect of the training program.

This insight saves your company from mistakenly attributing a $76,800 monthly revenue increase to the training. Without the paired t-test, you might have:

  • Incorrectly praised the training vendor and renewed an expensive contract.

  • Rolled out the same training to other departments, wasting more money.

  • Missed an opportunity to find truly effective training.

Instead, you can now investigate other factors that might actually boost sales performance or test a different training approach—saving potentially hundreds of thousands in misspent training dollars.

How to Run a T-test in Excel

A retail chain launched a new customer service approach in 15 stores. They measured customer satisfaction scores (out of 10) before and after implementing the new approach.

The company wants to know: Did the new approach truly improve customer satisfaction?

Here's the data they collected:

  • Before implementation: Average satisfaction score of 7.2 (std dev: 0.3)

  • After implementation: Average satisfaction score of 7.9 (std dev: 0.7)

The 0.7-point improvement looks good, but is it a real change or just luck?

Step 1: Goal. Is the improvement in satisfaction scores real or just random variation?

Step 2: Data. In Excel, enter the "Before" scores in column A and the "After" scores in column B, with each row representing the same store.

Step 3: Paired t-test. 

  • Go to Data > Data Analysis > t-Test: Paired Two Sample for Means.

  • Select “Before” scores as Variable 1 and “After” scores as Variable 2.

  • Set Alpha = 0.05 (this means we’re testing at a 95% confidence level).

Step 4: Interpret the results. This is a two-tailed test.

  • The p-value = 0.0004 → Since this is less than 0.05, the improvement is statistically significant.

  • This means that the new customer service approach genuinely improved satisfaction scores, and this wasn't just due to random fluctuation.

Pro Tip: To be even more precise, we calculate the Confidence Interval (CI).

  1. Standard Error (SE): 0.16

  2. Margin of Error (ME) = SE × t Critical (2.144) → 0.34

  3. Final CI = 0.7 ± 0.34 → [0.36, 1.04]

This tells us that the true improvement in customer satisfaction is between 0.36 and 1.04 points—a meaningful increase for satisfaction metrics.

Conclusion:

With this data, the company can confidently invest in the new customer service approach, knowing it will improve satisfaction scores by at least 0.36 points—and possibly up to 1.04 points. 

This improvement could translate to higher customer retention, increased repeat business, and enhanced brand reputation worth millions in long-term revenue.

Limitations

  • Timing matters – When you measure can skew results. Too soon misses delayed effects; too late captures unrelated factors.

  • Requires same samples – You must measure the exact same entities before and after. Any turnover or changes compromise your data.

  • Can't prove causation – Shows if change occurred, but not whether your intervention caused it. External factors could be responsible.

  • Sample size critical – Fewer than 15-20 pairs reduces reliability. Smaller samples make it harder to detect real effects.

  • Assumes normal distribution – Doesn't work well with highly skewed data or significant outliers.

Where Else Can You Use This?

  • Product Improvements – Test if specific updates truly enhance user satisfaction or engagement by measuring the same users before and after changes.

  • Price Optimization – Measure if price adjustments affect purchase behavior by comparing the same customers' buying patterns before and after.

  • Employee Performance Programs – Determine if coaching or training initiatives actually boost productivity by tracking individual performance metrics.

  • Process Optimization – Quantify if efficiency changes reduce errors or time spent by comparing the same workflows before and after implementation.

  • Store Layout Changes – Validate if physical space modifications increase sales by comparing performance in the same locations before and after renovations.

Want to go beyond today’s breakdown? Here are the best resources to master this topic:

  • Gies College of Business – Inferential and Predictive Statistics for Business. Link here.

  • Investopedia – T-Test: What It Is With Multiple Formulas and When To Use Them. Link here.

How did you like today's newsletter?

Login or Subscribe to participate in polls.