Incident Communication

How to Write a Post-Mortem That Actually Improves Your Process

K

Kirk Makse

Founder 7 min read
How to Write a Post-Mortem That Actually Improves Your Process

The incident is resolved. Systems are green. Customers are calm.

Now what?

For most teams, the answer is “move on.” Ship the next feature. Close the ticket. Pretend it didn’t happen.

That’s a missed opportunity. The teams that get better at handling incidents are the ones that take the time to write a post-mortem — not as a formality, but as a genuine tool for learning.

Here’s how to write a post-mortem that actually improves your process.

What a Post-Mortem Is (and Isn’t)

A post-mortem is a structured review of an incident after it’s been resolved. Its purpose is simple: understand what happened, why it happened, and how to prevent it from happening again.

It’s not a blame report. It’s not a disciplinary document. It’s not a way to prove someone messed up.

The best post-mortems are learning documents. They capture context that would otherwise be lost and turn painful incidents into durable improvements.

Why Blameless Post-Mortems Produce Better Outcomes

When people feel blamed, they hide information.

They downplay their involvement. They leave out details that might make them look bad. They optimise for self-preservation instead of truth.

A blameless post-mortem flips this dynamic. When there’s no fear of punishment, engineers share openly:

  • “I deployed without checking the migration”
  • “I saw the alert but assumed it was a false positive”
  • “I didn’t know that service had a hard dependency on the cache”

These honest details are exactly what you need to find the real root cause — and they only surface when people feel safe.

Blameless doesn’t mean accountability-free. It means you focus on fixing the system, not punishing the person. If one engineer can bring down production with a single command, that’s a system problem, not a people problem.

The Anatomy of a Great Post-Mortem

Every effective post-mortem covers the same ground. You don’t need a complex framework — just a clear, consistent structure.

Incident Summary

Start with the basics. What happened, when, and how severe was it?

  • Date and duration — When did it start? When was it resolved?
  • Severity — How many customers were affected? What was the business impact?
  • Components affected — Which parts of the product were impacted?

Keep this section short. Two to three sentences plus a few data points. The goal is orientation, not detail.

Timeline

The timeline is the backbone of the post-mortem. It should read like a chronological story:

  • 10:14 UTC — Monitoring alert fires for elevated API error rates
  • 10:18 UTC — On-call engineer acknowledges alert, begins investigation
  • 10:25 UTC — Root cause identified: database connection pool exhausted after deploy
  • 10:32 UTC — Fix deployed, connection pool configuration updated
  • 10:45 UTC — Error rates return to normal, incident marked resolved

Include timestamps, who did what, and key decisions made along the way. This is where the learning happens — you can see gaps in detection, delays in response, and communication breakdowns.

Root Cause

Go deeper than the surface. “The deploy broke production” isn’t a root cause — it’s a symptom.

Ask why repeatedly:

  • Why did the deploy break production? → The new code opened too many database connections
  • Why did it open too many connections? → The connection pool size wasn’t configured for the new query pattern
  • Why wasn’t this caught before deploy? → There’s no load testing step in the deployment pipeline

The real root cause is almost never “someone made a mistake.” It’s usually a missing safeguard, a blind spot in monitoring, or a process gap.

Customer Impact

Be specific about what customers experienced:

  • How many users were affected?
  • What did the failure look like from their perspective?
  • How long were they impacted?
  • Were they notified? How quickly?

This section connects the technical incident to the business reality. It’s also where you evaluate how well your incident communication worked — did customers know what was happening, or were they left in the dark?

Action Items

This is the most important section. Without action items, a post-mortem is just a story.

Good action items are:

  • Specific — “Add connection pool size to deployment checklist” not “improve deployment process”
  • Assigned — Every item has an owner
  • Time-bound — Every item has a deadline
  • Prioritised — Not everything needs to happen this week, but the highest-risk items should

Aim for 3–7 action items. If you have more than 10, you’re probably trying to fix too much at once.

A Ready-to-Use Post-Mortem Template

Here’s a template you can copy and adapt for your team:

Incident Post-Mortem: [Incident Title]

Date: [Date of incident] Duration: [Start time] – [End time] ([total duration]) Severity: [Critical / Major / Minor] Author: [Name]

Summary [2-3 sentences describing what happened and the customer impact]

Timeline

  • [HH:MM UTC] — [Event]
  • [HH:MM UTC] — [Event]
  • [HH:MM UTC] — [Event]

Root Cause [Detailed explanation of the underlying cause]

Customer Impact [Number of affected users, what they experienced, duration]

What Went Well

  • [Things that worked during the response]

What Could Be Improved

  • [Gaps in detection, response, or communication]

Action Items

  • [Action] — Owner: [Name] — Due: [Date]
  • [Action] — Owner: [Name] — Due: [Date]
  • [Action] — Owner: [Name] — Due: [Date]

Adapt it to your team’s needs, but resist the urge to overcomplicate it. A simple, consistently used template beats a detailed one that nobody fills out.

How to Share Post-Mortems to Build Trust

Most teams keep post-mortems internal. That’s a safe default — but sharing them publicly can be a powerful trust-building move.

When customers see a public post-mortem, they learn:

  • You take incidents seriously
  • You understand what went wrong
  • You have a plan to prevent it from happening again
  • You’re transparent enough to share the details

You don’t need to share every technical detail. A customer-facing post-mortem can be a simplified version:

  • What happened (in plain language)
  • How long it lasted
  • What you’re doing to prevent it
  • An apology

Your status page is a natural place to link public post-mortems. Customers who subscribe to your status page are already engaged — they’ll appreciate the follow-through.

Common Post-Mortem Mistakes

Writing it weeks later. Details fade fast. Aim to write the post-mortem within 48 hours of the incident while context is fresh.

Skipping the meeting. A written document is good. A 30-minute discussion where the team walks through it together is better. Questions surface details that the author missed.

Vague action items. “Improve monitoring” is not an action item. “Add alerting for connection pool usage above 80% by Friday” is.

Never following up. The post-mortem is only as valuable as the action items that get completed. Review open items in your next team sync.

Making it punitive. The moment people feel blamed, they stop sharing. Protect the blameless culture — it’s what makes the whole process work.

How CheckStatus Helps

CheckStatus gives you the foundation for effective post-incident communication:

  • Incident timelines that document exactly what happened and when
  • Component-level impact tracking so you know which services were affected
  • Chronological updates that serve as a starting point for your post-mortem timeline
  • A public status page where you can link to customer-facing post-mortem summaries

Your incident history in CheckStatus becomes a natural audit trail — making it easier to write accurate post-mortems and track patterns over time. See all CheckStatus features or learn how it works.

Final Thought

Incidents are inevitable.

Repeating the same incident is not.

A good post-mortem turns a painful experience into a lasting improvement. It builds trust within your team, strengthens your processes, and shows your customers that you take reliability seriously.

The best time to write a post-mortem is right after the incident. The second best time is now.

Create your status page in less than 5 minutes. No credit card required.

K

Kirk Makse

Founder of CheckStatus. Building tools to help SaaS teams communicate better during incidents.

Ready to get started?

Try CheckStatus free for 30 days. No credit card required.

Join teams who value transparency