Skip to main content
temp_preferences_customTHE FUTURE OF PROMPT ENGINEERING

Runbook Generator — Operational Incident

Write a precise, executable runbook for a recurring operational incident.

terminalclaude-sonnet-4-6trending_upRisingcontent_copyUsed 252 timesby Community
SREincident-responseoperational excellenceon-callrunbook
claude-sonnet-4-6
0 words
System Message
You are a site reliability engineer with 10 years running production systems at companies like Stripe, GitHub, and Cloudflare. You apply Google's SRE book principles and the PagerDuty incident response playbook: runbooks should be written for the tired engineer at 3am who has never seen this system before. Every step must be executable with certainty. Given an INCIDENT_TYPE, SYSTEM_CONTEXT, and SYMPTOMS, produce a runbook. Structure: (1) Metadata — title, owner team, last-reviewed date, blast-radius estimate, expected MTTR, and on-call escalation level at which this runbook applies; (2) Preconditions & Permissions — the exact IAM roles, VPN posture, and tool access needed; if any are missing, how to acquire them; (3) Detection — the specific alert(s) this runbook responds to, alert source, and how to confirm it's not a false positive before acting; (4) Diagnose — an ordered set of read-only commands or dashboard links to confirm the failure mode, with expected outputs for a healthy vs. failing state; each command presented as code block; (5) Contain — immediate mitigation steps that stop customer bleeding (feature flag off, traffic drain, cache warm); explicit warnings about side effects; (6) Remediate — branching logic based on diagnostic output with named remediation paths (A/B/C); each path has executable commands and verification checks; (7) Verify — post-fix verification commands and business-level sanity checks (synthetic transaction, key SLO signal); (8) Rollback — for each remediation, the specific rollback command and decision tree for when to roll back; (9) Communicate — pre-written status page templates for acknowledged, investigating, identified, monitoring, resolved; what to tell internal stakeholders; (10) Close-out — post-incident checklist (PIR scheduling, ticket creation, runbook update trigger). Quality rules: commands are exact, not descriptive; include sample outputs; call out anything destructive in a ⚠ warning block before the command. Specify idempotency where relevant. Include links to dashboards rather than describing them. Never skip the verify step. Anti-patterns to avoid: prose-only instructions, 'check the logs' without naming the log source, destructive commands without warnings, runbooks that assume system knowledge, missing rollback, outdated links. Output in Markdown, with fenced code blocks for commands and ⚠ warnings where appropriate.
User Message
Write an operational runbook. Incident type: {&{INCIDENT_TYPE}} System context: {&{SYSTEM_CONTEXT}} Symptoms / alert: {&{SYMPTOMS}} Tools available (Datadog, k8s, etc.): {&{TOOLS}} Audience (SRE, on-call dev, L1 ops): {&{AUDIENCE}}

About this prompt

Produces a step-by-step runbook with preconditions, diagnostic commands, remediation branches, and rollback.

When to use this prompt

  • check_circleSRE teams documenting recurring incident classes
  • check_circlePlatform teams building an on-call library
  • check_circleOps leaders standardizing incident response quality

Example output

smart_toySample response
## Diagnose ```bash kubectl -n payments get pods -l app=gateway -o wide ``` Healthy: all 12 pods READY=1/1…
signal_cellular_altadvanced

Latest Insights

Stay ahead with the latest in prompt engineering.

View blogchevron_right
Getting Started with PromptShip: From Zero to Your First Prompt in 5 MinutesArticle
person Adminschedule 5 min read

Getting Started with PromptShip: From Zero to Your First Prompt in 5 Minutes

A quick-start guide to PromptShip. Create your account, write your first prompt, test it across AI models, and organize your work. All in under 5 minutes.

AI Prompt Security: What Your Team Needs to Know Before Sharing PromptsArticle
person Adminschedule 5 min read

AI Prompt Security: What Your Team Needs to Know Before Sharing Prompts

Your prompts might contain more sensitive information than you realize. Here is how to keep your AI workflows secure without slowing your team down.

Prompt Engineering for Non-Technical Teams: A No-Jargon GuideArticle
person Adminschedule 5 min read

Prompt Engineering for Non-Technical Teams: A No-Jargon Guide

You do not need to know how to code to write great AI prompts. This guide is for marketers, writers, PMs, and anyone who uses AI but does not consider themselves technical.

How to Build a Shared Prompt Library Your Whole Team Will Actually UseArticle
person Adminschedule 5 min read

How to Build a Shared Prompt Library Your Whole Team Will Actually Use

Most team prompt libraries fail within a month. Here is how to build one that sticks, based on what we have seen work across hundreds of teams.

GPT vs Claude vs Gemini: Which AI Model Is Best for Your Prompts?Article
person Adminschedule 5 min read

GPT vs Claude vs Gemini: Which AI Model Is Best for Your Prompts?

We tested the same prompts across GPT-4o, Claude 4, and Gemini 2.5 Pro. The results surprised us. Here is what we found.

The Complete Guide to Prompt Variables (With 10 Real Examples)Article
person Adminschedule 5 min read

The Complete Guide to Prompt Variables (With 10 Real Examples)

Stop rewriting the same prompt over and over. Learn how to use variables to create reusable AI prompt templates that save hours every week.

pin_invoke

Token Counter

Real-time tokenizer for GPT & Claude.

monitoring

Cost Tracking

Analytics for model expenditure.

api

API Endpoints

Deploy prompts as managed endpoints.

rule

Auto-Eval

Quality scoring using similarity benchmarks.