Skip to main content
JG is here with you โœจ
Back to Blog
Blog Post

The Ops Mindset: Systems That Run Themselves

Automation scripts, rollback drills, AI governance, and the operational rituals that turn chaotic projects into calm systems. The expert playbook for AI-assisted development.

F
Frame Architect
Author
2025-12-01
Published
2025-12-01
Last Updated
โ—† โ—† โ—†

The Ops Mindset: Systems That Run Themselves

Part 1 taught you to set up calm. Part 2 taught you to ship with confidence.

Part 3 is about building systems that don't need you to babysit them.

We're going to automate your quality checks, practice what to do when things break, and establish rules for working with AI that keep you in control. This is the stuff that separates "I built a thing" from "I built a thing that keeps working."

You don't need a DevOps title. You just need the willingness to think in systems instead of one-off fixes.

Explain it three ways

๐Ÿ‘ถ Like I'm 5

We're turning our LEGO city into a real theme park. That means setting up rules so every ride works the same way, practicing what to do if a light goes out, and keeping a clipboard with all the "what happened?" stories. That way, when friends come to visit, everything runs smoothly without us having to fix things every five minutes.

๐Ÿ’ผ Like you're my boss

This is the production operations playbook. It covers automation scripts that enforce quality gates, rollback procedures with documented recovery times, AI governance policies that maintain accountability, and analytics loops that catch issues before users report them. ROI: reduces incident response time by 60% and prevents 80% of "it worked on my machine" deployments.

๐Ÿ’• Like you're my girlfriend

Remember how we plan the whole weekend before inviting friends overโ€”the food, the playlist, the backup plan if it rains? This is that, but for software. We're writing down the recipes, testing the smoke alarms, practicing what to do if the oven fails, and keeping notes on what worked so next time is even smoother. It's the difference between hosting a party and hosting a party that runs itself.

The incident that changed everything

Let me tell you about a Tuesday afternoon.

14:03  Deployed new feature to production โœ“
14:17  Vercel logs show environment variable missing ๐Ÿค”
14:19  Users see blank screen ๐Ÿ˜ฌ
14:25  Team scrambles, no one knows how to rollback ๐Ÿ˜ฑ
14:45  Finally rolled back, postmortem scheduled

42 minutes of chaos. Users affected. Stress through the roof.

Now here's the same scenario with ops rituals in place:

13:45  Pre-deploy checklist catches missing env var
13:50  Fix added to .env.example + Vercel dashboard
14:00  Deploy succeeds
14:05  Lighthouse + analytics verified โœ“

Same feature. Zero drama.

Ops rituals don't prevent all emergencies. They make emergencies boring.

What we're building today

๐Ÿ“ Part 3: Systems, Automation, and Ops Rituals

โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
โ”‚  1. Automation Ladder    โ†’ Scripts that check   โ”‚
โ”‚  2. Rollback Drills      โ†’ Practice recovery    โ”‚
โ”‚  3. AI Governance        โ†’ Stay in control      โ”‚
โ”‚  4. Analytics Loop       โ†’ Catch issues early   โ”‚
โ”‚  5. Retro Ritual         โ†’ Learn from mistakes  โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜
SectionTimeWhat you'll learn
---------
Automation Ladder12 minScripts that enforce quality
Rollback Drills10 minHow to recover when things break
AI Governance10 minRules for AI collaboration
Analytics Loop8 minMonitoring and accessibility
Retro Ritual4 minLearning from incidents
Total44 minComplete ops foundation

Step 1: The Automation Ladder (12 minutes)

You don't have to automate everything at once. Climb the ladder at your own pace.

The four levels

LevelWhat it meansTime to set up
---------
1. Manual checklistPaper or Markdown list you run each deploy5 min
2. Deploy scriptScript that runs checks and stops if something fails20 min
3. CI/CDGitHub Actions runs checks on every pull request1-2 hours
4. Zero-touchMerge to main auto-deploys when all checks pass1 day

Most people stay at Level 1 forever. Level 2 is the sweet spotโ€”you get 80% of the benefit with 20% of the setup.

Level 2: The deploy script

Create a file called scripts/deploy.ps1 (Windows) or scripts/deploy.sh (Mac/Linux):

Windows (PowerShell):

# scripts/deploy.ps1
# Run all quality checks before deploying

Write-Host "๐Ÿ” Running lint..." -ForegroundColor Cyan
npm run lint
if ($LASTEXITCODE -ne 0) { 
    Write-Host "โŒ Lint failed. Fix errors before deploying." -ForegroundColor Red
    exit 1 
}

Write-Host "๐Ÿ” Running type-check..." -ForegroundColor Cyan
npm run type-check
if ($LASTEXITCODE -ne 0) { 
    Write-Host "โŒ Type-check failed. Fix errors before deploying." -ForegroundColor Red
    exit 1 
}

Write-Host "๐Ÿงช Running tests..." -ForegroundColor Cyan
npm run test
if ($LASTEXITCODE -ne 0) { 
    Write-Host "โŒ Tests failed. Fix errors before deploying." -ForegroundColor Red
    exit 1 
}

Write-Host "๐Ÿ—๏ธ Building..." -ForegroundColor Cyan
npm run build
if ($LASTEXITCODE -ne 0) { 
    Write-Host "โŒ Build failed. Fix errors before deploying." -ForegroundColor Red
    exit 1 
}

Write-Host "๐Ÿš€ All checks passed! Deploying to Vercel..." -ForegroundColor Green
vercel --prod --yes

Mac/Linux (Bash):

#!/bin/bash
# scripts/deploy.sh

set -e  # Stop on any error

echo "๐Ÿ” Running lint..."
npm run lint

echo "๐Ÿ” Running type-check..."
npm run type-check

echo "๐Ÿงช Running tests..."
npm run test

echo "๐Ÿ—๏ธ Building..."
npm run build

echo "๐Ÿš€ All checks passed! Deploying to Vercel..."
vercel --prod --yes

How to use it

Instead of running vercel --prod directly, run your script:

# Windows
.\scripts\deploy.ps1

# Mac/Linux
./scripts/deploy.sh

If any check fails, the script stops. No more "oops, I forgot to run lint" deployments.

๐Ÿ’ฐ ROI calculation

If manual checks take 6 minutes per deploy, and you deploy 20 times per month, that's 120 minutes of manual work. The script takes 20 minutes to set up. It pays for itself in the first month, then saves you 2 hours every month after.

Step 2: Rollback Drills (10 minutes)

Things will break. The question isn't if but whenโ€”and whether you'll panic or calmly fix it.

The rollback decision matrix

ScenarioSeverityWhat to do
---------
404 on new pageMediumFix forward if quick, rollback if >15 min
Missing env varHighRollback immediately, add var, redeploy
Styling bugLowFix forward, note in devlog
Data corruptionCriticalRollback, open incident, notify stakeholders

Fire Drill Friday

Once a month, practice a rollback. Here's the drill:

1. Record current deployment

vercel ls
# Note the current production URL

2. Trigger rollback

vercel rollback
# Select the previous deployment

3. Verify site works

  • Open the production URL
  • Click around, check key features
  • Note how long the rollback took

4. Roll forward again

vercel --prod --yes

5. Log the drill

## Rollback Drill โ€” 2025-12-01

- Started: 14:00
- Rollback complete: 14:03 (3 minutes)
- Verification: 14:05
- Roll forward: 14:08

Lessons: Rollback command is fast. Main delay was finding the right deployment ID.
Next time: Bookmark the Vercel dashboard deployments page.

๐Ÿ”ฅ Chaos engineering (lite version)

For extra credit: intentionally break a preview deployment. Remove a required env var, deploy to preview, and confirm that your monitoring catches it before you would have promoted to production. Document what you learned.

Step 3: AI Governance (10 minutes)

AI makes us faster. It can also make us sloppy if we're not careful.

The governance scorecard

Rate yourself honestly:

PracticeYour statusTarget
---------
Session loggingDo you write down what AI helped with?Log every major session
Secrets in promptsEver pasted API keys into Cursor?Never (use .env.example)
Technical debt trackingDo you note when AI takes shortcuts?Weekly review
Manual override protocolKnow when to ignore AI suggestions?Documented in AI_PRACTICES.md
Refactor impact analysisCheck what AI changes might break?Review all AI refactors

The session log template

After any significant AI collaboration, log it:

## AI Session โ€” 2025-12-01 โ€” Deploy Script

**Goal:** Create automated deploy script with quality gates

**Constraints:** Windows PowerShell, must stop on failures

**What AI helped with:**
- Generated initial script structure
- Suggested error handling pattern
- Wrote the colored output messages

**What I changed:**
- Added the build step (AI forgot it)
- Changed exit codes to match our standards

**Decisions made:**
- Kept manual Vercel auth (didn't automate login)
- Added comments explaining each section

**Next time:**
- Ask AI to include coverage threshold flag

The escape hatch flow

Sometimes AI suggestions feel wrong. Here's what to do:

AI suggestion feels off?
        โ†“
Stop autocomplete (Cmd+. or Esc)
        โ†“
Check AI_PRACTICES.md for guidance
        โ†“
Make the change manually
        โ†“
Log why you overrode the suggestion
        โ†“
Update PROMPT_PLAYBOOK if it's a pattern

๐Ÿšฉ Red flag phrases from AI

When you see these, pause and verify manually:

โ€ข "I guessed the file path..."
โ€ข "This might work..."
โ€ข "I can't access that file..."
โ€ข "I assumed you wanted..."

These are moments to take over, not moments to trust blindly.

Step 4: Analytics Loop (8 minutes)

You can't fix what you can't see. Set up basic monitoring.

The metrics dashboard

MetricToolHow often to check
---------
Core Web VitalsVercel AnalyticsWeekly
Lighthouse scoresChrome DevToolsEvery major change
Error logsVercel Functions logsDaily (quick scan)
User engagementGA4 or PlausibleWeekly

Accessibility quick audit

Before any major release, run through this checklist:

  • [ ] Keyboard navigation โ€” Can you use the entire site without a mouse?
  • [ ] Color contrast โ€” Do text and backgrounds have enough contrast? (Use DevTools)
  • [ ] Reduced motion โ€” Do animations respect `prefers-reduced-motion`?
  • [ ] Screen reader โ€” Do images have alt text? Do buttons have labels?

The weekly review ritual

Every Friday (or whatever day works), spend 10 minutes:

  • Open Vercel Analytics โ†’ Note any performance drops
  • Check error logs โ†’ Note any new errors
  • Run Lighthouse on homepage โ†’ Log the scores
  • Update `PERFORMANCE_LOG.md` with findings
## Weekly Review โ€” 2025-12-01

**Lighthouse:** Mobile 94, Desktop 98 (no change)
**Errors:** None new
**Analytics:** 12% traffic increase, bounce rate stable

**Action items:**
- None this week
- Consider image optimization next sprint

Step 5: The Retro Ritual (4 minutes)

When something goes wrongโ€”and it willโ€”capture what you learned.

The retro template

## Incident: [Brief description]

**Date:** 
**Duration:** 
**Severity:** Low / Medium / High / Critical

### What happened
[Timeline of events]

### Impact
[Who was affected and how]

### Root cause
[Why did this happen]

### Fix applied
[What you did to resolve it]

### Prevention
[What changes prevent this from happening again]

### Rule update
[New checklist item or process change]

Example entry

## Incident: Preview build failure

**Date:** 2025-12-01
**Duration:** 30 minutes
**Severity:** Medium (blocked preview, not production)

### What happened
- 14:00 Pushed feature branch
- 14:05 Vercel build failed
- 14:20 Found missing Tailwind color
- 14:30 Fixed and redeployed

### Impact
Preview blocked for 30 minutes. No user impact.

### Root cause
Used `bg-sage-500` but `sage` wasn't in tailwind.config.ts

### Fix applied
Added sage color palette to Tailwind config

### Prevention
- Add "color token check" to QA template
- Include tailwind.config.ts in work order context

### Rule update
New checklist item: "Verify all color tokens exist in Tailwind config"

Ops Ritual Bingo ๐ŸŽฏ

Track your progress. How many can you check off this month?

---------
โ˜ Ran rollback drillโ˜ Caught bug in previewโ˜ Updated AI_PRACTICES.md
โ˜ Hit 90+ Lighthouseโ˜ Wrote retro within 24hโ˜ Automated a manual step
โ˜ Reviewed analyticsโ˜ Rehearsed DNS failoverโ˜ Pair-reviewed AI refactor

Bingo = any row, column, or diagonal completed

Your completion checklist

The full Vercel ร— Cursor Learning Ladder:

Part 1: Calm Rituals โœ“

  • [ ] Folder structure with docs/, src/, content/
  • [ ] Four documentation files created
  • [ ] Tooling shakedown habit established
  • [ ] Prompt spine template ready
  • [ ] First Vercel deploy complete

Part 2: Quality Gates โœ“

  • [ ] Work order workflow in place
  • [ ] Feature shipped with tests
  • [ ] Preview deploy reviewed
  • [ ] Documentation updated

Part 3: Ops Rituals โœ“

  • [ ] Deploy script created and tested
  • [ ] Rollback drill completed and logged
  • [ ] AI governance scorecard filled out
  • [ ] Analytics loop established
  • [ ] First retro entry written

What's next?

You've built the foundation. The rituals. The systems.

Now it's about repetition. Every feature you ship, run through the workflow. Every incident, write a retro. Every month, practice a rollback.

The goal isn't perfectionโ€”it's predictability. When you know what to do, stress goes down. When stress goes down, you build better things.

๐ŸŽ“ You've completed the Learning Ladder

From calm setup to confident shipping to operational systems. You now have the same workflow used by professional teamsโ€”without needing years of experience to get here.

Interactive lab available

Practice ops rituals hands-on:

  • Automation ladder with ROI calculator
  • Rollback drill timer and logging
  • AI governance scorecard
  • Ops Ritual Bingo tracker

<a href="/learn-ai-lab/vercel-cursor/expert" className="text-purple-300 hover:text-purple-100 underline">โ†’ Try the Expert Lab</a>

Navigate the series

โ—†END OF ARTICLEโ—†
๐ŸŽฎ Fun Reminder
touch me

Every deploy is saved. Every version is recoverable. Vercel has your back.

F

About Frame Architect

Builder of JMFG.ca โ€” an interactive learning platform with 80+ hands-on labs for cybersecurity, web development, and AI workflows. Passionate about making complex topics accessible through real-world examples and the "Explain 3 Ways" teaching method.

Open to AI-Focused Roles

AI Sales โ€ข AI Strategy โ€ข AI Success โ€ข Creative Tech โ€ข Toronto / Remote

Let's connect โ†’
Terms of ServiceLicense AgreementPrivacy Policy
Copyright ยฉ 2026 JMFG. All rights reserved.