Skip to main content

The AI Operating Rhythm

Running AI in production needs a rhythm: monitoring, evaluation, improvement, and governance. The weekly and monthly cadence for teams that ship AI.
8 October 2025·8 min read
Isaac Rolfe
Isaac Rolfe
Managing Director
Shipping AI to production is the beginning, not the end. Every AI system needs ongoing attention: monitoring, evaluation, improvement, and governance. Most organisations know this in theory. In practice, they ship the system and move on to the next project. The AI quietly degrades until someone notices. What is missing is not awareness. It is rhythm.

Why Rhythm Matters

AI systems are living systems. They interact with changing data, changing user behaviour, and changing model versions. Unlike traditional software, where the code does the same thing until someone changes it, AI systems change behaviour over time even when nobody touches them.
This means AI operations cannot be event-driven (fix things when they break). They must be cadence-driven (check things on a rhythm, whether they appear broken or not). The rhythm catches drift before it becomes a problem and ensures continuous improvement is systematic rather than reactive.

The Weekly Rhythm

Monday: Quality Review (30 minutes)

Review the automated quality metrics from the previous week. This is a structured check, not a deep dive:
  • Has output quality (as measured by your evaluation suite) changed?
  • Have user correction or rejection rates shifted?
  • Are any specific task types or user groups showing degradation?
  • Have any model provider updates been announced?
Most weeks, everything is fine. The review takes fifteen minutes and you move on. The weeks where something is off, you catch it early.

Wednesday: Cost and Performance Check (15 minutes)

Review the operational metrics:
  • What did AI cost this week? Is it tracking to budget?
  • Are latency metrics within acceptable bounds?
  • Are there any error spikes or availability issues?
  • Is usage growing, shrinking, or stable?
This check is about operational health. AI systems that cost too much get cut. AI systems that are too slow get abandoned. Catching these trends weekly prevents unpleasant surprises at the monthly review.

Friday: User Feedback Triage (30 minutes)

Review any user feedback, support tickets, or complaints related to AI capabilities. Categorise into:
  • Quality issues: The AI output was wrong or unhelpful. These feed into the quality evaluation.
  • Feature requests: Users want the AI to do something it currently does not. These feed into the product roadmap.
  • Workflow friction: The AI works but does not fit smoothly into the user's workflow. These feed into UX improvements.
  • Praise: The AI did something well. These are important for identifying what to protect and amplify.

The Monthly Rhythm

Month-Start: Performance Deep Dive (2 hours)

A thorough review of the previous month's AI performance:
Quality assessment. Run the full evaluation suite, not just the daily automated checks. Include human evaluation of a production sample. Compare against the previous month's baseline.
Cost analysis. Total cost, cost per task type, cost per business outcome. Trend analysis: are costs increasing, decreasing, or stable? Are there optimisation opportunities?
Usage patterns. Who is using AI? For what? Are adoption patterns changing? Are there user groups that have stopped using AI? Why?
Model evaluation. Are new models available that might perform better or cost less? Run comparative evaluations against your test suite.

Mid-Month: Improvement Sprint (varies)

Based on the month-start deep dive, execute one to three improvements:
  • Prompt refinements to address quality issues identified in evaluation
  • Model routing adjustments to optimise cost
  • Retrieval pipeline improvements to address relevance gaps
  • UX changes to address workflow friction
Keep improvements small and measurable. Each improvement should have a hypothesis ("this prompt change will reduce correction rates for classification tasks by 10%") and a measurement plan.

Month-End: Governance Review (1 hour)

Review AI governance:
  • Compliance check. Are AI capabilities operating within policy? Any incidents, near-misses, or policy gaps?
  • Access review. Are the right people accessing AI capabilities? Any unauthorised use or inappropriate data access?
  • Risk assessment. Have any new risks emerged? Are existing mitigations still appropriate?
  • Stakeholder update. Prepare a summary for leadership: what AI is doing, how it is performing, what it costs, what is next.

The Quarterly Rhythm

Strategic Review (half day)

Step back from operations and ask strategic questions:
  • Are our AI capabilities aligned with business priorities? Have priorities changed?
  • Are we investing in the right capabilities? Should we sunset any?
  • How does our AI maturity compare to where we wanted to be?
  • What capabilities should we build next quarter?
This review connects AI operations to business strategy. Without it, AI programmes drift into technical optimisation without business alignment.

Evaluation Suite Update

Update the golden test suite and evaluation criteria:
  • Add new test cases from production edge cases discovered this quarter
  • Remove test cases that are no longer relevant
  • Update evaluation criteria to reflect any business rule or policy changes
  • Recalibrate human evaluation rubrics
The evaluation suite is a living document. If it does not evolve with the system, it stops catching the issues that matter.

Making It Stick

The rhythm only works if it is actually followed. Three things that help:
Calendar it. Put recurring blocks on the calendar for every check. If it is not scheduled, it will not happen. Urgent work will always take priority over routine checks unless the routine is protected.
Assign ownership. Each check needs a named person responsible. Not a team. A person. Teams diffuse responsibility. A named owner ensures it happens.
Keep it lightweight. The weekly checks should take less than two hours total. If the rhythm becomes burdensome, people will skip it. Start with the minimum and add only what proves valuable.
The organisations that run AI well are not the ones with the best models or the most data. Consistent, lightweight, and relentless.
Isaac Rolfe
Managing Director

Starting from Nothing

If you currently have no AI operating rhythm:
Week 1: Set up automated quality monitoring. Even a basic dashboard that tracks output volume, error rates, and latency is a starting point.
Week 2: Create a lightweight weekly review template. Three questions: is quality stable, are costs on track, are users happy?
Week 3: Schedule the first monthly deep dive. Use it to establish baselines.
Week 4: Assign ownership and calendar the recurring checks.
Within a month, you have a functioning operating rhythm. It will not be perfect. But it will be infinitely better than what most organisations have, which is nothing.