
The short version
Day 1 of agent deployment: my morning report had 47 items at 8 AM. 60% noise. By Day 14, the same agents produced 5-7 actionable items I trusted. Six dial adjustments did the work: severity weighting on Day 3, channel filtering on Day 5, customer-tier weighting on Day 7, action-first report structure on Day 10, comparative context on Day 12, and edge-case refinement in week 3-4. The playbook: treat the first two weeks as training, not production. Day 1 output is a first draft. You need to tune before you trust. By Day 14, the report became the first thing I checked each morning. Before that, it was the last.
Day 1. I deployed the Daily Focus agent and the Red Flag agent. Six minutes later, I got the first report: 47 items, flagged as "critical" or "review."
Forty-seven things before 8 AM.
I spent forty-five minutes reading through them. Most of it was noise. A Slack message from a customer that mentioned a competitor's name (flagged as market intel). A routine contract renewal hitting the calendar (flagged as revenue risk). A support ticket from a power user with a feature suggestion (flagged as product opportunity).
The signal-to-noise ratio was approximately 60/40. As in, 60% noise.
I thought about ripping the whole thing out and going back to manual monitoring. But I'd built these agents for a reason, and I knew from the playbook that the first two weeks are training, not production. So I started adjusting.
Here's every dial I turned.
Day 1-2: The Problem Is Too Broad
The agents were pattern-matching on everything vaguely important. The Daily Focus agent was supposed to surface the three to five things that actually mattered that day. Instead it was surfacing thirty things that might matter.
The Red Flag agent was supposed to catch anomalies indicating churn risk or opportunity risk. Instead it was flagging any customer Slack mention as a red flag.
The issue: I hadn't tuned the thresholds. The agents were running on defaults.
Adjustment Day 3: I added severity weighting. Not every Slack mention counts the same. A message from a tier-1 customer escalated as URGENT gets flagged. A mention of a competitor in an offhand Slack comment does not. I set minimum severity levels for what even counts as a "flag."
By end of Day 3, reports went from 47 items to 23 items. Still too much, but less egregious.
Day 4-5: Channels Matter
The Red Flag agent was scraping every Slack channel for customer mentions. Including channels like #random, #wins, and #engineering-discussion where team members randomly chatted about work stuff.
Turns out, if you're in a 300-person company and you scan every Slack channel for "customer" mentions, you find a lot of noise.
Adjustment Day 5: I excluded internal channels from the agent's scope. Customer mentions in #customer-wins are handled by a different playbook anyway. We care about customer mentions in #customer-issues, #support-urgent, #sales-opportunities. Real signal channels.
This cut the noise by another 40%. By end of Day 5, reports were down to 14 items.
Suddenly the report became readable.
Day 7: Tier-Weighted Severity
But 14 items was still not three to five. And some of them were still... not quite right.
A small customer had filed a support ticket about a minor bug. It got flagged as a red flag because the pattern looked like escalation (two tickets in one day). But two tickets for a company with one integration point is normal. Two tickets for a tier-1 customer who usually has zero is abnormal.
Adjustment Day 7: I added customer tier weighting to the severity algorithm. A pattern that's concerning for a tier-1 account might be totally fine for a tier-3 account. I adjusted the thresholds based on historical baseline. For tier-1 customers, deviation from their normal pattern triggers a flag at under-50 percentile. For tier-3 customers, only severe deviations matter.
This didn't reduce volume that much, but it reordered volume. The 14 items were now ranked by what actually mattered.
Day 9: Structure Beats Data
By Day 9, I had the right items, but the report structure was wrong. I was getting a wall of context: ticket mentions, Slack excerpts, historical data, severity scores, pattern explanations. The report took 20 minutes to parse.
I was a data analyst, not a PM looking for decisions.
Adjustment Day 10: I restructured the report to lead with actions, not data. Instead of:
Red Flag: Customer X, Slack mention on 2026-03-12 at 14:23, matched pattern "urgent language + multiple channel mentions + tier-1 account = escalation risk", severity score 8.7, frequency baseline 0.3, current frequency 1.2...
I changed it to:
[ACTION] Customer X may be at escalation risk. Last two support tickets used urgent language. Recommend: call today to verify status.
Then I put the supporting data below the action, not above it.
The report went from "data I have to analyze" to "decisions I need to make." By end of Day 10, the morning report took seven minutes to read. I could actually act on it.
Day 12: Context Beats Absolutes
But even with the new structure, I was still missing context. The report said "Customer Y has filed three tickets this week. Review." But I didn't know if that was normal escalation or genuinely unusual.
Adjustment Day 12: I added comparison context. Instead of absolutes, I gave historical context:
[RED FLAG] Customer Y - 3 tickets this week
- Their average: 0.5 tickets/week
- This week: 3x normal volume
- Recommendation: call to understand what changed
Now I could instantly tell if a number meant something. Three tickets for someone who normally files three tickets a week? Not a flag. Three tickets for someone who normally files one a month? Different story.
Day 14: The Turning Point
By Day 14, the report had shrunk to 5-7 items per day. All of them were actually important. The structure was decision-focused instead of data-focused. The context was comparative instead of absolute.
I'd stopped skimming and started reading.
And here's what changed: the report became the first thing I checked. Not the last thing, not the "maybe if I have time" thing. The first.
Because by Day 14, I trusted it.
Day 15-30: Refinement
The next two weeks were calibration. I noticed the agent was catching product bugs early (good), but also flagging every minor bug as an "adoption risk" (bad). I tightened the pattern.
I noticed certain customer segments had different communication styles, so I added segment-specific baselines. A customer in a high-velocity space uses urgent language more often; their threshold moved.
I noticed "competitor mention" didn't need to be a separate red flag type; I merged it into market intel and made it lower priority unless it was coupled with churn risk signals.
None of these changes were dramatic. They were tuning.
What Actually Happened
By end of Day 30, I had a report that took five minutes to read, contained three to five real items, and was actionable. It caught the churn signal I talked about in an earlier post. It surfaced an integration issue before it became a support nightmare. It highlighted a market opportunity we'd missed in the noise.
It wasn't because the agent was magic. It was because I did what the playbook says: treat the first two weeks as training, not production.
The Lesson for Your Teams
If you're deploying agents, here's what that training looks like:
- Week 1: Deploy and accept high noise. You're learning what the system actually does.
- Day 3: First threshold adjustments. What channel noise can you exclude?
- Day 5-7: Segment-specific tuning. Different things matter for different customer types.
- Day 10: Restructure the output. Data is useless without decision clarity.
- Day 12: Add comparative context. Absolutes are meaningless without baseline.
- Week 3-4: Edge case refinement. You're not changing fundamentals; you're catching exceptions.
The common mistake is treating Day 1 output as representative. It's not. It's a first draft. You need to tune before you trust.
But once tuned, agents become the thing you build your workflow around. Not because they're perfect - they're not. Because they're calibrated to your actual operations, not generic defaults.
The first report was unusable. By Day 14, it was indispensable. The difference was two weeks of disciplined tuning.
That's the playbook working: acknowledging that new tools need training before they deliver, and structuring that training so it's systematic instead of chaotic.
Also on Medium
Full archive →AI Agents and the Future of Work: A Pixar-Inspired Journey
What product managers can learn about AI agents from how Pixar runs a film team.
Many AI Agents Are Actually Workflows or Automations in Disguise
How to tell agents from workflows from cron jobs, and why it matters for what you ship.