Computer-Using Agents: Notes on Human Checkpoints
When agents can click/type/navigate software, value goes up fast. Risk also goes up fast.
I treat these systems as supervised operators, not independent decision makers.
Useful mental model
A very fast operator that:
- can follow instructions
- can recover from minor UI changes
- can handle repetitive digital tasks
But still needs checkpoints near irreversible actions.
Typical loop
- Read screen state
- Choose next action
- Execute action
- Verify state changed as expected
- Repeat
This is more adaptive than classic RPA. Still weaker than humans on ambiguous/high-stakes tasks.
Action lanes I use
- Green: read/search/summarize
- Yellow: low-risk changes with audits
- Red: explicit human approval (money/data/permissions/legal)
If everything is "green," incident probability rises quickly.
Where checkpoints are non-negotiable
- payments/refunds
- data deletion
- permission changes
- compliance/legal submissions
- healthcare/safety-critical ops
Trend signals behind this note
- Google announced AI Mode and Project Mariner updates on May 20, 2025: Google Search AI Mode update.
- OpenAI added computer-use capabilities in Responses API updates on May 21, 2025: New tools and features in the Responses API.
Sticky takeaway
Let computer-using agents execute routine work. Keep humans on risky boundaries.