Skip to main content

Leverage Record: February 23, 2026

AI Time Record

About the author: I'm Charles Sieg, a cloud architect and platform engineer who builds apps, services, and infrastructure for Fortune 1000 clients through Vantalect. If your organization is rethinking its software strategy in the age of AI-assisted engineering, let's talk.

Daily accounting of what Claude Opus 4.6 built today, measured against how long a senior engineer familiar with each codebase would need for the same work. These are leverage factors, not time savings. Most of these projects are ones I would not have started without AI. The leverage factor measures how much more I can ship, not how much faster I finish.

About These Records
These time records capture personal project work done with Claude Code (Anthropic) only. They do not include work done with ChatGPT (OpenAI), Gemini (Google), Grok (xAI), or other models, all of which I use extensively. Client work is also excluded, despite being primarily Claude Code. The actual total AI-assisted output for any given day is substantially higher than what appears here.

The Numbers

# Task Human Est. Claude Leverage
1 Implement four engine subsystems with full test coverage 120 hours 40 min 180x
2 Write seven architecture deep-dive articles and apply writing style revisions across all content (4,198 line insertions) 60 hours 30 min 120x
3 Fix 77 diagram reference numerals, placeholder steps, and section headers across 11 documents 40 hours 25 min 96x
4 Write authentication architecture deep-dive article (~470 lines, 6 diagrams, 12 tables) 8 hours 8 min 60x
5 Fix diagram truncation and pseudocode overflow across 11 application documents 16 hours 25 min 38x
6 Create two domain specification documents with README 16 hours 25 min 38x
7 Write messaging architecture article (607 lines) and revise 175 files for AI detection scores 12 hours 25 min 29x
8 Fix Mermaid diagram numerals and labels (9 figures) 4 hours 12 min 20x
9 Fix Mermaid diagram numerals (8 figures) 4 hours 12 min 20x
10 Run AI content detection scoring across 215 files and add results to frontmatter 4 hours 15 min 16x

Aggregate

Metric Value
Tasks completed 10
Human equivalent 284 hours (~7.1 work weeks)
Claude wall-clock 217 minutes (~3.6 hours)
Tokens consumed ~1,495,000
Weighted leverage factor 78.5x

Analysis

The standout task was a subsystem implementation: four interconnected engine components with full test suites, estimated at 120 human-hours (three work weeks). Claude completed it in 40 minutes at 180x leverage. Greenfield implementation of well-scoped subsystems is where AI code generation delivers its highest returns. The requirements were precise, the interfaces were defined, and the test expectations were clear. Claude spent zero time on orientation or context-switching between the four subsystems.

The second highest factor (120x) came from writing seven architecture articles in a single session while simultaneously applying writing style revisions across the entire existing content library. A human would need to context-switch between research, writing, and editing across 31 files. Claude treats the batch as a single operation.

The lowest factor (16x) was AI content detection scoring: running every article and post through a detection API and writing results back to frontmatter. The bottleneck was API latency, not cognitive complexity. Claude waited for external responses the same way a human would. Tasks gated by external I/O compress less than tasks gated by thinking time.

Pattern from today: leverage correlates strongly with the ratio of thinking to waiting. Tasks that are 90% cognitive work (architecture, implementation, refactoring) produce leverage of 60x or higher. Tasks that involve external API calls, file-by-file mechanical edits, or rendering pipelines land at 15-40x. The human brain is the bottleneck that AI removes; when the bottleneck is something else (network latency, disk I/O), the advantage narrows.

Seven work weeks of output in a single afternoon. The aggregate 78.5x leverage factor means every minute of Claude time replaced roughly 1.3 hours of human engineering.

How This Works

I use Claude Code (Anthropic's CLI agent) for all engineering work across multiple projects. Claude Code runs in the terminal, reads and writes files, executes commands, runs tests, and iterates until the task is complete.

Every non-trivial task gets a leverage record. At the start of each task, Claude estimates how long a senior engineer already familiar with the codebase would need for the same work. After Claude finishes, it records the wall-clock time. The ratio is the leverage factor:

Leverage Factor = Human-Equivalent Hours × 60 / Claude Minutes

Claude logs each record to a CSV file (leverage_factor_log.csv) and to a per-project markdown summary. At the end of each day, I prompt Claude with "post today's leverage record" and it reads the CSV, filters for the current date, sanitizes the task descriptions (removing specific project names, client names, and proprietary details), calculates the aggregates, writes the analysis, and publishes this post. The entire daily record process, from prompt to published post, takes about two minutes.

The leverage factor is not a speedup metric. Most tasks in this record are projects I would never have started without AI. Writing seven architecture articles in a weekend, implementing four engine subsystems in an afternoon, scoring 215 files through an AI detection API: these are not things I was doing slowly before. They are things I was not doing at all. The factor measures expanded capability, not compressed schedules. That is why I call it leverage rather than time savings.


This is the first in a daily series of leverage records. See all records under the Time Record tag.

Let's Build Something!

I help teams ship cloud infrastructure that actually works at scale. Whether you're modernizing a legacy platform, designing a multi-region architecture from scratch, or figuring out how AI fits into your engineering workflow, I've seen your problem before. Let me help.

Currently taking on select consulting engagements through Vantalect.