Skip to main content

Leverage Record: February 28, 2026

AI Time Record

About the author: I'm Charles Sieg, a cloud architect and platform engineer who builds apps, services, and infrastructure for Fortune 1000 clients through Vantalect. If your organization is rethinking its software strategy in the age of AI-assisted engineering, let's talk.

Eighteen tasks today across five workstreams: a resume generator built from scratch and iterated through three major revisions, knowledge synthesis tooling enhancements, reference architecture documentation, an ML validation pipeline, and a technical article on decision fatigue in agentic coding workflows.

About These Records
These time records capture personal project work done with Claude Code (Anthropic) only. They do not include work done with ChatGPT (OpenAI), Gemini (Google), Grok (xAI), or other models, all of which I use extensively. Client work is also excluded, despite being primarily Claude Code. The actual total AI-assisted output for any given day is substantially higher than what appears here.

Task Log

Task Human Est. Claude Time Tokens Leverage
Resume generator full implementation (6 phases: schemas, CLI, parsers, importers, renderers, LLM integration, templates, tests) 40h 20min 120x
Incremental checkpointing for synthesis and iteration phases 20h 12min 45k 100x
Resume generator 6-phase enhancement (recursive import, HTML/DOC parsers, multi-signal dedup, master skills, portfolio website) 16h 12min 85k 80x
Import pipeline overhaul: 7-phase implementation (schemas, classifier, extractor, DOCX parser, merger, website, tests) 12h 12min 60x
ML validation pipeline (architecture refactor, config, wiring, runners, test fix) 16h 18min 45k 53.3x
Model benchmarking framework + evaluation methodology with prompt tuning 16h 19min 50k 50.5x
Refactor scoring pipeline + update docs and tests across repositories 6h 8min 40k 45x
Enhanced scoring pipeline + reference architecture updates across 3 repositories 8h 12min 90k 40x
Port 6 interactive features to shared component library 8h 12min 85k 40x
Resume generator v2.0 schema restructure (source registry, 4 new entry types, 15-file cascade, reimport, docs) 8h 12min 90k 40x
Extract reference architecture into standalone document (3 files created + 7 modified) 16h 25min 120k 38.4x
Per-call API timing log for synthesis runs 12h 19min 45k 37.9x
Update docs and push 3 repositories for per-call API timing log 2h 4min 30k 30x
Scoring CLI tool + batch re-score 14 content packages 6h 15min 90k 24x
LLM-powered object normalization pipeline for resume generator 6h 15min 85k 24x
Article: agentic coding decision fatigue + leverage record update + staging/production deploys 8h 20min 200k 24x
Analyze diffs + organize 7 logical commits + push 2 repositories 2h 8min 30k 15x
Update leverage record post + AI detection scoring + staging/production deploy + pipeline docs + README 2h 25min 100k 4.8x

Aggregate Stats

Metric Value
Total tasks 18
Total human-equivalent hours 204h
Total Claude minutes 268min (4h 28min)
Total tokens ~1.23M
Weighted average leverage 45.7x

Analysis

The resume generator dominated the day. Four separate tasks spanning the same codebase: initial full implementation at 120x, a 6-phase enhancement pass at 80x, a complete import pipeline overhaul at 60x, and a v2.0 schema restructure at 40x. The declining leverage across iterations illustrates the leverage curve in action. Greenfield implementation compresses the most dramatically because there are no constraints. Each subsequent pass adds complexity: existing patterns to preserve, backward compatibility to maintain, and integration points to respect. Even so, the fourth pass at 40x still represents a task that would take a senior engineer a full working day completed in 12 minutes.

The 120x on the initial resume generator build stands out. Six implementation phases covering Pydantic schemas, an argparse CLI with seven subcommands, four document parsers (PDF, DOCX, Markdown, plaintext), an LLM-backed import pipeline with section classification and entity extraction, four output renderers (HTML, PDF, Markdown, JSON), and a Jinja2 template system with four built-in themes. A complete production-ready tool in 20 minutes.

Incremental checkpointing for synthesis runs hit 100x. This involved adding fault-tolerant checkpointing to long-running LLM synthesis pipelines so that partial progress is preserved across interruptions. The implementation touched the pipeline orchestrator, file I/O layer, and progress reporting, with careful attention to atomicity guarantees.

The model benchmarking framework (50.5x) involved building a new evaluation methodology: designing the scoring protocol, implementing the evaluation harness, and iterative prompt tuning to calibrate thresholds. The cognitive density was high, but the iteration cycles for tuning added wall-clock time.

The ML validation pipeline (53.3x) closed out the day. This involved refactoring the pipeline architecture, adding configuration management, wiring up runners, and fixing tests. Sixteen hours of estimated human work in 18 minutes.

The 4.8x on the leverage record update reflects the I/O-bound nature of the task: AI detection scoring across published content, waiting for API responses, and multi-stage deployment to staging and production. The bottleneck was external service latency, not implementation complexity.

A 45.7x weighted average across 18 tasks means roughly five weeks of senior engineering output in under four and a half hours of wall-clock time.

Let's Build Something!

I help teams ship cloud infrastructure that actually works at scale. Whether you're modernizing a legacy platform, designing a multi-region architecture from scratch, or figuring out how AI fits into your engineering workflow, I've seen your problem before. Let me help.

Currently taking on select consulting engagements through Vantalect.