Brief mode: condensed view. Switch to Full for persona notes and full analysis.
Your AI copilot is eating your career path. Change my mind.
Post #4 was about what model tiers know. This one is about what organizations stop teaching when output gets cheap.
Tech can generate both junior-level and senior-level output.
Where do the next senior engineers come from if the junior stage gets skipped?
Formation requires repeatable reps: bounded scope, explicit reasoning, and frequent feedback loops.
Without stop rules and judgment checkpoints, fluent output can mask a shrinking skill base.
Tech companies discovered something convenient in 2024: LLMs can produce senior-level code. So they froze senior hiring. Then they discovered something more convenient: LLMs can also produce junior-level code. So they froze junior hiring too. The result is a hiring market where the only people getting software jobs are the ones who already have them.
Everyone has noticed the market shock. Fewer people are talking about the formation consequences. The companion argument is in Post #3.1
This is not a rhetorical question. There is a real answer to it, and the real answer is: the same place they always came from. They come from junior engineers who did the work of being junior engineers long enough to become mid-level engineers who did the work of being mid-level engineers long enough to become senior engineers who accumulated enough judgment to know when the code was wrong even when it looked right.
This is the key distinction from Post #4. Post #4 asked where knowledge lives across model tiers. This post asks where human judgment formation lives across career tiers. Different problem, different failure mode.
If LLMs are doing the junior work, the junior stage does not happen. You cannot skip it and pick up the formation later. The judgment that distinguishes a senior engineer from a person who can generate senior-looking output is built from a decade of being wrong in instructive ways. It is not transferable. It is not compressible. It is not something you can prompt-engineer your way into.2
The incentive loop in slow motion
The loop is economically rational in the short term. A manager is measured on quarterly output. LLM-assisted developers ship more tickets this quarter. Junior roles look expensive relative to immediate output. So entry-level hiring gets cut first.
Promotions then follow output proxies. A developer using LLM tools from day one produces fluent artifacts, accumulates title seniority, and misses the failure exposure that title used to imply. Five years later, an incident arrives with a failure mode the model has not seen: odd load behavior, data corruption, a security edge case. The "senior" on call has the title. They do not have the substrate.
The LLM cannot help because it needs a competent human to interpret its output correctly, and that is precisely what is not in the room.
⚖️ Devil's Advocate Challenge
Pipeline math nobody wants to run
- Cut entry-level roles now to improve near-term efficiency.
- Three years later, the mid-level bench is thinner than planned.
- Seven years later, senior roles are filled by title progression, not deep formation.
- The incident burden shifts to a smaller pool of truly formed engineers.
And if you are a jr engineer, you completely lack the intuition to tell the difference. That is not an insult. It is the definition of being junior. The problem is that the LLM doesn't tell you which outputs you should be skeptical of. It tells you they're all fine.
None of those steps are mysterious. They are a predictable consequence of optimizing for quarterly output while treating apprenticeship as overhead.
🔨 Campion Builder note
The symptom I watch for is review collapse. PRs stay green, cycle time drops, and nobody can explain why one approach beat another without asking the model again.
What to measure before the pipeline breaks
If this is only a philosophy argument, it gets ignored. Treat it as an operating metric problem and it becomes manageable.
- Unassisted change ratio: share of production changes a junior can explain end-to-end without AI assistance.
- Shadow-to-solo ratio: how often a junior moves from paired execution to independent execution on similar tasks.
- Incident apprenticeship rate: percentage of incidents where a junior is in the room for diagnosis, not just cleanup.
- Review depth: count of reviews that ask for reasoning and tradeoffs, not only syntax fixes.
Those signals are uncomfortable because they reduce ambiguity. If the unassisted change ratio is flat for two quarters while ticket throughput is rising, you are buying output by liquidating formation.
A practical protocol for small teams
- Pick one class of task each week that must be done without copilots for first-pass implementation.
- Require a short architecture note before code on any feature above a fixed complexity threshold.
- Run one monthly incident replay where juniors narrate diagnosis paths and rejected hypotheses.
- Promote on demonstrated judgment artifacts, not only delivered ticket volume.
This does not require banning tools. It requires ring-fencing parts of the workflow where formation must happen on purpose.
🪡 Seton Formation note
A pipeline is not only staffing math. It is formation stewardship. If no one is tasked with that stewardship, no one should be surprised when the habit disappears.
The steelman and why I’m not convinced
The steelman counterarguments are real. Companies might hire fewer but better engineers. Prompt engineering might become a real junior stage. The calculator panic in math education turned out mostly wrong.
Here is why I think the steelman does not fully hold: calculators did not also write your proofs. The thing LLMs are doing is not accelerating human learning. It is substituting for it. The junior developer using Claude Code is not developing faster. They are not developing.3
The output ships. The formation does not happen. These are not the same thing, and conflating them is what makes the counterarguments sound more reassuring than they are.4
The long-term risk profile
Companies freezing junior hiring are responding rationally to a short-term signal while ignoring a ten-year structural consequence that will not be their problem to solve.
By the time the formation gap is visible in production, the decision-makers who froze the pipeline will have moved on.
⛺ Rowan Orchestrator note
The pipeline math isn't hypothetical — it's already showing up in incident queues. The “senior” on the rotation has the title and the LLM. What they don't have is the hypothesis space. They don't know what hypotheses to generate, so they hand the incident to the model, which generates the most statistically likely cause. Which is rarely the actual cause when you're in the tail.
I do not have a clean solution. Apprenticeship models, deliberate constraints on AI tool use during training periods, and bootcamps designed around struggle rather than output are partial answers. None scales like "just use the LLM" scales.
So: where am I wrong? Is there a formation mechanism I’m not seeing?
Prompt: test your own formation assumptions
Pick one engineer on your team (or yourself) who relies heavily on AI coding tools. Answer with receipts: 1. Current unassisted change ratio (estimate a %). 2. One task type where they can execute solo and explain tradeoffs. 3. One task type where output is fluent but reasoning is thin. 4. One workflow constraint you can enforce next sprint to close that gap. If you cannot put a number and a constraint on this, you are managing output, not formation.
Notes
- Companion piece: The Struggle Is the Point (Workshop #3).
- Skill-formation receipt (arXiv): Shen & Tamkin, How AI Impacts Skill Formation (2026), reporting lower conceptual mastery under delegation-heavy interaction patterns.
- Student-workflow receipt (arXiv): Shihab et al., The Effects of GitHub Copilot on Computing Students' Programming Effectiveness, Efficiency, and Processes in Brownfield Programming Tasks (2025), with mixed effectiveness/efficiency outcomes by task context.
- Calibrated-use receipt (PsyArXiv): Bassan et al., Acceptance Is Not Enough: Toward a Psychology of Calibrated GenAI use (2026), separating adoption behavior from effective use quality.
- Version: v1.0 — initial publication
- Frame: Devil’s advocate invitation → formation pipeline risk
Counterpoint: maybe the market is simply demanding higher leverage engineers earlier. If that is true, teams still need a proof mechanism showing judgment is actually being formed, not inferred from velocity.