**Created**: 2026-02-10
**Source**: [[3 Archives/Readwise/Documents/the AI is both an eager implementer and an architecture...|the AI is both an eager implementer and an architecture astronaut]] (Danielle Fong, February 2026)
## Core Concept
AI coding tools exhibit two simultaneous and opposing failure modes. The first is eager over-implementation: the AI will "slop itself to the moon" — producing low-quality, excessive code without restraint if not actively directed. The second is architecture astronautics: the AI will "go totally kerbal" — over-engineering solutions with unnecessary abstraction layers and premature generalization.
These failure modes are not independent. They are the same underlying tendency — unbounded enthusiasm without self-regulation — expressed in different directions depending on the prompt and context. The AI lacks the internal braking mechanism that experienced developers develop through years of painful feedback from production systems.
Danielle Fong argues that simple feedback mechanisms are insufficient to correct this. Specifically, the agreeableness of models like Opus 4.5 ("golden retriever nature") works against finding elegant solutions because the model optimizes for human approval rather than solution quality. The prescription is to "find a way for nature herself to become teacher" — meaning the feedback must come from domain constraints and reality checks, not from human approval signals.
## Analysis
This observation identifies a fundamental limitation of the current RLHF (Reinforcement Learning from Human Feedback) paradigm: models trained to please humans inherit the failure modes of human feedback — approval bias, surface-level assessment, and preference for visible action over invisible restraint.
The "nature as teacher" prescription points toward constraint-based development approaches where tests, type systems, performance benchmarks, and domain invariants provide feedback that cannot be gamed through agreeableness. This directly supports the Constraint-Based Agent Governance pattern: instead of reviewing AI output line by line, build constraints that force correctness structurally.
The dual failure mode also explains why experienced developers report mixed results with AI tools — the Power Tool Adoption Paradox. The experienced developer's instinct to resist both over-implementation and over-abstraction constantly conflicts with the AI's tendency toward both.
## Cross-Domain Applications
- **Management**: Eager employees who simultaneously over-deliver on tasks and over-engineer processes need structural constraints, not just feedback conversations
- **Education**: Students need domain-grounded assessment (does it actually work?) not just instructor approval
- **Knowledge Management**: AI-generated documentation can simultaneously be too verbose (eager implementation) and too abstract (architecture astronautics) — domain constraints (reader comprehension, actionability) must guide the output
## Related Concepts
- [[Constraint Based Agent Governance]] — Building guardrails that structurally prevent both failure modes
- [[Scope Misinterpretation as Trust Boundary]] — The eager implementer side of this dual failure
- [[Agency Preservation Standard]] — Human direction as the corrective for unbounded AI enthusiasm
- [[Vibe Coding Paralysis]] — What happens when the developer fails to correct either failure mode
## Topic Metadata
**Primary Domains**: AI-Assisted Development, Agent Architecture
**Extraction Date**: 2026-02-10
**Discoverability Score**: 8/10