Files
taskgraph_ts/docs/architecture/decisions/005-no-depth-escalation-v1.md
glm-5.1 bde1cc4e70 Decompose monolithic architecture.md into modular docs/architecture/ documents
The 751-line architecture.md violated the SDD process modular documentation
target (~500 lines). It also had duplicate TaskGraph class definitions (one
monolith, one decomposed) that directly contradicted each other, and embedded
consumer-specific tool dispatch mappings that belong in downstream projects.

Changes:
- Split into 8 focused documents + 7 ADR records + redirect page
- Removed the monolithic TaskGraph class (kept only decomposed version)
- Moved CLI→plugin dispatch mapping out (belongs in plugin architecture)
- Extracted implementation code (frontmatter splitter, findCycles, DAG
  propagation) into WHAT/WHY descriptions per architect role spec
- Added proper ADR format for all resolved design decisions
- Fixed review issues: C_fail mapping, DuplicateNodeError/DuplicateEdgeError
  types, ValidationError/GraphValidationError definitions, mutation error
  handling contract, enum naming convention, validation timing clarification
2026-04-26 06:38:52 +00:00

1.4 KiB

ADR-005: No depth-escalation heuristic in v1

Status: Accepted

Context

In the DAG-propagation model, each hop compounds another <1.0 factor. This implicitly captures depth effects — deeper chains have more compounding. An explicit depth-escalation heuristic (increasing risk at deeper chain levels) would add another multiplicative penalty on top.

Decision

Defer depth-escalation to v2. The multiplicative propagation model already captures depth effects implicitly. Adding an explicit depth heuristic would double-count the depth effect until we have empirical calibration data from actual task outcomes.

Consequences

Positive

  • No double-counting of depth effects
  • Simpler model to explain, implement, and debug
  • Architecture supports future depth-escalation via per-edge qualityDegradation adjustments or risk categorical escalation without API changes

Negative

  • May underestimate cost for very deep dependency chains where risk genuinely escalates with depth
  • The model treats all "hops" as equivalent — a 5-hop chain where each step is moderate risk may actually be worse than the model predicts

Future

If empirical data from actual task outcomes shows that depth-escalation is needed, it can be added without API changes — either by adjusting qualityDegradation per depth, or by escalating the risk categorical. This is a calibration question, not an architecture question.