Tradtional thinking
Each phase produces specific artifacts in traditional formats
Continued from Task decomposition
Excellent catch. This is exactly the kind of first-principles thinking we need. Let’s interrogate this assumption.
Traditional reasons:
But in our decomposed task model, are these reasons still valid?
What does the NEXT task in the dependency chain actually need?
If the next task is DEV4: Dashboard UI Components, what does the developer (human or AI) actually need?
Designer creates Figma file → Developer opens Figma → Developer inspects:- Layout dimensions- Component hierarchy- Colors (hex codes)- Typography (font, size, weight)- Spacing values- Interaction states- Responsive breakpoints→ Developer manually translates into React codeThis is insane information loss and manual translation work.
What if the design task outputs machine-readable specification, not a visual mockup?
{ "screen": "predictive_analytics_dashboard", "layout": { "type": "grid", "columns": 12, "gap": "24px", "responsive_breakpoints": ["768px", "1024px", "1440px"] }, "components": [ { "type": "HeaderBar", "position": { "row": 1, "col": "1-12" }, "props": { "title": "Predictive Analytics", "actions": ["refresh", "export", "settings"] } }, { "type": "MetricCard", "position": { "row": 2, "col": "1-4" }, "data_binding": "prediction_accuracy", "props": { "trend": "up", "comparison": "vs_last_week" } }, { "type": "ChartContainer", "position": { "row": 3, "col": "1-8" }, "props": { "chart_type": "line", "data_source": "predictions_over_time", "interactive": true } } ], "design_tokens": { "colors": { "primary": "#2563eb", "success": "#10b981", "background": "#ffffff" }, "typography": { "heading": "Inter/600/24px/32px", "body": "Inter/400/16px/24px" } }, "interaction_flows": [ { "trigger": "metric_card.click", "action": "show_detail_modal", "transition": "fade_in_300ms" } ], "accessibility": { "aria_labels": {...}, "keyboard_navigation": {...}, "contrast_verified": true }}This specification can:
Instead of: “Figma Asset Production”
Subtask WF-9A: Component Specification Generation
Subtask WF-9B: Visual Preview Generation
Subtask WF-9: Design Token Mapping
Result:
Dashboard screen composition:- Use: HeaderBar component (v2.3)- Use: MetricCard component (v1.8) × 3- Use: LineChart component (v3.1)- Custom spacing: 32px between cards (override default 24px)- Color override: Use 'success' variant for middle cardNext task (DEV4) receives:
What if there’s no intermediate artifact at all?
Subtask WF-9: Design Context Packaging
Then DEV4 (claimed by AI agent) receives:
AI development agent directly generates code from this context, no visual mockup needed.
Only if human developer claims DEV4 might they want visual reference - generated on demand.
Let’s apply this thinking to other “obvious” deliverables:
Traditional: Product Owner writes user stories in specific format
Alternative:
What’s actually needed:
{ "requirement_id": "REQ-034", "actor": "small_business_owner", "goal": "understand_customer_churn_risk", "context": "reviewing_weekly_metrics", "success_criteria": [ "churn_probability_displayed_per_customer", "risk_factors_explained_in_plain_language", "actionable_recommendations_provided" ], "constraints": { "response_time": "< 2s", "data_freshness": "< 24h", "accuracy_threshold": "> 75%" }}AI can generate “As a small business owner, I want to…” text for humans who need it, but other AI agents work directly from structured data.
Traditional: Developer writes OpenAPI spec, generates docs site
Alternative:
What’s actually needed:
Documentation becomes generated view of the code, not separate artifact.
Traditional: QA Engineer writes test plan document
Alternative:
What’s actually needed:
Feature: Churn Risk Prediction Scenario: High-risk customer identified Given customer "Acme Corp" has not logged in for 45 days And their usage dropped 60% last month When I view the churn risk dashboard Then I should see "Acme Corp" with "HIGH" risk status And I should see "Declining engagement" as risk factor And I should see "Send re-engagement offer" as recommendationThis IS the test plan, the test case, AND the documentation. One artifact, multiple uses.
Traditional: Someone types notes during meeting, posts to Confluence
Alternative:
What’s actually needed:
Meeting notes as prose document become obsolete.
Tradtional thinking
Each phase produces specific artifacts in traditional formats
New thinking
Each task produces data in the format optimally consumed by the next task
The flow:
Customer conversation (audio) → AI transcription + synthesis (structured insights) → Requirement specification (machine-readable) → Component specification (executable) → Working code (directly deployed) → Runtime behavior (instrumented/observable) → Usage analytics (feeds back to insights)Human-readable views are generated on-demand when humans need to review, but they’re not the “official” artifact.
Old model:
Task A: Create wireframes (output: Figma file) → Task B: Review wireframes (human opens Figma) → Task C: Implement UI (developer translates from Figma)New model:
Task A: Define UI composition (output: component spec) → Optional: Generate preview (if human review needed) → Task B: Implement UI (AI reads spec directly, generates code) → Automated: Verify output matches specFewer tasks, fewer translation layers, less information loss.
The platform must support:
Revised task breakdown:
WF-1 through WF-7: (Same as before - research and analysis)
WF-8: Human Judgment & Design Decision
WF-9: Component Specification Generation (Replaces Figma production)
WF-10: Specification Validation
WF-11: Visual Preview Generation (Optional)
Total time: 20 minutes (vs 30 minutes for Figma production) Total cost: 35-45 tokens (vs 40 tokens) Bonus: Next task (DEV4) can start immediately from spec, no translation needed
When we eliminate unnecessary translation layers:
Saved time per task: Small (10-30 minutes) Saved time across project: Massive (cumulative) Error reduction: Huge (each translation introduces errors) Iteration speed: Much faster (change spec, code regenerates)
Example cascade:
So the deeper question:
For every task in our project lifecycle, should we ask:
Should we go through the entire task list and eliminate cargo-culted deliverables?
And should Future’s Edge be designed around canonical data + generated views rather than artifact storage?