1) Core thesis
- The primary problem: humans and AI do not share a provable definition of “done”
- The consequence: assistants simulate progress, but users cannot delegate responsibility
- The solution: a system that represents outcomes as claims about reality and closes them only via evidence or explicit human confirmation
2) Canonical definitions
- Loop: a verifiable claim about the future that stays open until reality proves it true or a human explicitly confirms it
- Closure: the set of conditions that must be satisfied to mark a loop closed
- Signal: an immutable fact emitted by integrations, user actions, or systems of record
- Action: an attempt to move a loop toward closure
- Contradiction: a signal that invalidates closure and reopens the loop
3) Product vision
- One “truth layer” above all tools that answers: what is still not actually done?
- The app becomes the canonical place to see unresolved reality
- AI helps propose actions and definitions; the system proves closure; humans arbitrate ambiguity
- Expansion path: from email + calendar to “any domain” via templates + sensors + confirmation
4) Principles and non-negotiables
- Reality closes loops, not AI
- Closure must be auditable (who/what/why/evidence)
- Evidence-closed preferred, asserted-closed allowed, AI-only closure prohibited
- Contradictions reopen loops automatically
- High-stakes loops require stricter closure and higher approval thresholds
- Extensibility must not degrade truthfulness
5) Conceptual model
- Goals describe “why” (optional)
- Loops define “what must become true” (strict)
- Actions attempt to make it true
- Signals prove whether it happened
6) Object model
6.1 Minimal Loop schema
-
Fields
- id
- objective (human-readable claim)
- status (OPEN, IN_PROGRESS, BLOCKED, AT_RISK, CLOSED, REOPENED, ABANDONED)
- closureSpec (how done is proven)
- context (people, threads, events, docs, etc.)
- risk (low/med/high), external boolean, irreversible boolean
- createdAt, updatedAt, dueBy optional
6.2 ClosureSpec (constrained, machine-evaluable)
- mode: evidence | asserted
- conditions: list of allowlisted conditions
- requiresUserApproval boolean
- confidence (informational, never authoritative)
- createdBy: system | ai | user
- versioning
6.3 Signal (facts)
- id, loopId
- type (open string)
- source (integration/system/user/agent)
- payload (optional)
- timestamp
- immutability
6.4 Action (attempts)
- id, loopId
- type
- preview / proposed change
- requiresApproval
- status: proposed/approved/executed/failed
- executedAt
6.5 DecisionTrace (audit + learning)
- proposed action, user response (approved/edited/rejected)
- diff or notes
- timestamp
- ties to outcomes for learning
6.6 Optional Goal object
- flexible tracking layer
- can spawn loops (templates or manual)
7) Closure semantics and risk control
7.1 Two closure modes
- Evidence-closed: conditions satisfied by signals
- Asserted-closed: explicit user confirmation (optionally with proof)
7.2 “Closure is a fact” enforcement
-
Loops cannot close without:
- required evidence signals, or
- explicit confirmation condition
-
No “AI thinks it’s done”
7.3 Contradiction and auto-reopen
- Define reopen signals per loop type/template
- Automatic transition: CLOSED → REOPENED on contradiction
- Record contradiction evidence and cause
7.4 Risk gating rules
-
If external or irreversible or high-risk:
- prefer evidence-closed only
- require HITL for actions
- require explicit confirmation if evidence incomplete
-
Low-risk internal loops may allow looser confirmation, but still auditable
8) AI’s role (strictly scoped)
-
AI may:
- detect candidate loops
- propose closure specs (in constrained DSL)
- propose next actions
- summarize why a loop is open
- rank risk/priority and detect stalling
-
AI may not:
- mark loops closed
- suppress contradictions
- invent evidence
- silently resolve ambiguity
-
The system, not AI, is the authority on closure state
9) Templates: how the system becomes universal
9.1 LoopTemplate
-
type (string namespace)
-
required context keys
-
default risk profile
-
closure spec
-
reopen rules
-
suggested actions
-
examples:
- scheduling.email
- followup.after_meeting
- contract.signature
- invoice.paid
- pr.merged
9.2 Template rollout safety ladder
- Observe: compute “would close” only
- Recommend: suggest closure, require approval
- Enforce: auto-close only when evidence-based and stable
- Graduation gates based on metrics (false closures, reopens)
10) Integrations as sensors (not features)
10.1 Integration selection principle
- Integrations are valuable only if they emit authoritative closure signals
10.2 Integration tiers
- Tier 1: Email, Calendar
- Tier 2: Docs/Drive + e-signature, Payments/Invoices
- Tier 3: Issue tracking (GitHub/Linear/Jira), CRM
- Tier 4: External reality feeds (shipping, support tickets, domain-specific webhooks)
10.3 Signal taxonomy design
- Keep signals open-string but governed
- Canonical naming conventions
- Provenance requirements (where signal came from, when, and how derived)
11) App UX concept
11.1 Primary screen: Open Loops
-
A list of unresolved claims about reality
-
Each item shows:
- statement of outcome
- waiting-on (you/them/system/time)
- source of truth (email/calendar/etc.)
- risk indicator
11.2 Loop detail view
- Evidence view: signals that exist vs missing
- Context view: relevant thread/event/doc
- Suggested action view: what to do next
- Audit view: how decisions were made
11.3 Closure UX
- Evidence-closed: disappears automatically when conditions met
- Asserted-closed: user confirms with optional proof note/link
- Reopen UX: loop reappears with “why it reopened” and the contradictory signal
11.4 Interaction model
- No checkboxes as default behavior
- “Silence is the reward”: absence indicates truly done
- “Why is this open?” is the core tap behavior
11.5 Notifications
-
Only for:
- new high-risk loops
- loops that became at-risk
- contradictions/reopens
- approval requests (write operations)
12) System behavior and lifecycle
-
Creation
- manual creation
- inference from integrations
- template-driven creation from events (meeting ended → follow-up loop)
-
Progression
- actions proposed and executed
- state transitions based on evidence and waiting states
-
Closure
- evidence or assertion
- full record of closure proof
-
Reopen
- contradiction arrives
- re-evaluation and escalation
13) Prioritization and ranking
-
Ranking model inputs
- risk level
- person importance (later)
- time sensitivity (dueBy)
- stalling duration
- irreversibility/external visibility
-
Views
- “At Risk”
- “Waiting on Me”
- “Waiting on Others”
- “This week”
14) Trust, safety, and accountability
-
Read vs write boundaries (if actions exist)
-
HITL requirements for high-stakes actions
-
Deterministic guardrails around:
- wrong recipient/thread/account
- irreversible ops
- ambiguity resolution
-
Audit log requirements:
- what was proposed
- what was approved/edited
- what was executed
- what outcome occurred
-
Reversibility patterns:
- draft instead of send
- tentative holds instead of booking
- confirm before notifying attendees
15) Metrics that prove it works
-
Core trust metric: “If it’s not on the list, it’s actually done”
-
Operational metrics
- false closure rate (closed then reopened)
- auto-reopen rate
- asserted vs evidence closure mix
- time-to-close by loop type
- stalling rate and resolution rate
- user correction rate on actions
- notification-to-resolution conversion
-
Product metrics
- loops closed per week per user
- retention correlated with loop closure volume
- trust resets and causes
16) Expansion roadmap
- Phase 1: Email + calendar loops with rock-solid closure
- Phase 2: Templates + observe/recommend rollout
- Phase 3: Docs/signature and payments sensors
- Phase 4: Domain-specific sensors + webhooks
- Phase 5: Autonomy expansion gated by measured correctness
17) Technical architecture (high level)
-
Ingestion layer for integrations (webhooks + polling where needed)
-
Normalization layer (thread/event identity resolution)
-
Loop engine
- creation rules
- closure evaluation
- contradiction detection
- state transitions
-
Storage
- loops
- signals
- actions
- decision traces
- templates
-
Serving layer
- Open Loops views
- “why open” explanations
- approval flows if acting is supported
18) Product positioning (internal truth)
- Not a task manager
- Not a productivity dashboard
- Not a notes app
- A system for verified follow-through and real responsibility
19) Open questions to resolve early (to avoid drift)
- What counts as an authoritative signal in each integration?
- What is the minimum evidence required to close each core loop type?
- When do we allow asserted closure and what proof do we require?
- What is the default reopen window for each loop type?
- How do we communicate “asserted” vs “proven” without confusing users?
20) Canonical one-line definition
- The system that gives humans and AI a shared, provable definition of “done.”