Skip to content
scsiwyg
sign insign up
get startedmcpcommunityapiplaygroundswaggersign insign up
Wiki/Intelligence Layers/

Context Intelligence

High confidence

Updated 29 Apr 2026 by David Olsson

Context Intelligence

Layer: Context | Phase: Discovery & Diagnosis (Weeks 0-2) | Status: Enriched from Daanaa-supplied PDP artifacts

This layer answers: what is the urgency, who matters, where do we enter, what are the constraints, is this feasible?


Urgency & Risk Drivers

What is forcing this conversation now? What is the cost of inaction?

  • ISO compliance โ€” Daanaa is "on the path" (Udi's email). External audit timing TBD; cost of poor traceability surfaces at audit.
  • Documentation as productivity tax โ€” explicitly raised by Udi as "necessary evil dragging productivity."
  • IP control non-negotiable โ€” power electronics IP, customer NDAs, foreground/background classification needs.
  • Scale stress โ€” 100+ engineer org; documentation cost compounds with growth.

Stakeholder Map

Known so far:

  • Udi Daon (CEO, Daanaa) โ€” initial contact; April 2026 meeting; receptive.
  • TBD: CTO / Eng VP / IP officer / CFO โ€” to be identified at initial meeting.

Org Entry Points

  • Primary: Udi (CEO sponsorship)
  • Pilot champion: TBD (must be identified โ€” gate-1 KPI)
  • Co-build engineers: 1-2 to be designated at pilot scoping (gate-2)
  • Audit/IP gatekeepers: to be mapped during discovery (CTO, IP officer, legal)

Constraint Envelope

The fixed lines we can't cross:

  • IP control โ€” data residency rules TBD; AI/LLM access rules TBD; default: data lives in Daanaa environment, no third-party LLM hops on confidential content
  • Tooling stack โ€” Jira, Smartsheet, timesheet workbook all in active use; we read/write, do not replace. Outsourcing to Neuronics mentioned for non-critical implementation/testing roles
  • PDP โ€” 10-phase process (Innovation through EOL). Each phase has Objectives / Activities / Outputs / RACI. Sound structure; preserved, not redesigned
  • Gate reviews โ€” Development Approval Request (DAR) is the canonical gate-review pack. Electra DAR example shows combined Concept + Planning presentation. Approval chain: Udi, Raheem, Robert
  • Reporting โ€” Electra Reporting workbook: rows per deliverable/task with hours entered vs hours planned, delta column. ~40 timesheet categories
  • ISO standards in scope โ€” 9001 confirmed; 26262 / AEC-Q100 to confirm (automotive customers Mercedes, BorgWarner)

Feasibility View

  • Conceptual fit: Strong โ€” translation tax across artifact boundaries is well-evidenced. The Electra Reporting workbook is exactly the "hand-consolidated planned-vs-actual" pattern the cockpit replaces. The DAR deck is hand-assembled from data that could be substrate-derived
  • Cultural fit: Encouraging โ€” PDP deck says "this process is being developed, and we will evolve it with collective input"
  • Technical fit: Likely good โ€” substrate adapters exist for Daanaa's tools; integration is read/write, not replace
  • Commercial fit: TBD โ€” depends on pilot envelope being acceptable at gate-2

Daanaa Project Portfolio

P1 (must do): Mercedes (POC3/Furud), BorgWarner (POC894)

P2 (funded, committed): CTC (Orion), Heliene Rev A (Zodiac), Midwest (Zodiac), GAF (Zodiac)

P3 (not funded): Mercedes (Dual DC/DC), Heliene Rev B, KDDI

Internal/research: Electra (Sirius Eval System), Maia, Taurus, Atlas SOM variants

Open Questions

  • ISO audit timeline & standards in scope (26262 / AEC-Q100)
  • Default IP classification scheme
  • Hosting & AI/LLM constraints
  • Champion identification (candidates: Jim, Mark, Ehsan, Javad)
  • Cohort size for pattern fluency
  • Document/product numbering databases (referenced but not provided)

Sources

  • intelligence/context.md
  • PDP Master Deck (48 slides)
  • Electra DAR
  • Project Priorities and Good Practices

See also