
SECA targets sub-7B parameter models running directly on Snapdragon · Dragonwing edge AI hardware. Achieving useful latency on a workstation-class NPU without compromising hardware-domain reasoning depends on KV cache compression, retrieval-based context packing and attention-aware token eviction. The visualization below shows the same prompt running with and without these optimizations.
For the exact same artifacts (Ver D spec, BOM, datasheets), a generic 70B cloud LLM lacks hardware-domain priors and either drops or hallucinates the revision-critical fields. SECA's domain-tuned 7B + retrieval recovers all five Ver D deltas. Wrong fields on the left are highlighted in red.
The mock folder contains a real customer cluster system specification (Ver C / Ver D), datasheets, a power-domain block diagram and a BOM. Customer-identifying information is masked for the demo. In production, every artifact stays on the engineer's workstation or on a Snapdragon / Dragonwing edge box — none of it is sent to a cloud LLM.
Drop files from demo_dump/ (or mock/) here — datasheets, BOM, cluster spec, BD — and click Run. SECA's on-device pipeline normalizes each artifact into hardware-domain JSON below, matched by filename. The accuracy comparison vs a generic cloud LLM is shown in 00 TECHNOLOGY.
Generic LLMs emit free-form JSON. To build a knowledge graph we constrain extraction to a hardware ontology — entity types like Spec, Component, BomLine, Datasheet, BdBlock and typed predicates like requires, has_attr, instance_of, contradicts. Each JSON field is matched against this schema and rewritten as a triple before it is added to the graph.
Each artifact contributes nodes (Spec, Component, BomLine, BdBlock, DsAttr) and edges (requires, instance_of, has_attr, driven_by, contradicts). The single view animates the staged build of one revision; revision-compare runs two cytoscape instances side-by-side so you can see exactly which nodes/edges were added, changed or now contradict the spec when going Ver C → Ver D.
The moment requirements move from Ver C to Ver D, SECA re-evaluates the affected slice of the graph and surfaces five cross-artifact conflicts. Each finding shows which two artifacts disagree, on which field, with what evidence — so the engineer can fix it before tape-out, not after a USD 50K–500K re-spin.
SECA's Value 1 stops here. 5 cross-artifact conflicts are surfaced with traceable evidence; 3 are introduced by the Ver D revision and 2 were latent in Ver C and likely to be missed in manual review. Every finding is evidence-grounded — engineers do the deciding, not the searching.
Beyond hard 1:1 mismatches, the scorer model proposes a ranked list of top-K candidate conflicts. The engineer's confirm / reject signals feed two parallel updates: ① the scorer is updated with RLHF so next round's ranking improves, and ② the Knowledge Graph acts as an external memory network — confirmed patterns become permanent rules edited directly into the graph.
Latent conflicts the model is unsure about. Confirm or reject each — your decisions drive both updates below.
The accumulated KG is no longer just a consistency-checking artifact. It becomes the external context / feature memory for a hardware-design agent. Type a question or pick one from the side — the agent decomposes it into sub-queries, traverses the KG, packs the relevant slice as context, and a local SLM streams an evidence-grounded answer. Every step is shown explicitly.