The SatCam System captures and reasons on the interview at the edge. The SatCam Cloud turns every session into compounding forensic intelligence, with governance and automation for federal-grade workflows.
A synchronized depth + vision + audio + biosignal capture system with on-device AI compute. The full credibility pipeline runs on the device — including the Cortex reasoner — with a 3-second end-to-end delay. No network required.
Every stream is time-stamped to correlate every facial micro-movement with a syllable of speech and the beat frequency of the heart.
| Sensor | Purpose |
|---|---|
| ToF Depth Indirect time-of-flight | Sub-mm facial muscle movement @ 0.5 m |
| RGB Global Shutter High-resolution colour | Face, landmarks, pose, action-unit classification |
| mmWave Radar Millimetre-wave | Non-contact heart & respiration rate |
| Beamforming Mic Array Multi-element | Transcript, prosody, turn-taking |
An indirect time-of-flight camera captures facial muscle displacement with sub-millimeter resolution at 0.5 m — the only way to pick up suppression, lid-tightening, and dimpler cues that never show on RGB alone.
A global-shutter colour sensor produces distortion-free frames for face detection, landmarking, pose estimation, and action-unit classification. Global shutter preserves geometry during fast motion.
Multi-element array with spatial beamforming isolates the subject's voice from room noise. Downstream: prosody features, turn-taking, speaker identification, and time-stamped transcript.
Millimetre-wave radar resolves chest-wall motion for respiration rate. Paired with rPPG from the RGB feed, the system produces contactless heart rate — no chest strap, no electrodes.
Four near-IR VCSEL emitters flood the subject with invisible infrared for ToF operation — Class-1 rated, so the subject perceives no light, no laser, no discomfort.
Hardware-anchored time stamps across all sensor streams. Downstream fusion can correlate a 40–200 ms action unit with a syllable and a heartbeat — frame-for-frame, no drift.
Face detection, landmark localization, pose estimation for frontalization, image normalization, and an action-unit classifier covering the full set of decision-relevant facial cues.
Multi-language speech-to-text, prosodic feature extraction, and speaker-embedding identification — all time-aligned with the vision chain.
Action units, prosody, transcript, and biosignals converge at the Cortex reasoner running locally on the on-device NPU — returning a score and a rationale every ~3 seconds.
Budgeted to preserve natural turn-taking in a live interview. The interviewer sees the score and next-question suggestion before they would have formulated their follow-up.
Operate fully offline — sensors, processing, inference, and the interviewer UI all run locally. Recordings and metadata are written only to the local NVMe SSD. Classified-environment ready.
Cortex attaches the biometric and linguistic evidence that drove each credibility score — required for 2026 AML governance and federal audit trails, exported with the session.
Cortex improves in real time — no retraining cycle, no catastrophic forgetting. Model updates propagate in-place; the device you field in 2028 will outperform the one shipped in 2027.
The speech chain supports multiple languages out of the box. Cortex adapts to new demographics and cohorts on-the-fly without a retraining cycle.
Every session bundle is cryptographically signed and encrypted at rest. Chain-of-custody metadata is part of the export — tampering is detectable.
Live interview signals and transcripts flow into the SatCam Cloud, where Cortex turns them into searchable evidence, cross-session correlation, automated reports, and governance artifacts — all traceable to the biometric signals that produced each claim.
We are running capability demonstrations with customers and launch partners in September 2026.