The Intel Inside for BCI

Real-time, Explainable Neural Inference

Nimbus BCI Logo

What is a BCI?

A Direct Pathway Between Brain and Machine

A Brain-Computer Interface (BCI) is a technology that captures brain signals, analyzes them, and translates them into commands that are relayed to an output device to carry out a desired action.

Essentially, BCI allows for the direct control of computers or other devices using only thought.

Non-invasive BCI headset

Non-Invasive BCI
(e.g., EEG Headset)

Invasive BCI implant

Invasive BCI
(e.g., Neural Implant)

Slide 3: The Problem

The BCI Revolution Faces a Bottleneck

  • High Latency: A 100–300ms gap between thought and action breaks immersion and usability. At that delay, the brain perceives a disconnect — making low latency not a feature, but a hard requirement for any viable BCI product.
  • Low Adaptability: Static models can't track the brain's changing states. Retraining costs $10K–$50K per session clinically, plus weeks of engineering time — making iteration prohibitively expensive.
  • The "Black Box" Issue: Uninterpretable models block regulatory approval and erode user trust.

Together, these bottlenecks stall adoption across medicine, assistive technology, and immersive VR/AR.

Slide 4: Our Technological Core

Introducing Nimbus SDK: The Reactive Inference Engine

Nimbus SDK is a production-ready inference engine built for real-time brain signal processing. Unlike static ML models, it runs probabilistic models that continuously update as new data arrives — adapting to the brain's changing states without retraining.

At its core, Nimbus SDK uses reactive message passing on factor graphs — a technique pioneered by our team through RxInfer, an open-source Bayesian inference library developed by Lazy Dynamics, our co-founding company.

docs.nimbusbci.com →

<10ms Inference Latency

Real-time processing of EEG and neural signals with sub-10ms end-to-end latency.

Continuous Adaptation

Models update on every new sample — no retraining cycles, no downtime.

Interpretable by Design

Probabilistic outputs with uncertainty estimates — built for regulatory transparency.

RxInfer Lazy Dynamics

Slide 5: The Competitive Landscape

Why BCI Processing Methods Fall Short

Criterion LDA SVM Deep Learning Bayesian Inference on NimbusSDK
Real-time Performance High Very fast for simple, linear problems. Medium Slower inference, especially with complex kernels. Low Computationally heavy, often requires offline processing. High (Reactive) Designed for streaming data and continuous updates.
Adaptability Low A static model that must be fully retrained. Medium Can be adapted, but it's often complex and inefficient. Medium Requires large new datasets and extensive retraining. High (By Design) Continuously learns and updates from new data points.
Robustness (to noise) Low Highly sensitive to outliers and noisy signals. Medium More robust than LDA but can still be affected. Medium Can learn to ignore noise but requires vast data. High (Explicit Modeling) Models uncertainty directly, making it resilient to noise.
Interpretability High Feature contributions are clear and understandable. Medium-Low Becomes a "black box" with non-linear kernels. Very Low ("Black Box") Decision-making process is completely opaque. Very High ("White Box") The entire model structure and reasoning is transparent.

Slide 6: The Solution: RxInfer Inside

Real-time, Explainable Inference for BCI

We are building an SDK/API powered by RxInfer to serve as the core processing engine for the next generation of BCI devices.

Factor Graphs

For "white box" models that regulators and users can trust.

Variational Inference

For scalable computation balancing accuracy and speed.

Reactive Message Passing

For adaptive, real-time performance on streaming data.

The result: We reduce BCI latency from over 200ms to a target of 10-20ms.

Slide 7: Product Offering

The "Intel Inside" for Neurotechnology

  • A Core Engine: A powerful inference engine deployable in the cloud or at the edge.
  • Pre-built Models: A library of models for common BCI paradigms (e.g., motor imagery, SSVEP, P300).
  • Developer Tools: A comprehensive SDK and API for seamless integration into any neurotech stack.
Nimbus Engine

Slide 8: Visual Pipeline Builder

Nimbus Visual BCI Pipeline Generator

⚡ Key Features:

• Drag-and-drop interface • Real-time pipeline execution • Visual execution logs • No-code BCI development

Slide 9: Nimbus Studio Competitors

Why Existing Tools Fall Short

Alternative Limitation Nimbus Advantage
NeuroPype $2,000–$5,000/year, desktop-only, no embeddable SDK 10× cheaper, web-based, embeddable SDK for hardware partners
OpenBCI GUI No ML pipeline, no model training, visualization only Full ML pipeline from offline training to real-time deployment
MNE-Python Code-only, requires deep Python expertise, no visual interface Visual drag-and-drop interface accessible to non-coders
MATLAB / BCILAB Expensive license ($2K–$5K/year), legacy codebase, no cloud Modern cloud-native stack, no license fees, open integration
Custom dev team $120K+/year per developer, 12–18 months to first working prototype SDK co-development at a fraction of the cost, 6× faster to market

No existing tool combines a visual interface, Bayesian probabilistic outputs, and an embeddable hardware SDK in a single platform.

Slide 10: Target Customers

Three Validated Customer Segments

SEGMENT 1

Researchers, PhD Students & BCI Consumers

Who Postdocs, PI-led labs, enthusiasts, r/neuralinterfaces community
Channel Academic conferences, Discord, NeurotechX, r/neuralinterfaces
Price / Validation $49–$199/month ✓ 80+ on waiting list (zero marketing)
SEGMENT 2

BCI Hardware Startups

Who Miruns, BrainBit, PiEEG and similar hardware-first neurotech companies
Channel Direct outreach, BCI Society events, hardware accelerators
Price / Validation $20K–$150K contract + equity ✓ 2 signed LOIs (Miruns, BrainBit+PiEEG)
SEGMENT 3 — Q4 2026

Medical Device & Clinical Teams

Who Medical device manufacturers, hospital neuro departments, clinical research organisations
Channel Enterprise direct sales, medical device conferences, regulatory consultants
Price / Validation $6K–$60K/year Planned Q4 2026 enterprise tier

Slide 11: Market Size & Growth

A $400 Billion Market Opportunity in the US Alone

$80.8B

Early TAM

~2.8M patients with critical impairments (ALS, stroke, SCI, epilepsy, depression)

First-wave adopters — horizon: 2035

$320B

Intermediate TAM

~6.8M follow-on adopters with moderate impairments as technology matures

Follow-on wave — horizon: 2045

~$400B

Total US TAM

Combined addressable market for BCI implants in the United States

Source: Morgan Stanley BCI Primer, 2024 →

▶ Nimbus Addressable Segment — The Software Layer Inside Every BCI Device

$50–100M

SAM Today (2025)

Developer tools & pipeline software for BCI labs and researchers

$500M–$1B

SAM at 5% BCI Penetration

SDK royalties + Studio SaaS as BCI devices reach early clinical scale

$1B–$2B+

SAM at Market Maturity

Per-unit royalties across millions of deployed BCI devices globally

Slide 12: Business Model

Two Independent Revenue Streams

STREAM 1

Nimbus Studio

Freemium SaaS — researchers & BCI consumers

$49 – $199

/month per seat

  • Free tier — unlimited pipeline building, community support
  • Pro ($49/mo) — real-time execution, export, priority support
  • Team ($199/mo) — shared workspaces, advanced models, SLA

✓ 80+ on waiting list — zero paid marketing

STREAM 2

Nimbus SDK

B2B co-development — hardware startups

$20K – $150K

per contract + equity/royalties

  • 9-month co-development — full team + advisors access
  • RxInfer Pro license — embedded in partner hardware
  • Per-unit royalty — recurring revenue on every device shipped

✓ 2 signed LOIs: Miruns & BrainBit+PiEEG

Slide 13: Partnership Model

Why Partner with Nimbus?

EEG Hardware

Your specialized sensors and acquisition systems

Nimbus Software Layer

Real-time inference engine with explainable AI

Applications

Medical devices, gaming, AR/VR, assistive tech, meditation, stress monitoring

Complete Computational Expertise

No need to hire expensive AI/ML specialists. We provide the entire software stack.

Faster Time-to-Market

Launch products 6-12 months faster. Focus on hardware while we handle software.

Cheaper than In-House

Save 60-80% compared to building an internal team.

All Benefits of RxInfer

10x latency reduction, explainable AI, real-time adaptability.

Slide 14: Value Proposition

Why We Win

  • 10x Performance Boost: A massive reduction in latency creates a fundamentally better user experience.
  • Regulatory Advantage: Our "white box" approach provides the explainability required for FDA/CE approval.
  • Operational Efficiency: High adaptability eliminates the need for frequent, expensive retraining campaigns.

Slide 15: Traction

Early Validation — Before Funding

80+

Waiting List

Researchers & labs signed up with zero paid marketing

2

Signed LOIs

Miruns & BrainBit+PiEEG — hardware SDK co-development

Live Product

Nimbus Studio publicly accessible with working pipeline builder

RxInfer.jl — 500+ GitHub stars, used in 10+ academic publications

The inference engine powering Nimbus already has proven scientific traction

Advisory board with Jeff Beck (Duke), Ryan Smith (LIBR), Bert de Vries (TU/e), Alexander Kuck

World-class scientific and industry validation secured

Nimbus SDK & Nimbus Studio — production-ready

Both core products are built and available for partner integration today

Slide 16: Revenue Projections

Path to $1.1M+ ARR by Year 3

Year 1 (2025)

$20K

Studio: Free beta — 0 revenue

SDK: 1 co-dev contract ($20K)

Year 2 (2026)

$120K

Studio: 20 Pro seats — $20K/yr

SDK: 2 contracts — $100K

Year 3 (2027)

$1.1M+

Studio: 300 seats — $770K/yr

SDK: 3+ contracts — $300K+

Upside Scenario: $4.5M ARR by Year 3

If one SDK partner ships 10,000+ devices with per-unit royalty, or enterprise clinical tier launches on schedule in Q4 2026, cumulative Year 3 revenue reaches $4.5M. This is structurally similar to the Arm licensing model applied to BCI hardware.

Slide 17: Long-Term Vision

The Figma of BCI

Years 1–3 — Foundation

Establish Nimbus Studio as the standard tool for BCI research. Sign 5+ SDK hardware partners. Reach $1M+ ARR.

Years 3–5 — Scale

Launch enterprise clinical tier. Expand SDK royalty base to 50K+ devices. Build a community marketplace of BCI pipeline templates.

Years 7–10 — Platform

Become the default inference layer for BCI hardware globally — the “Intel Inside” of neurotechnology. Target: $50M+ valuation.

$50M+

Target valuation by Year 7–10

Every BCI device ships with Nimbus SDK

Researchers worldwide use Nimbus Studio

Nimbus = the standard for explainable neural inference

Slide 18: The Team

Our Founding Team & Expertise

Sergey Musienko

Sergey Musienko

CEO

Drives strategy and partnerships in deep-tech ventures.

Albert Podusenko

Albert Podusenko

CTO

Tech development & probabilistic programming.

Kirill Parinov

Kirill Parinov

CBDO

Market strategy & partnerships.

Alex Sidorov

Alex Sidorov

Chief Revenue Officer

15+ years management consulting and B2B SaaS.

İsmail Şenöz

İsmail Şenöz

Chief Scientist

Ph.D. ML from TU/e, leads scientific discovery.

Bart van Erp

Bart van Erp

Chief Product Officer

Ph.D. EE from TU/e, probabilistic programming tools.

Slide 19: Advisory Board

Guidance from World-Class Scientists

For context: Yann LeCun (ex-Meta Chief AI Scientist) averages ~50K citations/year on Google Scholar.

Duke University

Jeff Beck

Professor of Computational Neuroscience

Duke University

7,169 total citations

~360/year • h-index 28

Computational neuroscience & Bayesian brain models. Co-author of landmark paper cited 1,955×.

LIBR

Ryan Smith

Research Associate Professor

Laureate Institute for Brain Research

8,229 total citations

~1,400/year • h-index 48

Active Inference & BCI translation. Collaborates with Karl Friston (UCL).

TU/e

Bert de Vries

Professor of Signal Processing & ML

Eindhoven University of Technology

3,603 total citations

~95/year • h-index 26

Co-founder of GN Hearing (Jabra). Creator of RxInfer.jl — the core engine of Nimbus.

Medtronic

Alexander Kuck

Principal Engineer

Medtronic

Industry practitioner

Neurotechnology & medical devices

Expert in BCI hardware integration, medical device software, and neurotechnology commercialization.

Thank You

Let's build the future of neurotechnology together.