furcate

Science

Methodology, cited end-to-end.

Furcate is a composition of open-source projects, not a new framework. Every layer of the stack maps to projects that already work at production scale; every interface is explicit. We do not lock customers into proprietary runtimes, opaque orchestrators, or vendor-only mesh protocols.

The stack

Open foundations, cited and named.

Furcate composes the best open-source projects in edge AI into one coherent fabric — never a black box. Every decision the runtime, orchestrator, or agent makes is traceable to the project that produced it.

Read the methodology

TensorRT Edge-LLM

NVIDIA

LLM + VLM inference on Jetson / DRIVE

High-performance C++ runtime for LLM and VLM inference on resource-constrained NVIDIA platforms — FP8, NVFP4, INT4 quantization, EAGLE-3 speculative decoding, KV-cache compression. Demonstrated at CES 2026 with Bosch, ThunderSoft, MediaTek partner showcases.

LiteRT

Google

Lightweight cross-platform inference

TensorFlow Lite's evolution into LiteRT — built-in quantization + compression, runs on Android, embedded Linux, microcontrollers via TFLite Micro. Default for cross-vendor mobile/edge deployments.

ONNX Runtime

Microsoft

Cross-hardware AI inference

Cross-platform inference engine optimising AI models across CPUs, GPUs, NPUs, and specialised accelerators with minimal model modification. Deployed widely as a hardware-agnostic runtime.

ExecuTorch

Meta

PyTorch on microcontrollers + mobile

Bytecode VM with AOT compilation for PyTorch models — built for microcontrollers and embedded edge devices. Pairs with NVIDIA FLARE for federated fine-tuning on mobile.

OpenVINO

Intel

Intel hardware-tuned inference

Optimised for Intel CPUs / GPUs / VPUs / FPGAs. Strong for industrial IoT vision (smart cameras, intelligent retail) and any deployment that's standardised on Intel silicon.

NVIDIA FLARE

NVIDIA

Production federated-learning runtime

Domain-agnostic SDK for federated learning. Hierarchical FL architecture for thousands of edge devices. Production deployments include Eli Lilly TuneLab, Taiwan MOHW national healthcare FL, and a Tri-Labs (Sandia / LANL / LLNL) federated AI pilot. Integrates with Flower; pairs with ExecuTorch for mobile FL.

Flower

Flower Labs

Open FL framework + community

Cohesive approach to federated learning, analytics, and evaluation. Strong research community + extensive strategy library. Interoperates with FLARE so Flower-built apps run inside the FLARE runtime without modification.

OpenFL

Intel

Intel federated learning

Open-source FL implementation focused on sensitive-data deployments. Used in healthcare and regulated-industry pilots.

KubeEdge

CNCF / open-source

Kubernetes for edge devices

CNCF incubation project. Scales to 100,000 concurrent edge nodes managing 1,000,000+ active pods. The default when a customer's edge fleet is large enough to need cloud-native ops.

OpenYurt

Alibaba / open-source

Edge-native K8s with offline ops

Brings edge computing capabilities to Kubernetes — edge nodes can run K8s without continuous cloud connectivity. Strong choice for intermittent / offline edge.

K3s

Rancher / SUSE

Lightweight Kubernetes

Optimised K8s with significantly reduced memory footprint — full K8s experience in resource-constrained environments. Lowest resource consumption among lightweight K8s distributions in 2026 benchmarks.

Akri

open-source

K8s leaf-device discovery + lifecycle

Built on Kubernetes Device Plugins; discovers small edge devices via ONVIF / udev / OPC UA handlers. Creates K8s services per device with HA when nodes lose network or fail.

EdgeX Foundry

LF Edge

Vendor-neutral edge IoT framework

LF Edge open-source edge platform. Modular reference services for device-data ingestion, normalisation, analysis, and sharing. The default integration substrate when a customer's deployment spans many vendor protocols.

Dapr

CNCF / open-source

Distributed application runtime

Portable runtime for distributed applications across cloud and edge. Built-in workflow, pub/sub, state, secrets, bindings, actors, distributed lock, cryptography. State of Dapr 2026 reports 20-40% developer productivity uplift.

Eclipse Hono

Eclipse Foundation

Multi-protocol IoT messaging

Uniform API surface over MQTT, AMQP, HTTP, LoRaWAN — abstracts the protocol mess so application code can stay protocol-agnostic.

Wasmtime

Bytecode Alliance

Standalone WebAssembly runtime

Bytecode Alliance reference runtime — leads in cold-start performance among JIT/AOT compilers (Jan 2026). 1-5 ms cold starts vs 100ms-1s+ for Linux containers — the 100x improvement that lets edge inference become serverless.

WasmEdge

CNCF / open-source

Edge-optimised WebAssembly runtime

Lightweight, high-performance WASM runtime for cloud-native, edge, and decentralised apps. Powers serverless apps, embedded functions, microservices, smart contracts, IoT devices. ~100x faster startup and ~1/100 the size of equivalent Linux containers.

Matter / Thread

Connectivity Standards Alliance

Residential interoperability

Matter 1.4 (November 2024) added energy-management device classes. Thread provides the IPv6 mesh underneath. The default for Furcate's residential / SMB device class.

LoRaWAN

LoRa Alliance

Long-range low-power IoT

Sub-GHz mesh for outdoor / industrial / agricultural deployments where Wi-Fi and cellular don't reach. Ultra-low-power, multi-km range, kilobit-class throughput.

Private 5G + TSN

3GPP + IEEE

Industrial-grade wireless

Private 5G slices + Time-Sensitive Networking (TSN) for deterministic latency on the factory floor. The wireless backbone Furcate's industrial customers run on.

TPM 2.0 + TEE

Trusted Computing Group + ARM + Intel

Hardware root of trust

Trusted Platform Module 2.0 + Trusted Execution Environment (Intel SGX, ARM TrustZone) for secure boot, attested device identity, and confidential inference. The hardware backstop for Furcate's sovereign-by-design posture.

ROS 2

Open Source Robotics Foundation

Robot middleware

Robot Operating System 2 — DDS-based publish/subscribe middleware for cobots, AMRs, manipulators, drones. Native integration so Industry's robotic workflows speak the same protocol as the rest of the fleet.

NVIDIA Triton + vLLM + Ray Serve

NVIDIA / Anyscale / open-source

Distributed model serving

Triton for general inference serving, vLLM for LLM throughput, Ray Serve for distributed Python services. The serving substrate behind Furcate's edge gateways.

Benchmarks

Documented in the field.

Edge AI silicon: Jetson Orin Nano Super delivers 67 TOPS at $249 and 7-25 W (NVIDIA, 2025). The Hailo-10H AI HAT+ 2 lifts a Raspberry Pi 5 to 40 TOPS INT4 at 2.5 W and runs 2B-parameter LLMs at ~10 tok/s for $130 (Hailo, 2025). Google Coral Edge TPU sits at 4 TOPS / 2 W for quantised TFLite vision workloads.

WebAssembly cold start: Wasmtime leads JIT/AOT cold-start performance among standalone WASM runtimes (January 2026 benchmarks). 1-5 ms WASI cold starts vs 100ms-1s+ for traditional Linux containers — a 100× improvement that lets edge inference become serverless. Cloudflare Workers runs ~10M WebAssembly requests per second across 300+ edge locations.

Kubernetes-at-edge: KubeEdge published scaling to 100,000 concurrent edge nodes managing 1,000,000+ active pods. K3s exhibits the lowest resource consumption among lightweight K8s distributions; OpenYurt offers the strongest offline-capable edge story. Akri extends K8s-native device discovery to OPC UA / udev / ONVIF leaf devices.

Federated learning in production: NVIDIA FLARE deploys hierarchical FL across thousands of edge devices. Real production deployments include Eli Lilly TuneLab (built by Rhino Federated Computing on FLARE), Taiwan MOHW national healthcare FL, and a Tri-Labs (Sandia/LANL/LLNL) federated AI pilot. Flower interoperates natively with FLARE; ExecuTorch handles mobile-side fine tuning.

Sovereign edge inference: Microsoft Sovereign Private Cloud scales to thousands of nodes per sovereign environment (announced April 27, 2026). Azure Local Disconnected Operations enables fully air-gapped deployment with consistent management UX. HPE Private Cloud AI offers turnkey air-gapped AI training and inference.

Standards

Native to the protocols of the edge.

Edge AI runs on a thicket of standards that span radio, security, orchestration, and ML frameworks. A fabric that doesn’t speak them natively is doomed to integration debt. We speak them natively.

  • Matter 1.4 + Thread

    Energy-class device support, IPv6 mesh

  • LoRaWAN

    Long-range low-power IoT

  • Private 5G + TSN

    Deterministic latency wireless

  • MQTT 5 + Sparkplug B

    Pub/sub bus, ISA-95 namespace

  • OPC UA + DDS + ROS 2

    Industrial + robotics middleware

  • TPM 2.0 + TEE

    Hardware root of trust, secure boot

  • WASI 0.3 / 1.0

    WASM systems interface (Feb 2026 / late 2026)

  • IEC 62443

    Industrial cybersecurity zones-and-conduits

  • NIS2 (EU mandatory)

    Critical-infra cybersecurity

  • FIPS 140-3 + NIST SP 800-series

    Cryptography compliance

  • GDPR + HIPAA + CMMC

    Data-residency + handling regimes

  • ONNX + GGUF + GGML

    Model interchange formats

Selected references

Where the work comes from.

  • TensorRT Edge-LLM: Accelerating LLM and VLM Inference for Automotive and Robotics

    NVIDIA Technical Blog, CES 2026

  • NVIDIA FLARE: Federated Learning from Simulation to Real-World

    Roth et al., arXiv:2210.13291

  • Supercharging Federated Learning with Flower and NVIDIA FLARE

    arXiv:2407.00031

  • KubeEdge: Performance Test — Scaling to 100,000 Edge Nodes

    CNCF / KubeEdge community

  • WASI 0.3.0 Release — WebAssembly Replaces Containers for Edge

    Bytecode Alliance, February 2026

  • Effortless Federated Learning on Mobile with NVIDIA FLARE and Meta ExecuTorch

    NVIDIA Technical Blog 2025

  • Microsoft Sovereign Private Cloud Scales to Thousands of Nodes

    Microsoft Official Blog, April 2026

  • Build sovereign AI at the edge with Azure Local

    Microsoft Azure Blog 2026

  • 8 CNCF Tools to Run Kubernetes at the Edge and Bare Metal

    Cloud Native Now, 2026

  • Comparative Analysis of Lightweight Kubernetes Distributions for Edge Computing

    Springer Nature 2024

  • WebAssembly Runtime Benchmarks 2026: Wasmtime vs Wasmer vs WasmEdge

    wasmRuntime.com

  • TinyML on ESP32 with TensorFlow Lite Micro

    Hackster / EloquentArduino