Solutions · Builders
From a Pi on your desk to your first thousand devices.
Developers shipping edge AI shouldn't have to assemble a stack from a dozen GitHub repos. Furcate ships an SDK, REST + gRPC + MQTT APIs, a Wasmtime / WasmEdge runtime, examples for Pi / Jetson / ESP32 / Coral / Hailo, and the orchestration plane for taking your prototype to production.
WASM cold start · 100× containers
1-5 ms
Jetson Orin Nano Super dev kit
67 TOPS
API surface
REST + gRPC + MQTT
OpenVLA, Octo, Flower
Apache 2.0
Use cases
What the platform actually does, here.
Prototype → first deployment
Start on a Pi 5 with Hailo-10H AI HAT+ 2 ($130) or Jetson Orin Nano Super ($249). Furcate SDK ships a working example for Qwen3-VL or Pixtral inference, OpenVLA for cobot tasks, LiteRT for vision, TFLM for sensor anomaly. Iterate on the device; deploy to a fleet when it's ready.
Bring your own model
ONNX, GGUF, GGML, SavedModel, TorchScript — Furcate's quantization pipeline takes a foundation-model checkpoint and produces edge-ready binaries with documented quality bounds. Lineage tracked from source weights to deployed binary.
Multi-modal applications
Vision + audio + time-series + text + telemetry, all addressable through the same SDK. Compose multi-model agents on-device — Qwen3-VL for vision, PANNs for acoustic, TimesFM for time-series — and let the runtime route inference to the right silicon.
Federated fine-tuning
When user data shouldn't leave the device, ExecuTorch + NVIDIA FLARE handle on-device fine-tuning and aggregate model deltas centrally. The pattern Meta + NVIDIA shipped in 2025 — adopt it via Furcate's SDK without reimplementing the protocol.
Wasm serverless edge
Wrap your application logic as WASI 0.3 components. Cold-start is 1-5 ms vs 100ms-1s+ for containers. Cloudflare-class scale (10M+ Wasm requests/sec across 300+ edge locations) is the proof-point; Furcate brings that pattern to your fleet.
Local development → production parity
The same SDK runs on your laptop, on a Pi on your desk, and on the customer's air-gapped industrial gateway. Same APIs, same observability, same provenance trail. No surprises in production.
How a deployment runs
Day 0 → Day 30 → Day 90.
- 01
Day 0: Pull the Furcate SDK. Run the on-device example for your hardware (Pi 5 + Hailo, Jetson Orin Nano, or ESP32). One npm-equivalent command.
- 02
Day 1-7: Adapt the example to your model and data. Iterate on a single device.
- 03
Day 8-30: Move to a small pilot fleet (5-50 devices). Furcate handles enrolment, OTA, observability, and the dispatch guard. Measure inference latency, uplink reliability, and battery life.
- 04
Day 30+: Production. KubeEdge or OpenYurt orchestration. Canary + TPM gating on rollouts. Replay-grade decision logs. Federated learning when you're ready.
- 05
Day 90: Your first thousand devices in production with auditable provenance, sovereign-by-default operation, and a roadmap to the first 100k.
Stack active in this configuration
- Furcate SDK
- WasmEdge / Wasmtime
- TensorRT Edge-LLM
- LiteRT
- OpenVLA
- Octo
- NVIDIA FLARE
- Gemini 3.1 Flash Lite