furcate

Solutions · Builders

From a Pi on your desk to your first thousand devices.

Developers shipping edge AI shouldn't have to assemble a stack from a dozen GitHub repos. Furcate ships an SDK, REST + gRPC + MQTT APIs, a Wasmtime / WasmEdge runtime, examples for Pi / Jetson / ESP32 / Coral / Hailo, and the orchestration plane for taking your prototype to production.

WASM cold start · 100× containers

1-5 ms

Jetson Orin Nano Super dev kit

67 TOPS

API surface

REST + gRPC + MQTT

OpenVLA, Octo, Flower

Apache 2.0

Use cases

What the platform actually does, here.

Prototype → first deployment

Start on a Pi 5 with Hailo-10H AI HAT+ 2 ($130) or Jetson Orin Nano Super ($249). Furcate SDK ships a working example for Qwen3-VL or Pixtral inference, OpenVLA for cobot tasks, LiteRT for vision, TFLM for sensor anomaly. Iterate on the device; deploy to a fleet when it's ready.

Bring your own model

ONNX, GGUF, GGML, SavedModel, TorchScript — Furcate's quantization pipeline takes a foundation-model checkpoint and produces edge-ready binaries with documented quality bounds. Lineage tracked from source weights to deployed binary.

Multi-modal applications

Vision + audio + time-series + text + telemetry, all addressable through the same SDK. Compose multi-model agents on-device — Qwen3-VL for vision, PANNs for acoustic, TimesFM for time-series — and let the runtime route inference to the right silicon.

Federated fine-tuning

When user data shouldn't leave the device, ExecuTorch + NVIDIA FLARE handle on-device fine-tuning and aggregate model deltas centrally. The pattern Meta + NVIDIA shipped in 2025 — adopt it via Furcate's SDK without reimplementing the protocol.

Wasm serverless edge

Wrap your application logic as WASI 0.3 components. Cold-start is 1-5 ms vs 100ms-1s+ for containers. Cloudflare-class scale (10M+ Wasm requests/sec across 300+ edge locations) is the proof-point; Furcate brings that pattern to your fleet.

Local development → production parity

The same SDK runs on your laptop, on a Pi on your desk, and on the customer's air-gapped industrial gateway. Same APIs, same observability, same provenance trail. No surprises in production.

How a deployment runs

Day 0 → Day 30 → Day 90.

  1. 01

    Day 0: Pull the Furcate SDK. Run the on-device example for your hardware (Pi 5 + Hailo, Jetson Orin Nano, or ESP32). One npm-equivalent command.

  2. 02

    Day 1-7: Adapt the example to your model and data. Iterate on a single device.

  3. 03

    Day 8-30: Move to a small pilot fleet (5-50 devices). Furcate handles enrolment, OTA, observability, and the dispatch guard. Measure inference latency, uplink reliability, and battery life.

  4. 04

    Day 30+: Production. KubeEdge or OpenYurt orchestration. Canary + TPM gating on rollouts. Replay-grade decision logs. Federated learning when you're ready.

  5. 05

    Day 90: Your first thousand devices in production with auditable provenance, sovereign-by-default operation, and a roadmap to the first 100k.

Stack active in this configuration

  • Furcate SDK
  • WasmEdge / Wasmtime
  • TensorRT Edge-LLM
  • LiteRT
  • OpenVLA
  • Octo
  • NVIDIA FLARE
  • Gemini 3.1 Flash Lite