falcon-perception / README.md
shreyask's picture
readme: clickable model badge, sharper title + gradient
42e1de5 verified
metadata
title: Falcon-Perception-0.6B WebGPU
emoji: πŸ¦…
colorFrom: indigo
colorTo: pink
sdk: static
pinned: false
license: apache-2.0
short_description: Open-vocab detection + segmentation, all in the browser
models:
  - tiiuae/Falcon-Perception
  - onnx-community/falcon-perception-onnx-webgpu

πŸ¦… Falcon-Perception-0.6B WebGPU

A browser demo for tiiuae/Falcon-Perception β€” a 0.6B open-vocabulary VLM that turns natural-language queries into bounding boxes and pixel-accurate segmentation masks, running fully client-side via WebGPU + ONNX Runtime Web.

Model Weights

What's inside

  • Detection β€” draw bounding boxes for any natural-language query ("athletes", "the runner in front", "mangoes").
  • Segmentation β€” pixel-accurate masks via the AnyUp upsampler, all in-browser.
  • Tracker (preview) β€” HUD-style reticles on video. Limited by VLM latency between detections; see the in-space disclaimer.

How it runs

2.4 GB of ONNX weights are fetched once on first visit, then cached by your browser β€” no backend, no API keys, no network round-trip after load. Multi-threaded WASM is enabled via coi-serviceworker.