Bee Edge AI
Program the Bee camera to collect data from the physical world you want

Bee Edge AI
Program the Bee camera to collect data from the physical world you want
Build and deploy custom AI workloads on the Bee's edge computing platform. Write Python modules, push them OTA to Bee devices worldwide or your own devices, and stream results to your cloud.
Key Differentiation
Edge compute, not cloud latency — Run object detection and ML inference directly on the 5.1 TOPS NPU. Results in milliseconds, not round-trips.
Zero hardware and telecom ops — Deploy to a global fleet instantly. No procurement, no installation, no firmware headaches.
Programmable data collection — Change what you capture with a code push. New use case? New model? Deploy OTA in hours, not hardware cycles.
Pricing
See pricing here
Overview
Edge Modules are Python programs that run on the Bee camera to collect the data you need from the physical world. They have access to all onboard sensors and can run custom ML models alongside the native Map AI stack.
Capabilities
Sensors
12.3MP camera, stereo depth imagery, GPS, IMU, accelerometer
Compute
5.1 TOPS NPU, runs inference offline on-device
Deployment
OTA via Bee Maps infrastructure
Targeting
Country, state, metro, city
Output
Structured JSON, Imagery, Imagery + Depth, Video, Telemetry
Sensor Access
Edge Modules can access all onboard sensors:
Camera
12.3MP RGB frames at 30 FPS
Depth
Stereo depth imagery with distance estimation
GPS
Latitude, longitude, altitude, speed, heading
IMU
Accelerometer and gyroscope (6-axis)
Calibration
Device-specific calibration data for accurate object positioning
Running Custom Models
Run object detection and classification directly on the edge. The 5.1 TOPS NPU handles inference offline — no cloud dependency.
Detection Models
Detect and position objects in a scene: "speed limit sign at coordinates X,Y"
Classification Models
Binary or multi-class classification: "is there a baby stroller in frame?"
Geographic Targeting
Deploy modules to specific regions. Devices receive your module only when operating within targeted areas.
Country
All devices in a country
State/Province
All devices in a state or province
Metro
All devices in a metropolitan area
City
All devices in a given city (e.g., Santa Monica, CA)
Data Offload
Edge Module output streams to your cloud via Bee Connectivity Services, or upload to Bee Maps and consume via API.
Connectivity Channels
LTE
Real-time critical data, small payloads
Always on, immediate delivery
WiFi
Bulk imagery, large payloads
Batched delivery when connected
Output Configuration
Detections (JSON)
~1 KB per event
Real-time via LTE
Frame crop
~50 KB
Real-time via LTE
Full frame (12.3MP)
~2 MB
Batched via WiFi
Depth crop
Variable
Batched via WiFi
Video clip
Variable
Batched via WiFi
Deployment Workflow
Create Module — Define configuration and upload your model via Bee Maps console
Configure Output — Set your endpoint and select data types
Set Targeting — Define geographic regions
Staging Deploy — Push to a small device subset to validate accuracy
Production Deploy — Roll out to full target region via OTA
Example Use Cases
Retail & Places Churn
Monitor storefronts to detect business changes — new openings, closures, rebrands.
What you detect: "For lease" signs, changed storefront signage, boarded windows, new business openings
Output: Structured change events with imagery, fed into places databases to keep POI data fresh without manual surveys.
Complex Intersection Video
Capture video clips at specific intersections for traffic analysis, urban planning, or safety studies.
How it works: Define target intersections via GeoJSON or auto-detect based on traffic light count. Trigger recording when devices enter the zone. Collect multi-angle footage as different vehicles traverse the same intersection over time.
Output: Geotagged video clips from multiple perspectives, timestamped for temporal analysis.
Long-Tail Event Capture
Detect rare but critical events for autonomous vehicle training and safety validation.
Example events: Pedestrians with strollers, wheelchair users crossing, animals in road, unusual vehicle types, construction zone edge cases, adverse weather conditions
How it works: Run lightweight classifiers on-device. Upload only when target events are detected. Build datasets of real-world edge cases at global scale.
Output: Annotated imagery and video of rare events, with full sensor context.
World Model Training Data
Collect synchronized video + depth + IMU data for training world models and vision foundation models.
Data captured: High-resolution video (12.3MP @ 30fps), stereo depth imagery, full IMU telemetry, precise GPS positioning
Targeting options: Road types (highway, urban, rural), weather conditions, geographic regions, time of day
Output: Synchronized multimodal sensor data at scale — the raw ingredients for physical world simulations.
Developer SDK
Build and test Edge Modules locally before deployment.
Repository: github.com/Hivemapper/bee-plugins
Quick Start
Local Development
Connect to the Bee over WiFi (password: hivemapper) and iterate locally with hot-reload.
python3 devtools.py -dI
Enable dev mode (prevents OTA overwrites)
python3 devtools.py -i myplugin.py
Upload plugin (auto-restarts service)
python3 devtools.py -R
Restart plugin service
python3 devtools.py -dO
Disable dev mode
Fixture Data
Test with pre-built datasets before deploying to real devices:
python3 devtools.py -f sf
Load San Francisco fixture
python3 devtools.py -f tokyo
Load Tokyo fixture
python3 devtools.py -d
Dump cache to local machine
Device Utilities
Getting Started
Clone the SDK: github.com/Hivemapper/bee-plugins
Build and test locally with fixture data
Email us to get API credentials and deploy
Last updated
