YOLO11 Detection β EdgeFirst Edge AI
NXP i.MX 8M Plus | NXP i.MX 93 | NXP i.MX 95 | NXP Ara240 | RPi5 + Hailo-8/8L | NVIDIA Jetson YOLO11 Detection models optimized for edge AI deployment across multiple hardware platforms. All sizes from Nano to XLarge, in ONNX FP32 and TFLite INT8 formats, with platform-specific compiled models for NPU acceleration.
Trained on COCO 2017 (80 classes). Part of the EdgeFirst Model Zoo.
Training session: View on EdgeFirst Studio β dataset, training config, metrics, and exported artifacts.
Newer architecture with attention blocks.
Size Comparison
All models validated on COCO val2017 (5000 images, 80 classes).
| Size | Params | GFLOPs | ONNX FP32 mAP@0.5 | ONNX FP32 mAP@0.5-0.95 | TFLite INT8 mAP@0.5 | TFLite INT8 mAP@0.5-0.95 |
|---|---|---|---|---|---|---|
| Nano | 2.6M | 6.5 | 53.4% | 37.9% | 50.1% | 34.5% |
| Small | 9.4M | 21.5 | β | β | β | β |
| Medium | 20.1M | 68.0 | β | β | β | β |
| Large | 25.3M | 87.6 | β | β | β | β |
| XLarge | 56.9M | 195.0 | β | β | β | β |
On-Target Performance
Full pipeline timing: pre-processing + inference + post-processing.
| Size | Platform | Pre-proc (ms) | Inference (ms) | Post-proc (ms) | Total (ms) | FPS |
|---|---|---|---|---|---|---|
| β | β | β | β | β | β | β |
Measured with EdgeFirst Perception stack. Timing includes full GStreamer pipeline overhead.
Downloads
ONNX FP32 β Any platform with ONNX Runtime.
TFLite INT8 β CPU or NPU via runtime delegate (i.MX 8M Plus VX Delegate).
Deploy with EdgeFirst Perception
Copy-paste GStreamer pipeline examples for each platform.
NXP i.MX 8M Plus β Camera to Detection with Vivante NPU
gst-launch-1.0 \
v4l2src device=/dev/video0 ! video/x-raw,width=640,height=480 ! \
edgefirstcameraadaptor ! \
tensor_filter framework=tensorflow-lite \
model=yolo11n-det-coco.tflite \
custom=Delegate:External,ExtDelegateLib:libvx_delegate.so ! \
edgefirstdetdecoder ! edgefirstoverlay ! waylandsink
RPi5 + Hailo-8L
gst-launch-1.0 \
v4l2src device=/dev/video0 ! video/x-raw,width=640,height=480 ! \
hailonet hef-path=yolo11n-det-coco.hailo8l.hef ! \
hailofilter function-name=yolo11_nms ! \
hailooverlay ! videoconvert ! autovideosink
NVIDIA Jetson (TensorRT)
gst-launch-1.0 \
v4l2src device=/dev/video0 ! video/x-raw,width=640,height=480 ! \
edgefirstcameraadaptor ! \
nvinfer config-file-path=yolo11n-det-coco-config.txt ! \
edgefirstdetdecoder ! edgefirstoverlay ! nveglglessink
Full pipeline documentation: EdgeFirst GStreamer Plugins
Foundation (HAL) Python Integration
from edgefirst.hal import Model, TensorImage
# Load model β metadata (labels, decoder config) is embedded in the file
model = Model("yolo11n-det-coco.tflite")
# Run inference on an image
image = TensorImage.from_file("image.jpg")
results = model.predict(image)
# Access detections
for det in results.detections:
print(f"{det.label}: {det.confidence:.2f} at {det.bbox}")
EdgeFirst HAL β Hardware abstraction layer with accelerated inference delegates.
CameraAdaptor
EdgeFirst CameraAdaptor enables training and inference directly on native sensor formats (GREY, YUYV, etc.) β skipping the ISP color conversion pipeline entirely. This reduces latency and power consumption on edge devices.
CameraAdaptor variants are included alongside baseline RGB models:
| Variant | Input Format | Use Case |
|---|---|---|
yolo11n-det-coco.onnx |
RGB (3ch) | Standard camera input |
yolo11n-det-coco-grey.onnx |
GREY (1ch) | Monochrome / IR sensors |
yolo11n-det-coco-yuyv.onnx |
YUYV (2ch) | Raw sensor bypass |
Train CameraAdaptor models with EdgeFirst Studio β the CameraAdaptor layer is automatically inserted during training.
Train Your Own with EdgeFirst Studio
Train on your own dataset with EdgeFirst Studio:
- Free tier includes YOLO training with automatic INT8 quantization and edge deployment
- Upload datasets via EdgeFirst Recorder or COCO/YOLO format
- AI-assisted annotation with auto-labeling
- CameraAdaptor integration for native sensor format training
- Deploy trained models to edge devices via EdgeFirst Client
See Also
Other models in the EdgeFirst Model Zoo:
| Model | Task | Best Nano Metric | Link |
|---|---|---|---|
| YOLOv5 Detection | Detection | 49.6% mAP@0.5 (ONNX) | EdgeFirst/yolov5-det |
| YOLOv8 Detection | Detection | 50.2% mAP@0.5 (ONNX) | EdgeFirst/yolov8-det |
| YOLOv8 Segmentation | Segmentation | 34.1% Mask mAP@0.5-0.95 (ONNX) | EdgeFirst/yolov8-seg |
| YOLO11 Segmentation | Segmentation | 35.5% Mask mAP@0.5-0.95 (ONNX) | EdgeFirst/yolo11-seg |
| YOLO26 Detection | Detection | 54.9% mAP@0.5 (ONNX) | EdgeFirst/yolo26-det |
| YOLO26 Segmentation | Segmentation | 37.0% Mask mAP@0.5-0.95 (ONNX) | EdgeFirst/yolo26-seg |
Technical Details
Quantization Pipeline
All TFLite INT8 models are produced by EdgeFirst's custom quantization pipeline (details):
- ONNX Export β Standard Ultralytics export with
simplify=True - TF-Wrapped ONNX β Box coordinates normalized to [0,1] inside DFL decode via
tf_wrapper(~1.2% better mAP than post-hoc normalization) - Split Decoder β Boxes, scores, and mask coefficients split into separate output tensors for independent INT8 quantization scales
- Smart Calibration β 500 images selected via greedy coverage maximization from COCO val2017
- Full INT8 β
uint8input (raw pixels),int8output (per-tensor scales), MLIR quantizer
Split Decoder Output Format
Detection (e.g., yolo11n):
- Boxes:
(1, 4, 8400)β normalized [0,1] coordinates - Scores:
(1, 80, 8400)β class probabilities
Each tensor has independent quantization scale and zero-point. EdgeFirst HAL handles dequantization and reassembly automatically.
Metadata
- TFLite:
edgefirst.json,labels.txt, andedgefirst.yamlembedded via ZIP (notflite-supportdependency) - ONNX:
edgefirst.jsonembedded viamodel.metadata_props
No standalone metadata files β models are self-contained.
Limitations
- COCO bias β Models trained on COCO (80 classes) inherit its biases: Western-centric scenes, specific object distributions, limited weather/lighting diversity
- INT8 accuracy loss β Full-integer quantization typically degrades mAP by 6-12% relative to FP32; actual loss depends on model architecture and dataset
- Thermal variation β On-target performance varies with device temperature; sustained inference may throttle on passively-cooled devices
- Input resolution β All models expect 640Γ640 input; other resolutions require letterboxing or may reduce accuracy
- CameraAdaptor variants β GREY/YUYV models trade color information for latency; accuracy may differ from RGB baseline depending on the task
Citation
@software{edgefirst_yolo11_det,
title = { {YOLO11 Detection β EdgeFirst Edge AI} },
author = {Au-Zone Technologies},
url = {https://huggingface.co/EdgeFirst/yolo11-det},
year = {2026},
license = {Apache-2.0},
}
EdgeFirst Studio Β· GitHub Β· Docs Β· Au-Zone Technologies
Apache 2.0 Β· Β© Au-Zone Technologies Inc.
Space using EdgeFirst/yolo11-det 1
Evaluation results
- mAP@0.5 (Nano ONNX FP32) on COCO val2017self-reported53.400
- mAP@0.5-0.95 (Nano ONNX FP32) on COCO val2017self-reported37.900
- mAP@0.5 (Nano TFLite INT8) on COCO val2017self-reported50.100
- mAP@0.5-0.95 (Nano TFLite INT8) on COCO val2017self-reported34.500