Blazing-fast, production-ready YOLO inference for .NET
YoloDotNet is a modular, lightweight C# library for real-time computer vision and YOLO-based inference in .NET.
It provides high-performance inference for modern YOLO model families (YOLOv5u through YOLOv26, YOLO-World, YOLO-E, and RT-DETR), with explicit control over execution, memory, and preprocessing.
Built on .NET 8, ONNX Runtime, and SkiaSharp, YoloDotNet intentionally
avoids heavy computer vision frameworks such as OpenCV.
There is no Python runtime, no hidden preprocessing, and no implicit behavior —
only the components required for fast, predictable inference on Windows,
Linux, and macOS.
No Python. No magic. Just fast, deterministic YOLO — done properly for .NET.
YoloDotNet is designed for developers who need:
- ✅ Pure .NET — no Python runtime, no scripts
- ✅ Real performance — CPU, CUDA / TensorRT, OpenVINO, CoreML, DirectML
- ✅ Explicit configuration — predictable accuracy and memory usage
- ✅ Production readiness — engine caching, long-running stability
- ✅ Large image support — not limited to toy resolutions
- ✅ Multiple vision tasks — detection, OBB, segmentation, pose, classification
Ideal for desktop apps, backend services, and real-time vision pipelines that require deterministic behavior and full control.
- Added support for
Yolo26model suite - Added support for
RT-DETRmodels - Improved performance across all tasks, with reduced allocation pressure and lower per-frame latency.
- Improved performance on video inference
- Relicensed to MIT
📖 Full release history: CHANGELOG.md
dotnet add package YoloDotNet# CPU (recommended starting point)
dotnet add package YoloDotNet.ExecutionProvider.Cpu
# Hardware-accelerated execution (choose one)
dotnet add package YoloDotNet.ExecutionProvider.Cuda
dotnet add package YoloDotNet.ExecutionProvider.OpenVino
dotnet add package YoloDotNet.ExecutionProvider.CoreML
dotnet add package YoloDotNet.ExecutionProvider.DirectML💡 Note: The CUDA execution provider includes optional TensorRT acceleration.
No separate TensorRT package is required.
using SkiaSharp;
using YoloDotNet;
using YoloDotNet.ExecutionProvider.Cpu;
using var yolo = new Yolo(new YoloOptions
{
ExecutionProvider = new CpuExecutionProvider("model.onnx")
});
using var image = SKBitmap.Decode("image.jpg");
var results = yolo.RunObjectDetection(image, confidence: 0.25, iou: 0.7);
image.Draw(results);
image.Save("result.jpg");You’re now running YOLO inference in pure C#.
YOLO inference accuracy is not automatic.
Preprocessing settings such as image resize mode, sampling method, and confidence/IoU thresholds must match how the model was trained.
These settings directly control the accuracy–performance tradeoff and should be treated as part of the model itself.
📖 Before tuning models or comparing results, read:
👉 Accuracy & Configuration Guide
| Classification | Object Detection | OBB Detection | Segmentation | Pose Estimation |
|---|---|---|---|---|
![]() |
![]() |
![]() |
![]() |
![]() |
| pexels.com | pexels.com | pexels.com | pexels.com | pexels.com |
Hands-on examples are available in the demo folder:
Includes image inference, video streams, GPU acceleration, segmentation, and large-image workflows.
| Provider | Windows | Linux | macOS | Documentation |
|---|---|---|---|---|
| CPU | ✅ | ✅ | ✅ | CPU README |
| CUDA / TensorRT | ✅ | ✅ | ❌ | CUDA README |
| OpenVINO | ✅ | ✅ | ❌ | OpenVINO README |
| CoreML | ❌ | ❌ | ✅ | CoreML README |
| DirectML | ✅ | ❌ | ❌ | DirectML README |
ℹ️ Only one execution provider package may be referenced.
Mixing providers will cause native runtime conflicts.
YoloDotNet focuses on stable, low-overhead inference where runtime cost is dominated by the execution provider and model.
📊 Benchmarks: /test/YoloDotNet.Benchmarks
- Stable latency after warm-up
- Clean scaling from CPU → GPU → TensorRT
- Predictable allocation behavior
- Suitable for real-time and long-running services
- Core package is provider-agnostic
- Execution providers are separate NuGet packages
- Native ONNX Runtime dependencies are isolated
Why this matters: fewer conflicts, predictable deployment, and production-safe behavior.
⭐ Star the repo
💬 Share feedback
🤝 Sponsor development
MIT License
Copyright (c) Niklas Swärd
Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
YoloDotNet is designed as a low-level, production-grade YOLO inference engine. It does not include pretrained models, training pipelines, or hosted services.
The MIT license was chosen to maximize freedom and clarity for developers and organizations:
- No copyleft requirements
- No network-use clauses
- No restrictions on commercial or proprietary deployment
This makes YoloDotNet suitable for use in enterprise software, embedded systems, backend services, and closed-source products without licensing friction.
Model licensing is entirely separate and is determined by the source and terms of the ONNX models supplied by the user.
-
YoloDotNet is licensed under the MIT License and provides an ONNX inference engine for YOLO models exported using Ultralytics YOLO tooling.
-
This project does not include, distribute, download, or bundle any pretrained models.
-
Users must supply their own ONNX models.
-
YOLO ONNX models produced using Ultralytics tooling are typically licensed under AGPL-3.0 or a separate commercial license from Ultralytics.
-
YoloDotNet does not impose, modify, or transfer any license terms related to user-supplied models.
-
Users are solely responsible for ensuring that their use of any model complies with the applicable license terms, including requirements related to commercial use, distribution, or network deployment.




