The YOLO (You Only Look Once) series has made real-time object identification possible. The most recent version, YOLOv11, improves performance and efficiency. This article provides in-depth discussions of YOLOv11’s main advancements, parallels to earlier YOLO models, and practical uses. By comprehending its developments, we may observe why YOLOv11 is expected to become a key tool in real-time object detection.
Learning Objectives
- Understand the core principles and evolution of the YOLO object detection algorithm.
- Identify the key features and innovations introduced in YOLOv11.
- Compare YOLOv11’s performance and architecture with earlier YOLO versions.
- Explore the practical applications of YOLOv11 in various real-world scenarios.
- Learn how to implement and train a YOLOv11 model for custom object detection tasks.
This article was published as a part of the Data Science Blogathon.
What is YOLO?
It is a real-time object detection system and can also be called the family of object detection algorithms. Unlike traditional methods, which would trigger multiple passes over an image, YOLO can instantly detect objects and their locations in just one pass, resulting in something efficient for tasks that need to be done at high velocity without any compromise on accuracy. Joseph Redmon introduced YOLO in 2016, and it changed the object detection field by processing images as entire, not region-wise, which makes detections much faster while keeping a decent accuracy.
Evolution of YOLO Models
YOLO has evolved through multiple iterations, each improving upon the previous version. Here’s a quick summary:
YOLO Version | Key Features | Limitations |
---|---|---|
YOLOv1 (2016) | First real-time detection model | Struggles with small objects |
YOLOv2 (2017) | Added anchor boxes and batch normalization | Still weak in small object detection |
YOLOv3 (2018) | Multi-scale detection | Higher computational cost |
YOLOv4 (2020) | Improved speed and accuracy | Trade-offs in extreme cases |
YOLOv5 | User-friendly PyTorch implementation | Not an official release |
YOLOv6/YOLOv7 | Enhanced architecture | Incremental improvements |
YOLOv8/YOLOv9 | Better handling of dense objects | Increasing complexity |
YOLOv10 (2024) | Introduced transformers, NMS-free training | Limited scalability for edge devices |
YOLOv11 (2024) | Transformer-based, dynamic head, NMS-free training, PSA modules | Challenging scalability for highly constrained edge devices |
Each version of YOLO has brought improvements in speed, accuracy, and the ability to detect smaller objects, with YOLOv11 being the most advanced yet.
Also read: YOLO: An Ultimate Solution to Object Detection and Classification
Key Innovations in YOLOv11
YOLOv11 introduces several groundbreaking features that distinguish it from its predecessors:
- Transformer-Based Backbone: Unlike traditional CNNs, YOLOv11 uses a transformer-based backbone, which captures long-range dependencies and improves small object detection.
- Dynamic Head Design: This allows YOLOv11 to adapt based on the complexity of the image, optimizing resource allocation for faster and more efficient processing.
- NMS-Free Training: YOLOv11 replaces Non-Maximum Suppression (NMS) with a more efficient algorithm, reducing inference time while maintaining accuracy.
- Dual Label Assignment: Improves detection in overlapping and densely packed objects by using a one-to-one and one-to-many label assignment approach.
- Large Kernel Convolutions: Enables better feature extraction with fewer computational resources, enhancing the model’s overall performance.
- Partial Self-Attention (PSA): Selectively applies attention mechanisms to certain parts of the feature map, improving global representation learning without increasing computational costs.
Comparison of YOLO Models
YOLOv11 outperforms previous YOLO versions in terms of speed and accuracy, as shown in the table below:
Model | Speed (FPS) | Accuracy (mAP) | Parameters | Use Case |
---|---|---|---|---|
YOLOv3 | 30 FPS | 53.0% | 62M | Balanced performance |
YOLOv4 | 40 FPS | 55.4% | 64M | Real-time detection |
YOLOv5 | 45 FPS | 56.8% | 44M | Lightweight model |
YOLOv10 | 50 FPS | 58.2% | 48M | Edge deployment |
YOLOv11 | 60 FPS | 61.5% | 40M | Faster and more accurate |
With fewer parameters, YOLOv11 manages to improve speed and accuracy, making it ideal for a range of applications.
Also read: YOLOv7- Real-time Object Detection at its Best
Performance Benchmark
YOLOv11 demonstrates significant improvements in several performance metrics:
- Latency: 25-40% lower latency compared to YOLOv10, perfect for real-time applications.
- Accuracy: 10-15% improvement in mAP with fewer parameters.
- Speed: Capable of processing 60 frames per second, making it one of the fastest object detection models.
Model Architecture of YOLOv11
YOLOv11’s architecture integrates the following innovations:
- Transformer Backbone: Enhances the model’s ability to capture global context.
- Dynamic Head Design: Adapts processing to the complexity of each image.
- PSA Module: Boosts global representation without adding much computational cost.
- Dual Label Assignment: Improves detection of multiple overlapping objects.
This architecture allows YOLOv11 to run efficiently on high-end systems and edge devices like mobile phones.
YOLOv11 Sample Usage
Step 1: Install YOLOv11 Dependencies
First, install the necessary packages:
!pip install ultralytics
!pip install torch torchvision
Step 2: Load YOLOv11 Model
You can load the YOLOv11 pretrained model directly using the Ultralytics library.
from ultralytics import YOLO
# Load a COCO-pretrained YOLO11n model
model = YOLO('yolo11n.pt')
Step 3: Train the Model on the Dataset
Train model on your dataset with appropriate no of epochs
# Train the model on the COCO8 example dataset for 100 epochs
results = model.train(data="coco8.yaml", epochs=100, imgsz=640)
Test the model
You can save the model and test it on unseen images as required.
# Run inference on an image
results = model("path/to/your/image.png")
# Display results
results[0].show()
Original and Output image
I have unseen images to check model prediction, and it has provided the most accurate output
Applications of YOLOv11
YOLOv11’s advancements make it suitable for various real-world applications:
- Autonomous Vehicles: Improved detection of small and occluded objects enhances safety and navigation.
- Healthcare: YOLOv11’s precision helps in medical imaging tasks such as tumor detection, where accuracy is critical.
- Retail and Inventory Management: Tracks customer behaviour, monitors inventory, and enhances security in retail environments.
- Surveillance: Its speed and accuracy make it perfect for real-time surveillance and threat detection.
- Robotics: YOLOv11 enables robots to navigate environments better and interact with objects autonomously.
Conclusion
YOLOv11 sets a new standard for object detection, combining speed, accuracy, and flexibility. Its transformer-based architecture, dynamic head design, and dual label assignment allow it to excel in a range of real-time applications, from autonomous vehicles to healthcare. YOLOv11 is poised to become a critical tool for developers and researchers, paving the way for future advancements in object detection technology.
If you are looking for Generative AI course online then, explore: GenAI Pinnacle Program.
Key Takeaways
- YOLOv11 introduces a transformer-based backbone and dynamic head design, enhancing real-time object detection with improved speed and accuracy.
- It outperforms earlier YOLO models by achieving 60 FPS and a 61.5% mAP with fewer parameters, making it more efficient.
- Key innovations like NMS-free training, dual label assignment, and partial self-attention improve detection accuracy, especially for overlapping objects.
- Practical applications of YOLOv11 span across autonomous vehicles, healthcare, retail, surveillance, and robotics, benefiting from its speed and precision.
- YOLOv11 reduces latency by 25-40% compared to YOLOv10, solidifying its position as a leading tool for real-time object detection tasks.
The media shown in this article is not owned by Analytics Vidhya and is used at the Author’s discretion.
Frequently Asked Question
Ans. YOLO, or “You Only Look Once,” is a real-time object detection system that can identify objects in a single pass over an image, making it efficient and fast. It was introduced by Joseph Redmon in 2016 and revolutionized the field of object detection by processing images as a whole instead of analyzing regions separately.
Ans. YOLOv11 introduces several innovations, including a transformer-based backbone, dynamic head design, NMS-free training, dual label assignment, and partial self-attention (PSA). These features improve speed, accuracy, and efficiency, making it well-suited for real-time applications.
Ans. YOLOv11 outperforms previous versions with 60 FPS processing speed and a 61.5% mAP accuracy. It has fewer parameters (40M) compared to YOLOv10’s 48M, offering faster and more accurate object detection while maintaining efficiency.
Ans. YOLOv11 can be used in autonomous vehicles, healthcare (e.g., medical imaging), retail and inventory management, real-time surveillance, and robotics. Its speed and precision make it ideal for scenarios requiring fast and reliable object detection.
Ans. The use of a transformer-based backbone, dynamic head design that adapts to image complexity, and NMS-free training helps YOLOv11 reduce latency by 25-40% compared to YOLOv10. These improvements allow it to process up to 60 frames per second, ideal for real-time tasks.