JUHE API Marketplace
DATASET
Open Source Community

DOTA v1.5

The DOTA v1.5 dataset is designed for object detection in aerial and satellite imagery and supports rotated bounding boxes. It is compatible with YOLOv9 models.

Updated 7/15/2024
github

Description

Aerial‑YOLO‑DOTA: Advanced Aerial Image Object Detection

Dataset

We prepared an easy‑to‑use comprehensive DOTA v1.5 dataset, including YOLOv9‑compatible labels. The dataset can be downloaded from our Google Drive.

Experiments and Analysis

100e‑16b‑0.01lr (Best‑performing Model)

Normalized Confusion Matrix

The normalized confusion matrix of the best model reveals several interesting patterns:

  1. Strong diagonal performance: Many classes exhibit high true‑positive rates, evident from deep‑blue diagonal cells. Tennis court (0.89), airplane (0.86) and basketball court (0.69) are especially well classified.
  2. Consistent mis‑classifications: Off‑diagonal errors make intuitive sense in terms of visual and contextual similarity:
    • Ships are occasionally mis‑classified as ports (0.20), reflecting their co‑occurrence.
    • Small vehicles sometimes are labeled as large vehicles (0.17), likely due to scale ambiguity in aerial views.
    • Some confusion between airplanes and helicopters (helicopter → airplane 0.29) stems from shared aerial features.
  3. Sports‑facility confusion: Similar sports venues are sometimes mixed, e.g., football fields being labeled as other field types.
  4. Background interactions: The “background” class strongly interacts with bridge (0.84) and container crane (1.00), indicating difficulty separating these structures from background in certain scenes.
  5. No severe errors: Notably, there are no extreme mis‑classifications such as airplanes being labeled as ships, indicating the model has learned discriminative features.
  6. Challenging classes: Certain categories like “roundabout” (0.33) and “small vehicle” (0.26) show lower true‑positive rates, suggesting they could benefit from additional training data or feature engineering.

Overall, the confusion matrix reflects a model that has learned to differentiate categories in a context‑aware and semantically sensible manner, with most errors occurring between visually or functionally similar objects.

F1‑Score Curve

The F1‑score curve displays model performance across confidence thresholds, revealing key insight into its behavior and effectiveness. The overall F1‑score peaks at 0.54 when the confidence threshold is 0.202, indicating an optimal balance between precision and recall.

Precision Curve

The precision curve shows that precision improves as the confidence threshold rises for all categories, which is expected. This demonstrates that higher‑confidence predictions are more accurate across the board.

Precision‑Recall Curve

The precision‑recall curve illustrates the trade‑off between precision and recall at varying thresholds, highlighting model performance on aerial image detection tasks. Key findings include:

  1. Overall performance: The model achieves a mean average precision (mAP) of 0.512 across all classes, indicating moderate overall performance.
  2. Class‑wise differences:
    • Tennis court (0.940) and airplane (0.909) display excellent performance, maintaining high precision even at high recall.
    • Basketball court (0.758) and port (0.716) also perform well.
    • Bridge (0.195), football field (0.301) and small vehicle (0.364) perform poorly, with precision dropping sharply as recall increases.

Recall Curve

The recall curve shows how the model’s ability to detect objects (recall) changes as we adjust the confidence threshold. Key observations include:

  1. Overall performance: The model attains a mean recall of 0.53 across all classes, indicating moderate detection capability.
  2. Class‑wise differences:
    • Tennis court, airplane and port achieve strong recall even at high confidence thresholds.
    • Football field, bridge and container crane show poor recall, with recall decreasing quickly as confidence rises.

Overall Result Summary

The overall result summary provides a comprehensive view of model performance, highlighting key metrics and their implications.

  1. Loss reduction: All three loss components (box_loss, cls_loss, dfl_loss) consistently decrease over training epochs, indicating effective learning.
  2. Convergence: Losses appear to stabilize toward the end of training, suggesting the model converges after 100 epochs.
  3. Validation performance: Validation loss closely follows training loss, indicating good generalization without significant over‑fitting.
  4. Precision and recall: Both metrics steadily improve over time, with recall showing a slightly more pronounced early‑stage improvement.
  5. mAP performance: Mean average precision improves consistently under both mAP50 (IoU = 0.5) and mAP50‑95 (IoU = 0.5‑0.95), reaching approximately 0.52 for mAP50 and 0.35 for mAP50‑95.
  6. Learning dynamics: Most metrics improve rapidly at first and then plateau, suggesting the model quickly learns major features before fine‑tuning.
  7. Stability: The metrics/precision and metrics/recall plots exhibit minor fluctuations, which are normal, while the overall trend is upward.
  8. Improvement space: Although performance is solid, the final mAP values indicate room for improvement, potentially via longer training or architectural refinements.

These results demonstrate a successful training process with good generalization and consistent improvements across multiple metrics, providing a robust foundation for aerial image object detection.

AI studio

Generate PPTs instantly with Nano Banana Pro.

Generate PPT Now

Access Dataset

Login to Access

Please login to view download links and access full dataset details.

Topics

Object Detection
Aerial Image Analysis

Source

Organization: github

Created: 7/14/2024

Power Your Data Analysis with Premium AI Models

Supporting GPT-5, Claude-4, DeepSeek v3, Gemini and more.

Enjoy a free trial and save 20%+ compared to official pricing.