Skip to main content
TjMakeBot Blogtjmakebot.com

Drone Aerial Image Labeling: A Complete Practical Guide from Collection to Training

TjMakeBot Technical TeamTechnical Tutorial25 min
Technical TutorialAerial Image Labeling
Drone Aerial Image Labeling: A Complete Practical Guide from Collection to Training

Introduction: When AI Gets a "God's-Eye View"

The proliferation of UAV (Unmanned Aerial Vehicle) technology has finally freed computer vision from the constraints of the ground. Looking down from hundreds of meters above, the world presents an entirely different geometric logic. In fields like agricultural crop protection, urban illegal construction inspection, and solar panel defect detection, aerial AI is solving pain points that traditional manual methods simply cannot reach.

But any data engineer who has worked on aerial projects will tell you: aerial data is a "rose with thorns."

A single 4K-resolution aerial image might contain hundreds of vehicles, pedestrians reduced to a few thousand pixels blending into complex background noise, targets half-hidden under tree shadows, and dramatic scale variations from different flight altitudes that make models struggle to adapt.

This article skips the empty concepts. Drawing on our team's three years of hands-on experience, we'll break down every detail of the full pipeline — from the moment the drone takes off to the final model deployment. This isn't just a labeling guide; it's a pitfall avoidance handbook.

Rethinking Your Data: The Unique Nature of Aerial Images

1. The Double-Edged Sword of the God's-Eye View: Perspective and Scale

When we switch from ground level to the sky, the logic of features is completely restructured.

  • The "dimensional reduction" of shapes: From a ground-level perspective, a car has rich side textures, contours, and wheel features; but from an aerial perspective, it often degrades into a rectangular color block. Pedestrians are an even more extreme example — transforming from an upright being into a moving dot (the top of the head). This requires us to clearly define "top-view features" when establishing labeling rules. For instance, should side mirrors be included? Does a pedestrian's backpack count as part of the body? These details determine the model's generalization ability.

  • Inverted occlusion logic: In ground-level photography, occlusion is typically front-to-back; in aerial imagery, occlusion is vertical. Dense tree canopies may hide cars parked underneath, and overpasses may cut off roads below. When labeling a "car 50% occluded by a tree," should you label the visible portion or mentally reconstruct the full outline? Our experience: If the goal is counting, label the full outline (amodal); if the goal is visual localization, label the visible area (modal).

  • Dramatic scale jumps: This is the biggest headache in aerial imagery. The same object photographed at 50 meters and 200 meters altitude can differ by 16x in pixel area.

    Practical tip: If your dataset mixes data collected at different altitudes, be sure to analyze the Object Scale Distribution before training. If small targets (<32x32 pixels) are overrepresented, standard YOLO or SSD models without targeted modifications (such as adding high-resolution feature layers) will have very poor recall.

2. Easily Overlooked Image Quality Traps

  • Motion Blur: Drones aren't tripods — air turbulence and flight speed both cause blurry images. For cameras with insufficient shutter speed, ground textures may streak. Labeling advice: For samples severely blurred to the point where the human eye can barely identify the category, decisively remove them (Hard Negative). Don't force-label them, as this introduces noise into the model.
  • The deception of lighting and shadows: Long shadows at dawn and dusk are the biggest source of interference. Many beginner models mistake long shadows for the objects themselves, or miss detections due to shadow coverage. Collection advice: Try to collect during the "golden hours" — outside the 2 hours around noon and before sunset — when lighting is neither too harsh nor too angled.

3. Special Considerations for Data Organization

Aerial data typically consists of large images (e.g., 8000x6000 resolution) that would blow up GPU memory if fed directly into a model.

  • Tiling is mandatory: You can't just naively slice — there must be overlap. We generally recommend 15%-20% overlap to prevent targets at tile edges from being split in half and missed.
  • Geo-Tagging: Each image's EXIF data contains GPS coordinates. When labeling, it's best to preserve this information, because in the final application, the client doesn't care about "there's a fire in the image" — they care about "there's a fire at latitude XX, longitude XX."

Plan Your Collection Like a Director: The Art of Zero Rework

Many projects fail not because the algorithm is bad, but because the data source was flawed from the start.

1. The Math Behind Flight Path Planning

Don't fly randomly. You need to calculate flight altitude based on your target size. Formula: Flight altitude ≈ (Target actual size × Focal length) / (Minimum detectable pixels × Sensor pixel size)

Example: You want to detect safety helmets on the ground (approximately 0.3m diameter), and the algorithm requires minimum targets of at least 15x15 pixels. If you're using a lens with an equivalent 24mm focal length and a pixel size of about 3 microns, your maximum flight altitude is approximately 160 meters. Fly any higher, and the helmet becomes noise.

2. The "Golden Window" for Collection Conditions

  • Time: 10:00-11:30 AM, 1:30-3:00 PM. Avoid noon's overhead light (lacks depth perception) and sunrise/sunset's long shadows.
  • Weather: Overcast conditions are actually better than clear skies, because light diffused through clouds eliminates harsh ground shadows and reveals the most detail.
  • Flight parameters: Recommended lateral overlap of 70% and forward overlap of 80%. While this increases data volume, it's crucial for subsequent stitching or selecting the best-angle images.

3. Iron Rules of Data Management

  • File naming: Reject DJI_0001.jpg. Recommended format: {Location}_{Date}_{Altitude}_{FlightLineID}_{Sequence}.jpg. For example, FarmA_20260206_H50m_L1_0023.jpg. At a glance, you can tell where, when, and at what altitude the image was captured.
  • On-site verification: After landing, always spot-check a few original images on a computer. Any focus failures? Overexposure? The cost of re-flying on-site is a few hundred dollars; the cost of discovering unusable data back at the office and returning is thousands.

Labeling Strategies: From Rough to Refined

Strategy 1: The "Unwritten Rules" of Object Detection (Bounding Box)

Rule 1: Box Tightness This is the most common mistake beginners make. Boxes drawn too loosely include too much background (like road surface), causing the model to learn "gray road surface" as a feature of cars.

Standard: Box edges should tightly hug the target boundary, with pixel error controlled within 2-3px. For targets with shadows, do not include the shadow! Shadows change with time; the object itself doesn't.

Rule 2: The "Hell Mode" of Dense Targets In parking lots or crowded gatherings, targets are packed tightly together.

Tip: Carefully check box overlap (IoU). If two target boxes have IoU exceeding 0.7, consider whether to merge categories (e.g., "row of vehicles") or use Oriented Bounding Boxes (OBB) for labeling. In aerial imagery, rotated boxes often perform much better than horizontal boxes because they perfectly fit diagonally parked vehicles, reducing background interference.

Rule 3: Handling Truncated Targets Should you label an object at the image edge that's only half visible?

Recommendation: If more than 50% is visible, label it and tag it as truncated; if less than 30% is visible, don't label it, and set that area as ignore (if the tool supports it) to prevent the model from learning it as a negative sample.

Strategy 2: The Efficiency Battle of Semantic Segmentation

Pixel-level labeling is extremely time-consuming — a complex aerial image can take 2 hours of pure manual labeling. Efficiency boosters:

  1. Superpixel pre-segmentation: Using color and texture similarity, first divide the image into small blocks. Annotators only need to click-select these blocks and assign categories, improving efficiency 5-10x.
  2. Polygon vs. Brush: For buildings and roads with straight edges, use the polygon tool; for vegetation and water bodies with irregular shapes, use the brush tool.
  3. Hierarchical labeling: First roughly label major categories (e.g., "vegetation"), then subdivide into subcategories (e.g., "trees," "grass").

Strategy 3: The Registration Challenge of Change Detection

Finding differences between two images requires them to be properly aligned first. Practical pain point: GPS from two drone flights may have several meters of error, preventing pixel-level alignment. Solutions:

  • Register first, label second: Use feature point matching algorithms like SIFT/SURF, or dedicated registration software, to forcibly correct the T1 image to the T2 coordinate system.
  • Labeling is more than drawing boxes: Change detection typically requires labeling "Change Pairs" — identifying what in Image A changed to what in Image B, along with the type of change (e.g., "newly built," "demolished").

Real-World Cases: Lessons from the Trenches

Case 1: Smart Agriculture — "Spot the Difference" in Wheat Fields

Background: Identifying stripe rust infection centers across 500 acres of wheat fields. Challenge: Early-stage disease only causes leaf yellowing, which is hard to distinguish from uneven lighting. Breakthrough:

  1. Multispectral sensors: Standard RGB cameras couldn't see clearly enough, so we introduced the NDVI (Normalized Difference Vegetation Index) channel. In false-color images, diseased areas showed distinctly abnormal red features.
  2. Graded labeling: Instead of just labeling "diseased," we labeled "mild," "moderate," and "severe." While this increased labeling difficulty, it taught the model to recognize disease progression features. Result: Early disease detection rate improved from 60% to 92%.

Case 2: Urban Illegal Parking — Misjudgments from Above

Background: Identifying fire lane obstruction. Challenge: From high above, how do you tell if a vehicle is "parked" or "moving"? Breakthrough:

  1. Introducing the time dimension: A single image can't determine state. We switched to capturing short videos or taking 3 consecutive shots at 5-second intervals.
  2. Logical labeling: Only vehicles that remained nearly stationary across 3 consecutive frames were labeled as "stationary."
  3. Scene association: We specifically labeled the "fire lane" as a Region of Interest (ROI). An alert was only triggered when a "stationary vehicle's" center point fell within the "fire lane" area.

TjMakeBot's Aerial-Specific Features

We developed a dedicated toolchain in TjMakeBot to address the pain points described above:

  1. Ultra-large image tiling engine: Upload hundreds of megabytes of TIF orthoimages, and the system automatically creates pyramid tiles on the frontend. You can zoom and browse like using Google Maps, with labeling results automatically mapped back to original full-image coordinates.

  2. Native Oriented Bounding Box (OBB) support: Hold a shortcut key and directly drag out an angled rectangular box. Export formats perfectly support mainstream aerial dataset formats including DOTA and YOLOv8-OBB.

  3. Geographic projection sync: Import GeoTIFF files with coordinates, draw a box on the image, and the system displays the corresponding latitude/longitude range and actual physical area (square meters) in real time. This is extremely useful for estimating "affected area" or "building footprint."

  4. AI-assisted labeling (SAM integration): Integrated with a Segment Anything Model fine-tuned for remote sensing. For targets like solar panels and building rooftops, a single click automatically generates a perfect contour — no manual point-by-point tracing needed.

Optimization Tips: Squeezing Out the Last 1%

Data Augmentation Should Be "Moderate"

  • Recommended: Random rotation (0-360 degrees, since aerial images have no absolute up or down), random cropping (simulating different fields of view), Mosaic augmentation (improving small object detection).
  • Use with caution: Excessive Color Jitter. Aerial image colors often carry important information (e.g., water color indicates pollution level, vegetation color indicates health). Distorting colors too aggressively destroys features.

Solving Class Imbalance

In aerial images, background (Negative) often accounts for 99%, with targets at only 1%.

  • Copy-Paste technique: Extract rare targets (such as a specific rare vehicle type) from original images and randomly paste them onto other background images. This works much better than simply duplicating images because it changes the target's background context.

Conclusion

Drone aerial image labeling is essentially teaching machines to understand the human world from a god's-eye view. This requires not only precise manual skills but also a deep understanding of physical world imaging principles.

Good data is cultivated, and more importantly, "designed." We hope this guide helps you avoid detours and lets your models fly higher and see more accurately.


Try TjMakeBot's Aerial-Specific Labeling Tools

Keywords: drone labeling, remote sensing AI, rotated object detection, OBB labeling, aerial small targets, TjMakeBot in practice