SkySeaLand Satellite Object Detection Dataset

Published: 26 November 2025| Version 2 | DOI: 10.17632/d42n3cp86p.2
Contributor:

Description

This dataset was created to support a research idea that transportation-related objects in satellite imagery can be detected with steady accuracy when they are labeled with clear and consistent annotations. It focuses on four object classes: airplane, boat, car, and ship, and includes scenes from airports, highways, harbors, marinas, and coastal regions. These locations were selected to cover different backgrounds, object scales, and environmental conditions, which helps in studying model performance in realistic satellite settings. Images were collected using Google Earth Pro under fair use and academic research guidelines. Candidate regions were explored manually and exported in high resolution through the built-in tools of the software. After collection, each image was reviewed for clarity, relevance, and presence of target objects. Files with heavy noise, strong cloud cover, or duplicate viewpoints were removed. When needed, basic preparation such as cropping and resizing was applied to keep the focus on relevant areas while maintaining visual quality. Annotation was carried out using CVAT and Roboflow. Each object instance was marked with a bounding box and assigned one of the four class labels. A separate verification pass was performed to maintain consistent box placement and correct class assignment across the dataset. For this release, all final annotations are stored in a single COCO format JSON file that follows the standard object detection structure. The images are organized into separate folders for training, validation, and testing. A Roboflow project link is provided so that users can view the dataset online, apply their own preprocessing pipeline, and export the same annotations into formats such as YOLO or Pascal VOC if required. The dataset can be interpreted by loading the COCO JSON file, reading the category identifiers, and mapping them to airplane, boat, car, and ship. Each annotation entry provides bounding box coordinates and class information that can be used directly with common computer vision libraries. Researchers can use this dataset to test model generalization across land and sea scenes, evaluate multi-class detection performance, study small object behavior, compare detection architectures, and explore transfer learning strategies in aerial imagery.

Files

Steps to reproduce

1. Open Google Earth Pro and manually explore airports, highways, coastal zones, harbors, and marinas in different regions. 2. Use the built in export tool to save high resolution images and store them in a working directory. 3. Remove low-quality images that contain blur, heavy cloud cover, or duplicates. 4. Apply basic preprocessing through simple editing tools or Python scripts, including cropping or resizing when necessary. 5. Upload the cleaned images to CVAT or Roboflow. 6. Annotate each airplane, boat, car, and ship with a bounding box and assign the correct class. 7. Perform a second review to correct box placement and ensure consistent labeling. 8. Export all annotations in a single COCO format JSON file that follows standard object detection structure. 9. Split the dataset into separate folders for training, validation, and testing using an 80, 10, and 10 structure. 10. Optionally upload the dataset to Roboflow to enable automatic preprocessing, augmentation, and export to other formats such as YOLO or Pascal VOC.

Institutions

Green University of Bangladesh

Categories

Vehicle, Ships, Boat, Airplane

Licence