Research Article

Split Viewer

Korean J. Remote Sens. 2024; 40(6): 1141-1148

Published online: December 31, 2024

https://doi.org/10.7780/kjrs.2024.40.6.1.21

© Korean Society of Remote Sensing

Ship Detection in Dual-Polarization SAR Imagery Based on YOLO and Late Fusion

Yunjee Kim1* , Jinsoo Kim2 , Ki-mook Kang3

1Researcher, Maritime Digital Transformation Research Center, Korea Research Institute of Ships & Ocean Engineering, Daejeon, Republic of Korea
2CTO, DEEP.I Inc., Asan, Republic of Korea
3Senior Researcher, Water Resource Satellite Center, K-water Research Institute, Daejeon, Republic of Korea

Correspondence to : Yunjee Kim
E-mail: yunjee0531@kriso.re.kr

Received: November 15, 2024; Revised: December 5, 2024; Accepted: December 15, 2024

This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (https://creativecommons.org/licenses/by-nc/4.0/) which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

With the recent development of satellites and microsatellite constellations, higher temporal resolution satellite imagery and larger numbers of satellites have become available. As the availability of satellite data increases, it is necessary to develop ship detection models that consider specific data characteristics, such as polarization, spatial resolution, and frequency bands. Among these, to leverage the distinct features of polarization, we employed a late fusion approach to merge the ship detection results from Sentinel-1 dual-polarization data. To evaluate the effectiveness of the fusion model, we built four single training models using two polarizations (VV, VH) and two colormaps (gray, parula), as well as six multimodal models with late fusion. As a result of comparing the accuracy of the single model and the fusion model, we found that the accuracy of the fusion model consisting of 1) VH gray colormap and VV parula colormap, and 2) VH parula colormap and VV parula colormap is higher than that of the single model (based on intersection over union (IoU) thresholds of 0.4 and 0.5). Each fusion model achieved a relative accuracy improvement of at least 1.5% and up to 6.5% compared to the single model with the highest accuracy among the two. The significance of this study is that the late fusion was applied using both polarization data and colormap information simultaneously. These results suggest that the fusion model can detect ships more accurately and that colormaps, which have been underexplored in SAR research, can be a factor in improving accuracy.

Keywords Ship detection, SAR, Sentinel-1, Late fusion, YOLO, Dual-polarization

Ship monitoring in coastal and maritime areas is important for ensuring marine safety, maritime security, prompt shipwreck and accident responses, and preventing marine pollution (Nie et al., 2020; Qin et al., 2021; Wang et al., 2019; Zhang et al., 2021a). Coastal ship monitoring is primarily conducted through vessel traffic services (VTSs) and patrol vessels; however, this type of monitoring is limited due to the broad expanse of the ocean. As of late 2019, there were more than 90,000 merchant ships and 5,000 warships worldwide, complicating the rapid detection and accurate identification of the wide variety of ships in the ocean (Zhang et al., 2021c). Not all ships are obligated to carry transponders, and these may be switched off for various reasons. In areas with high ship traffic, it is difficult to monitor numerous ships with limited equipment (Tian et al., 2022); therefore, satellite data are essential for ship monitoring over wide areas. Many studies have investigated ship detection using satellite images, particularly using SAR satellites, which can acquire data regardless of the weather or time of day (Li et al., 2022; Miao et al., 2022; Pelich et al., 2019; Song et al., 2020).

Satellites are being gradually miniaturized, and synthetic aperture radar (SAR) microsatellite constellations, such as those produced by ICEYE and Capella, are becoming more common. This process has significantly improved the spatial and temporal resolution of satellite data producing an abundance of highquality satellite data. The increasing availability of various types of data from multiple satellites is creating new opportunities to utilize multi-satellite data, and any related research is considered essential. Effectively leveraging multi-satellite data requires the development of sophisticated algorithms capable of handling diverse data types and extracting meaningful insights. Therefore, in this study, we aimed to suggest a detection algorithm that uses dual-polarization data with different characteristics as input.

Ship detection research using SAR images is now mainly conducted through deep learning methods, including the You Only Look Once (YOLO) algorithm, which has high speed and detection accuracy (Chang et al., 2019; Guo et al., 2022; Sun et al., 2021; Tang et al., 2021; Wang et al., 2021; Wang et al., 2022). The YOLO v2 algorithm outperformed the generally faster region-based convolutional neural network (CNN) algorithm, showing 5.8-fold faster ship detection; an extension of this algorithm was 2.5-fold faster than the original (Chang et al., 2019). A subsequent novel ship detection approach integrated noise level classification and potential target area extraction based on the YOLO algorithm, and showed superior detection of objects directly from images; however, its data loss during edge extraction requires further refinement (Tang et al., 2021).

An analysis by Im et al. (2023) has shown the potential of utilizing various polarization data in ship detection by comparing the accuracy of YOLO-based detection models built with each polarization data and YOLO-based detection models built by integrating all polarization data. This study demonstrated that VV polarization typically shows higher backscatter coefficients than VH polarization (Im et al., 2023). Noting that dual-polarization data each has its unique characteristics, we compared the accuracy of a single model built with VV and VH images from the Sentinel-1 with a fusion model using a late fusion technique. While constructing the fusion model, we used gray and parula colormaps to demonstrate that the colormap of an image can also influence accuracy.

2.1. Outline of Datasets

In this study, training and test datasets were generated from a set of 31 Sentinel-1A/B images, captured between July 3, 2019, and March 18, 2021, over Busan in South Korea. The original image was cropped into sub-images according to the condition and prerequisite in Table 1.

Table 1 Description of sub-image data in this studyDescription of sub-image data in this study

Sub-imageConditionPrerequisiteNo. of sub-images
400s × 400sRandomly crop the sub-image from the original image so that both sides have lengths between 400 and 499Contains at least three ships322
500s × 500sRandomly crop the sub-image from the original image so that both sides have lengths between 500 and 599
600s × 600sRandomly crop the sub-image from the original image so that both sides have lengths between 600 and 699
700s × 700sRandomly crop the sub-image from the original image so that both sides have lengths between 700 and 799


The dataset was partitioned into training (60%), validation (20%), and testing (20%) subsets to balance training data availability with adequate samples for evaluation and hyperparameter tuning. Especially, randomization was employed during allocation to minimize selection bias, ensuring a representative sample across all subsets. This rigorous partitioning approach enabled a robust evaluation of model performance across diverse conditions, laying a strong foundation for reliable ship detection in SAR images.

2.2. Colormaps

Previous studies have shown that image segmentation performance varies depending on the color of the image (Cheng et al., 2001; Lucchese and Mitra, 2001). We took note of these points and considered that the colormap of the image might be related to the object detection accuracy when detecting ships. According to previous research, the Autumn, Viridis, and Parula colormaps demonstrated superior segmentation performance in microwave tomography (Zhang et al., 2021b). Among these, the parula colormap was found to have mean and standard deviation (SD) values similar to those of the gray colormap, and thus, it was compared with the gray colormap in this study. The following graph represents the RGB values of two colormaps (Fig. 1).

Fig. 1. RGB values: (a) gray and (b) parula colormaps.

If an image is represented using the gray colormap, its colors will range between black and white because its RGB values are distributed between 0 and 1. By contrast, parula colormap exhibits larger blue values for smaller values, and an increase in red and green values for larger values (Fig. 1). Although it is challenging to intuitively distinguish SAR images based on different colormaps, we used two distinct colormaps for all training data because they may exhibit different characteristics in the deep learning process (Fig. 2). Both colormaps used in this study were scaled to have values in the range of 0 to 255.

Fig. 2. SAR images of ships captured using different polarization modes, expressed as (a) VV gray, (b) VV parula, (c) VH gray, and (d) VH parula colormaps.

3.1. YOLO

YOLO is a single-stage object detection algorithm based on deep learning; it is based on the human glance, which instantly grasps the types and correlations of objects included in image data (Liu et al., 2016; Redmon, 2016). Also, it uses a pipeline that that infers object information from the entire image, and therefore learns information about the background region, in addition to the specific region where the object is located; this results in smaller background errors compared to two-stage object detection algorithms (Qing et al., 2021; Zhou et al., 2022). Zhu et al. (2021) compared the detection speed and accuracy of three region-based convolutional neural network (R-CNN)-based two-stage detectors and four single-stage detectors, confirming that the single-stage detector, YOLOv5s, achieved the fastest detection speed at 63.3 frames per second (FPS).

The YOLO algorithm has evolved through various versions. In a study by Kim (2023), YOLOv1 and YOLOv2 are suitable for tasks where accuracy is critical, while YOLOv3 and YOLOv4 are primarily used for tasks that prioritize speed. On the other hand, YOLOv5 and YOLOv6 are effective in environments requiring a balance between accuracy and speed, whereas YOLOv7 and YOLOv8 are known to be applicable in scenarios with limited memory and computational resources. In this study, YOLOv5, which shows an excellent balance between accuracy and speed, was used considering the real-time nature of ship detection.

3.2. Late Fusion

Late fusion involves applying object detection models to data from various sensors and combining the detection results at the decision-making stage. This method maintains high detection accuracy and robustness, even when the performance of individual sensors deteriorates (Kim and Cho, 2021; Kim et al., 2019; Zhao et al., 2021). In this study, late fusion was used to consider the features of VV and VH images with different features, and the detection results of each single model were fused using a soft non-maximum suppression (NMS)-based late fusion method (Fig. 3).

Fig. 3. Flowchart of the late fusion method.

A study by Farahnakian and Heikkonen (2020) explored ship detection using late fusion. They compared the accuracy of different fusion methods by integrating ship detection results from RGB and IR images, employing not only late fusion but also early and middle fusion techniques. Although various studies have explored fusing sensor detection results, this research is the first to improve ship detection accuracy by employing late fusion of dual-polarization data from SAR images.

Before comparing the accuracy of single models and fused models, we first evaluated the accuracy of single models based on polarization and color maps to determine whether the polarization and color map of SAR images affect the accuracy. This study evaluates mean average precision (mAP) at IoU thresholds between 0.3 and 0.5, reflecting varying levels of localization precision, to comprehensively assess model performance.

4.1. Comparison of Single Model Accuracy based on Polarization and Colormap

As a result of comparing the accuracy according to polarization, the VH image had higher accuracy in more than 79% of all cases. In particular, it had significantly higher detection accuracy than VV, for the gray colormap; the difference was even greater at a threshold of 0.5. The accuracy of the VH image was 26% higher than that of the VV image. However, for the parula colormap, the detection accuracy was generally similar between polarization modes (Table 2). Next, we compared the accuracy of the gray and parula colormaps. The detection accuracy of the parula colormap was higher than that of the gray colormap in more than 80% of cases. The difference in accuracy between the gray and parula colormaps was greater for VV images than for VH images and became larger as the IoU threshold increased (Table 2).

Table 2 Comparison of single model accuracy based on polarization and colormaps

Input sizePolarizationColormapIoU threshold
0.30.40.5
400s × 400sVVGray73.6547.8621.65
VH89.7072.5746.54
VVParula91.3973.8744.37
VH93.2677.4646.73
500s × 500sVVGray89.5067.2636.49
VH94.3984.3061.38
VVParula94.3384.7864.76
VH96.1287.3468.55
600s × 600sVVGray92.2774.8050.88
VH97.1986.3067.04
VVParula95.6488.6672.36
VH96.8385.4871.42
700s × 700sVVGray94.2779.5948.57
VH96.2487.6074.61
VVParula97.3189.5275.19
VH96.1488.4574.09


4.2. Comparison of the Accuracy of Fusion Models based on Single Model Combinations

This study proposed a ship detection method that fuses the detection results of dual-polarization data of different colormap types to improve the accuracy of deep training models by utilizing both training data with different characteristics. To improve detection accuracy, we compared the accuracy of all fusion models that could be combined into four single models (VV gray, VV parula, VH gray, VH parula). Specifically, the detection accuracy of a fusion model, which consists of a training model with an image size of and a threshold of 0.4 for late fusion, was compared (Table 3).

Table 3 Ship detection accuracy comparison according to multimodal types and IoU thresholds (Input size: 700s×700s / Multimodal threshold: 0.4)

Multimodal thresholdMultimodal inputIoU threshold
0.30.40.5
0.4VV grayVH gray95.8886.7369.79
VH grayVV parula96.7491.5978.81
VH grayVH parula97.8388.9175.97
VV grayVH parula95.7988.7068.97
VV grayVV parula94.8286.8064.88
VH parulaVV parula97.9290.8678.87


As a result, it was found that the fused model with 1) VH gray colormap and VV parula colormap, and 2) VH parula colormap and VV parula colormap as multimodal inputs has higher accuracy than the single model (for IoU threshold of 0.4 and 0.5). For an IoU threshold of 0.4, the accuracy of the VH gray colormap single model was 87.60%, and that for the parula VV single model had an accuracy of 89.54%, both of which were below 90%; by contrast, a fusion of these single models resulted in an accuracy level higher than 90%. Similarly, the accuracy levels of the single VH parula and VV parula models were 88.45% and 89.52%, respectively, whereas the fusion of these models resulted in > 90% accuracy. Both multimodal models had higher true positive rates than the single models, despite a nonsignificant difference in the number of detections (‘Det’ in Table 4) as shown in Table 4.

Table 4 Comparison of results produced by two multimodal models with high detection accuracy and their corresponding single model predictions

Predict resultInput
VH grayVH parulaVV parulaVH grayVH parula
VV parulaVV parula
Ground Truth (GT)731731731731731
Detected (Det)801776860840823
True Positive (TP)657662671678684
False Positive (FP)144114189162139


Improving ship detection accuracy consists of two components: increasing the rate of true positives and eliminating false positives. Factors such as ship wake and offshore structures can cause ships to be mistaken for noise whereas the multimodal model is more likely to detect ships among these confounding factors (top of Fig. 4a, blue circle in Fig. 4d). Conversely a bright area on a SAR image may be falsely detected as a ship due to noise from small islands or strong backscattering signals whereas the multimodal model will correctly recognize that these signals are not ships (left in Figs. 4b, c; yellow circle in Figs. 4e, f).

Fig. 4. Part of ship detection results for (a–c) single and (d–f) multimodal models.

In this study, we compared the accuracy of four single models using dual-polarization and gray and parula color maps, as well as six fused models utilizing late fusion techniques to integrate data with diverse characteristics. The comparison of 24 cases of single models revealed that, in 79% of the cases, the VH image was more accurate than the VV image, with the difference being more pronounced in the gray colormap. This indicates that the VH gray demonstrates high accuracy. Conversely, when comparing results based on the colormap, the parula colormap showed higher accuracy in 83% of the cases, with the difference being more significant in the VV image. This suggests that the VV parula provides high accuracy. To summarize, it can be concluded that VH parula demonstrates the highest accuracy. This finding aligns with the accuracy comparisons observed in the fused models.

Results from comparing the accuracy of six fusion models show that the best model for the YOLOv5-based late fusion algorithm proposed in this study involved fusing two pairs of detection results: 1) VH gray colormap and VV parula colormap, and 2) VH parula colormap and VV gray colormap, using a multimodal threshold of 0.4 and an IoU threshold of 0.4 or 0.5. All single models showed <90% accuracy for an IoU threshold of 0.4, whereas the optimal fusion models showed >90% accuracy, indicating effective compensation for both false positives and non detections. The YOLOv5-based late fusion method reduced the false detection rate by eliminating multiple detection results, which was a limitation of the NMS method, through the adoption of soft-NMS. The results of this study can be used for fundamental data to develop a multisensor-based ship detection algorithm. Future research will explore ways to utilize various satellite data to improve ship detection accuracy.

This research was funded Korea Institute of Marine Science & Technology Promotion (KIMST) and by the Ministry of Oceans and Fisheries (RS-2024-00415504).

No potential conflict of interest relevant to this article was reported.

  1. Chang, Y. L., Anagaw, A., Chang, L., Wang, Y. C., Hsiao, C. Y., and Lee, W. H., 2019. Ship detection based on YOLOv2 for SAR imagery. Remote Sensing, 11(7), 786. https://doi.org/10.3390/rs11070786
  2. Cheng, H. D., Jiang, X. H., Sun, Y., and Wang, J., 2001. Color image segmentation: Advances and prospects. Pattern Recognition, 34(12), 2259-2281. https://doi.org/10.1016/S0031-3203(00)00149-7
  3. Farahnakian, F., and Heikkonen, J., 2020. Deep learning based multi-modal fusion architectures for maritime vessel detection. Remote Sensing, 12(16), 2509. https://doi.org/10.3390/rs12162509
  4. Guo, Y., Chen, S., Zhan, R., Wang, W., and Zhang, J., 2022. LMSD-YOLO: A lightweight YOLO algorithm for multi-scale SAR ship detection. Remote Sensing, 14(19), 4801. https://doi.org/10.3390/rs14194801
  5. Im, Y., Youn, Y., Kang, J., Kim, S., Jeong, Y., and Choi, S., et al, 2023. Ship detection from SAR images using YOLO: Model constructions and accuracy characteristics according to polarization. Korean Journal of Remote Sensing, 39(5-3), 997-1008. https://doi.org/10.7780/kjrs.2023.39.5.3.9
  6. Kim, J., and Cho, J., 2021. A set of single YOLO modalities to detect occluded entities via viewpoint conversion. Applied Sciences, 11(13), 6016. https://doi.org/10.3390/app11136016
  7. Kim, J., Kim, J., and Cho, J., 2019 December. An advanced object classification strategy using YOLO through camera and LiDAR sensor fusion. In In Proceedings of the 2019 13th International Conference on Signal Processing and Communication Systems (ICSPCS), Gold Coast, Australia, Dec. 16−18, pp. 1-5. https://doi.org/10.1109/ICSPCS47537.2019.9008742
  8. Kim, J. Y., 2023. A comparative study on the characteristics of each version of object detection model YOLO. In Proceedings of the 2023 Korean Society of Computer Information Conference, Jeju, Korea, July 12-15, pp. 75-78.
  9. Li, J., Xu, C., Su, H., Gao, L., and Wang, T., 2022. Deep learning for SAR ship detection: Past, present and future. Remote Sensing, 14(11), 2712. https://doi.org/10.3390/rs14112712
  10. Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C. Y., and Berg, A. C., 2016. SSD: Single Shot MultiBox Detector. In: Leibe, B., Matas, J., Sebe, N., Welling, M., (eds.), Computer Vision -ECCV 2016, Springer, pp. 21-37. https://doi.org/10.1007/978-3-319-46448-0_2
  11. Lucchese, L., and Mitra, S. K., 2001. Color image segmentation: A state-of-the-art survey. Proceedings-Indian National Science Academy Part A, 67(2), 207-222.
  12. Miao, T., Zeng, H., Yang, W., Chu, B., Zou, F., Ren, W., and Chen, J., 2022. An improved lightweight RetinaNet for ship detection in SAR images. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 15, 4667-4679. https://doi.org/10.1109/JSTARS.2022.3180159
  13. Nie, X., Duan, M., Ding, H., Hu, B., and Wong, E. K., 2020. Attention mask R-CNN for ship detection and segmentation from remote sensing images. IEEE Access, 8, 9325-9334. https://doi.org/10.1109/ACCESS.2020.2964540
  14. Pelich, R., Chini, M., Hostache, R., Matgen, P., Lopez-Martinez, C., and Nuevo, M., et al, 2019. Large-scale automatic vessel monitoring based on dual-polarization Sentinel-1 and AIS data. Remote Sensing, 11(9), 1078. https://doi.org/10.3390/rs11091078
  15. Qin, P., Cai, Y., Liu, J., Fan, P., and Sun, M., 2021. Multilayer feature extraction network for military ship detection from high-resolution optical remote sensing images. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 14, 11058-11069. https://doi.org/10.1109/JSTARS.2021.3123080
  16. Qing, Y., Liu, W., Feng, L., and Gao, W., 2021. Improved Yolo network for free-angle remote sensing target detection. Remote Sensing, 13(11), 2171. https://doi.org/10.3390/rs13112171
  17. Redmon, J., 2016. You only look once: Unified, real-time object detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, June 27-30, pp. 257-276. https://doi.org/10.1109/JPROC.2023.3238524
  18. Song, J., Kim, D. J., and Kang, K. M., 2020. Automated procurement of training data for machine learning algorithm on ship detection using AIS information. Remote Sensing, 12(9), 1443. https://doi.org/10.3390/rs12091443
  19. Sun, Z., Leng, X., Lei, Y., Xiong, B., Ji, K., and Kuang, G., 2021. BiFA-YOLO: A novel YOLO-based method for arbitrary-oriented ship detection in high-resolution SAR images. Remote Sensing, 13(21), 4209. https://doi.org/10.3390/rs13214209
  20. Tang, G., Zhuge, Y., Claramunt, C., and Men, S., 2021. N-Yolo: A SAR ship detection using noise-classifying and complete-target extraction. Remote Sensing, 13(5), 871. https://doi.org/10.3390/rs13050871
  21. Tian, Y., Liu, J., Zhu, S., Xu, F., Bai, G., and Liu, C., 2022. Ship detection in visible remote sensing image based on saliency extraction and modified channel features. Remote Sensing, 14(14), 3347. https://doi.org/10.3390/rs14143347
  22. Wang, J., Lin, Y., Guo, J., and Zhuang, L., 2021. SSS-YOLO: Towards more accurate detection for small ships in SAR image. Remote Sensing Letters, 12(2), 93-102. https://doi.org/10.1080/2150704X.2020.1837988
  23. Wang, S., Gao, S., Zhou, L., Liu, R., Zhang, H., and Liu, J., et al, 2022. YOLO-SD: Small ship detection in SAR images by multiscale convolution and feature transformer module. Remote Sensing, 14(20), 5268. https://doi.org/10.3390/rs14205268
  24. Wang, Y., Wang, C., Zhang, H., Dong, Y., and Wei, S., 2019. A SAR dataset of ship detection for deep learning under complex backgrounds. Remote Sensing, 11(7), 765. https://doi.org/10.3390/rs11070765
  25. Zhang, T., Zhang, X., Liu, C., Shi, J., Wei, S., and Ahmad, I., et al, 2021a. Balance learning for ship detection from synthetic aperture radar remote sensing imagery. ISPRS Journal of Photogrammetry and Remote Sensing, 182, 190-207. https://doi.org/10.1016/j.isprsjprs.2021.10.010
  26. Zhang, Y., Fjeld, M., Fratarcangeli, M., Said, A., and Zhao, S., 2021b. Affective colormap design for accurate visual comprehension in industrial tomography. Sensors, 21(14), 4766. https://doi.org/10.3390/s21144766
  27. Zhang, Z., Zhang, L., Wang, Y., Feng, P., and He, R., 2021c. ShipRSImageNet: A large-scale fine-grained dataset for ship detection in high-resolution optical remote sensing images. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 14, 8458-8472. https://doi.org/ 10.1109/JSTARS.2021.3104230
  28. Zhao, J., Wang, Y., Cao, Y., Guo, M., Huang, X., and Zhang, R., et al, 2021. The fusion strategy of 2D and 3DD information based on deep learning: A review. Remote Sensing, 13(20), 4029. https://doi.org/10.3390/rs13204029
  29. Zhou, X., Jiang, L., Hu, C., Lei, S., Zhang, T., and Mou, X., 2022. YOLO-SASE: An improved YOLO algorithm for the small targets detection in complex backgrounds. Sensors, 22(12), 4600. https://doi.org/10.3390/s22124600
  30. Zhu, H., Xie, Y., Huang, H., Jing, C., Rong, Y., and Wang, C., 2021. DB-YOLO: A duplicate bilateral YOLO network for multiscale ship detection in SAR images. Sensors, 21(23), 8146. https://doi.org/10.3390/s21238146

Research Article

Korean J. Remote Sens. 2024; 40(6): 1141-1148

Published online December 31, 2024 https://doi.org/10.7780/kjrs.2024.40.6.1.21

Copyright © Korean Society of Remote Sensing.

Ship Detection in Dual-Polarization SAR Imagery Based on YOLO and Late Fusion

Yunjee Kim1* , Jinsoo Kim2 , Ki-mook Kang3

1Researcher, Maritime Digital Transformation Research Center, Korea Research Institute of Ships & Ocean Engineering, Daejeon, Republic of Korea
2CTO, DEEP.I Inc., Asan, Republic of Korea
3Senior Researcher, Water Resource Satellite Center, K-water Research Institute, Daejeon, Republic of Korea

Correspondence to:Yunjee Kim
E-mail: yunjee0531@kriso.re.kr

Received: November 15, 2024; Revised: December 5, 2024; Accepted: December 15, 2024

This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (https://creativecommons.org/licenses/by-nc/4.0/) which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

With the recent development of satellites and microsatellite constellations, higher temporal resolution satellite imagery and larger numbers of satellites have become available. As the availability of satellite data increases, it is necessary to develop ship detection models that consider specific data characteristics, such as polarization, spatial resolution, and frequency bands. Among these, to leverage the distinct features of polarization, we employed a late fusion approach to merge the ship detection results from Sentinel-1 dual-polarization data. To evaluate the effectiveness of the fusion model, we built four single training models using two polarizations (VV, VH) and two colormaps (gray, parula), as well as six multimodal models with late fusion. As a result of comparing the accuracy of the single model and the fusion model, we found that the accuracy of the fusion model consisting of 1) VH gray colormap and VV parula colormap, and 2) VH parula colormap and VV parula colormap is higher than that of the single model (based on intersection over union (IoU) thresholds of 0.4 and 0.5). Each fusion model achieved a relative accuracy improvement of at least 1.5% and up to 6.5% compared to the single model with the highest accuracy among the two. The significance of this study is that the late fusion was applied using both polarization data and colormap information simultaneously. These results suggest that the fusion model can detect ships more accurately and that colormaps, which have been underexplored in SAR research, can be a factor in improving accuracy.

Keywords: Ship detection, SAR, Sentinel-1, Late fusion, YOLO, Dual-polarization

1. Introduction

Ship monitoring in coastal and maritime areas is important for ensuring marine safety, maritime security, prompt shipwreck and accident responses, and preventing marine pollution (Nie et al., 2020; Qin et al., 2021; Wang et al., 2019; Zhang et al., 2021a). Coastal ship monitoring is primarily conducted through vessel traffic services (VTSs) and patrol vessels; however, this type of monitoring is limited due to the broad expanse of the ocean. As of late 2019, there were more than 90,000 merchant ships and 5,000 warships worldwide, complicating the rapid detection and accurate identification of the wide variety of ships in the ocean (Zhang et al., 2021c). Not all ships are obligated to carry transponders, and these may be switched off for various reasons. In areas with high ship traffic, it is difficult to monitor numerous ships with limited equipment (Tian et al., 2022); therefore, satellite data are essential for ship monitoring over wide areas. Many studies have investigated ship detection using satellite images, particularly using SAR satellites, which can acquire data regardless of the weather or time of day (Li et al., 2022; Miao et al., 2022; Pelich et al., 2019; Song et al., 2020).

Satellites are being gradually miniaturized, and synthetic aperture radar (SAR) microsatellite constellations, such as those produced by ICEYE and Capella, are becoming more common. This process has significantly improved the spatial and temporal resolution of satellite data producing an abundance of highquality satellite data. The increasing availability of various types of data from multiple satellites is creating new opportunities to utilize multi-satellite data, and any related research is considered essential. Effectively leveraging multi-satellite data requires the development of sophisticated algorithms capable of handling diverse data types and extracting meaningful insights. Therefore, in this study, we aimed to suggest a detection algorithm that uses dual-polarization data with different characteristics as input.

Ship detection research using SAR images is now mainly conducted through deep learning methods, including the You Only Look Once (YOLO) algorithm, which has high speed and detection accuracy (Chang et al., 2019; Guo et al., 2022; Sun et al., 2021; Tang et al., 2021; Wang et al., 2021; Wang et al., 2022). The YOLO v2 algorithm outperformed the generally faster region-based convolutional neural network (CNN) algorithm, showing 5.8-fold faster ship detection; an extension of this algorithm was 2.5-fold faster than the original (Chang et al., 2019). A subsequent novel ship detection approach integrated noise level classification and potential target area extraction based on the YOLO algorithm, and showed superior detection of objects directly from images; however, its data loss during edge extraction requires further refinement (Tang et al., 2021).

An analysis by Im et al. (2023) has shown the potential of utilizing various polarization data in ship detection by comparing the accuracy of YOLO-based detection models built with each polarization data and YOLO-based detection models built by integrating all polarization data. This study demonstrated that VV polarization typically shows higher backscatter coefficients than VH polarization (Im et al., 2023). Noting that dual-polarization data each has its unique characteristics, we compared the accuracy of a single model built with VV and VH images from the Sentinel-1 with a fusion model using a late fusion technique. While constructing the fusion model, we used gray and parula colormaps to demonstrate that the colormap of an image can also influence accuracy.

2. Datasets

2.1. Outline of Datasets

In this study, training and test datasets were generated from a set of 31 Sentinel-1A/B images, captured between July 3, 2019, and March 18, 2021, over Busan in South Korea. The original image was cropped into sub-images according to the condition and prerequisite in Table 1.

Table 1 . Description of sub-image data in this studyDescription of sub-image data in this study.

Sub-imageConditionPrerequisiteNo. of sub-images
400s × 400sRandomly crop the sub-image from the original image so that both sides have lengths between 400 and 499Contains at least three ships322
500s × 500sRandomly crop the sub-image from the original image so that both sides have lengths between 500 and 599
600s × 600sRandomly crop the sub-image from the original image so that both sides have lengths between 600 and 699
700s × 700sRandomly crop the sub-image from the original image so that both sides have lengths between 700 and 799


The dataset was partitioned into training (60%), validation (20%), and testing (20%) subsets to balance training data availability with adequate samples for evaluation and hyperparameter tuning. Especially, randomization was employed during allocation to minimize selection bias, ensuring a representative sample across all subsets. This rigorous partitioning approach enabled a robust evaluation of model performance across diverse conditions, laying a strong foundation for reliable ship detection in SAR images.

2.2. Colormaps

Previous studies have shown that image segmentation performance varies depending on the color of the image (Cheng et al., 2001; Lucchese and Mitra, 2001). We took note of these points and considered that the colormap of the image might be related to the object detection accuracy when detecting ships. According to previous research, the Autumn, Viridis, and Parula colormaps demonstrated superior segmentation performance in microwave tomography (Zhang et al., 2021b). Among these, the parula colormap was found to have mean and standard deviation (SD) values similar to those of the gray colormap, and thus, it was compared with the gray colormap in this study. The following graph represents the RGB values of two colormaps (Fig. 1).

Figure 1. RGB values: (a) gray and (b) parula colormaps.

If an image is represented using the gray colormap, its colors will range between black and white because its RGB values are distributed between 0 and 1. By contrast, parula colormap exhibits larger blue values for smaller values, and an increase in red and green values for larger values (Fig. 1). Although it is challenging to intuitively distinguish SAR images based on different colormaps, we used two distinct colormaps for all training data because they may exhibit different characteristics in the deep learning process (Fig. 2). Both colormaps used in this study were scaled to have values in the range of 0 to 255.

Figure 2. SAR images of ships captured using different polarization modes, expressed as (a) VV gray, (b) VV parula, (c) VH gray, and (d) VH parula colormaps.

3. Methods

3.1. YOLO

YOLO is a single-stage object detection algorithm based on deep learning; it is based on the human glance, which instantly grasps the types and correlations of objects included in image data (Liu et al., 2016; Redmon, 2016). Also, it uses a pipeline that that infers object information from the entire image, and therefore learns information about the background region, in addition to the specific region where the object is located; this results in smaller background errors compared to two-stage object detection algorithms (Qing et al., 2021; Zhou et al., 2022). Zhu et al. (2021) compared the detection speed and accuracy of three region-based convolutional neural network (R-CNN)-based two-stage detectors and four single-stage detectors, confirming that the single-stage detector, YOLOv5s, achieved the fastest detection speed at 63.3 frames per second (FPS).

The YOLO algorithm has evolved through various versions. In a study by Kim (2023), YOLOv1 and YOLOv2 are suitable for tasks where accuracy is critical, while YOLOv3 and YOLOv4 are primarily used for tasks that prioritize speed. On the other hand, YOLOv5 and YOLOv6 are effective in environments requiring a balance between accuracy and speed, whereas YOLOv7 and YOLOv8 are known to be applicable in scenarios with limited memory and computational resources. In this study, YOLOv5, which shows an excellent balance between accuracy and speed, was used considering the real-time nature of ship detection.

3.2. Late Fusion

Late fusion involves applying object detection models to data from various sensors and combining the detection results at the decision-making stage. This method maintains high detection accuracy and robustness, even when the performance of individual sensors deteriorates (Kim and Cho, 2021; Kim et al., 2019; Zhao et al., 2021). In this study, late fusion was used to consider the features of VV and VH images with different features, and the detection results of each single model were fused using a soft non-maximum suppression (NMS)-based late fusion method (Fig. 3).

Figure 3. Flowchart of the late fusion method.

A study by Farahnakian and Heikkonen (2020) explored ship detection using late fusion. They compared the accuracy of different fusion methods by integrating ship detection results from RGB and IR images, employing not only late fusion but also early and middle fusion techniques. Although various studies have explored fusing sensor detection results, this research is the first to improve ship detection accuracy by employing late fusion of dual-polarization data from SAR images.

4. Results

Before comparing the accuracy of single models and fused models, we first evaluated the accuracy of single models based on polarization and color maps to determine whether the polarization and color map of SAR images affect the accuracy. This study evaluates mean average precision (mAP) at IoU thresholds between 0.3 and 0.5, reflecting varying levels of localization precision, to comprehensively assess model performance.

4.1. Comparison of Single Model Accuracy based on Polarization and Colormap

As a result of comparing the accuracy according to polarization, the VH image had higher accuracy in more than 79% of all cases. In particular, it had significantly higher detection accuracy than VV, for the gray colormap; the difference was even greater at a threshold of 0.5. The accuracy of the VH image was 26% higher than that of the VV image. However, for the parula colormap, the detection accuracy was generally similar between polarization modes (Table 2). Next, we compared the accuracy of the gray and parula colormaps. The detection accuracy of the parula colormap was higher than that of the gray colormap in more than 80% of cases. The difference in accuracy between the gray and parula colormaps was greater for VV images than for VH images and became larger as the IoU threshold increased (Table 2).

Table 2 . Comparison of single model accuracy based on polarization and colormaps.

Input sizePolarizationColormapIoU threshold
0.30.40.5
400s × 400sVVGray73.6547.8621.65
VH89.7072.5746.54
VVParula91.3973.8744.37
VH93.2677.4646.73
500s × 500sVVGray89.5067.2636.49
VH94.3984.3061.38
VVParula94.3384.7864.76
VH96.1287.3468.55
600s × 600sVVGray92.2774.8050.88
VH97.1986.3067.04
VVParula95.6488.6672.36
VH96.8385.4871.42
700s × 700sVVGray94.2779.5948.57
VH96.2487.6074.61
VVParula97.3189.5275.19
VH96.1488.4574.09


4.2. Comparison of the Accuracy of Fusion Models based on Single Model Combinations

This study proposed a ship detection method that fuses the detection results of dual-polarization data of different colormap types to improve the accuracy of deep training models by utilizing both training data with different characteristics. To improve detection accuracy, we compared the accuracy of all fusion models that could be combined into four single models (VV gray, VV parula, VH gray, VH parula). Specifically, the detection accuracy of a fusion model, which consists of a training model with an image size of and a threshold of 0.4 for late fusion, was compared (Table 3).

Table 3 . Ship detection accuracy comparison according to multimodal types and IoU thresholds (Input size: 700s×700s / Multimodal threshold: 0.4).

Multimodal thresholdMultimodal inputIoU threshold
0.30.40.5
0.4VV grayVH gray95.8886.7369.79
VH grayVV parula96.7491.5978.81
VH grayVH parula97.8388.9175.97
VV grayVH parula95.7988.7068.97
VV grayVV parula94.8286.8064.88
VH parulaVV parula97.9290.8678.87


As a result, it was found that the fused model with 1) VH gray colormap and VV parula colormap, and 2) VH parula colormap and VV parula colormap as multimodal inputs has higher accuracy than the single model (for IoU threshold of 0.4 and 0.5). For an IoU threshold of 0.4, the accuracy of the VH gray colormap single model was 87.60%, and that for the parula VV single model had an accuracy of 89.54%, both of which were below 90%; by contrast, a fusion of these single models resulted in an accuracy level higher than 90%. Similarly, the accuracy levels of the single VH parula and VV parula models were 88.45% and 89.52%, respectively, whereas the fusion of these models resulted in > 90% accuracy. Both multimodal models had higher true positive rates than the single models, despite a nonsignificant difference in the number of detections (‘Det’ in Table 4) as shown in Table 4.

Table 4 . Comparison of results produced by two multimodal models with high detection accuracy and their corresponding single model predictions.

Predict resultInput
VH grayVH parulaVV parulaVH grayVH parula
VV parulaVV parula
Ground Truth (GT)731731731731731
Detected (Det)801776860840823
True Positive (TP)657662671678684
False Positive (FP)144114189162139


Improving ship detection accuracy consists of two components: increasing the rate of true positives and eliminating false positives. Factors such as ship wake and offshore structures can cause ships to be mistaken for noise whereas the multimodal model is more likely to detect ships among these confounding factors (top of Fig. 4a, blue circle in Fig. 4d). Conversely a bright area on a SAR image may be falsely detected as a ship due to noise from small islands or strong backscattering signals whereas the multimodal model will correctly recognize that these signals are not ships (left in Figs. 4b, c; yellow circle in Figs. 4e, f).

Figure 4. Part of ship detection results for (a–c) single and (d–f) multimodal models.

5. Conclusions

In this study, we compared the accuracy of four single models using dual-polarization and gray and parula color maps, as well as six fused models utilizing late fusion techniques to integrate data with diverse characteristics. The comparison of 24 cases of single models revealed that, in 79% of the cases, the VH image was more accurate than the VV image, with the difference being more pronounced in the gray colormap. This indicates that the VH gray demonstrates high accuracy. Conversely, when comparing results based on the colormap, the parula colormap showed higher accuracy in 83% of the cases, with the difference being more significant in the VV image. This suggests that the VV parula provides high accuracy. To summarize, it can be concluded that VH parula demonstrates the highest accuracy. This finding aligns with the accuracy comparisons observed in the fused models.

Results from comparing the accuracy of six fusion models show that the best model for the YOLOv5-based late fusion algorithm proposed in this study involved fusing two pairs of detection results: 1) VH gray colormap and VV parula colormap, and 2) VH parula colormap and VV gray colormap, using a multimodal threshold of 0.4 and an IoU threshold of 0.4 or 0.5. All single models showed <90% accuracy for an IoU threshold of 0.4, whereas the optimal fusion models showed >90% accuracy, indicating effective compensation for both false positives and non detections. The YOLOv5-based late fusion method reduced the false detection rate by eliminating multiple detection results, which was a limitation of the NMS method, through the adoption of soft-NMS. The results of this study can be used for fundamental data to develop a multisensor-based ship detection algorithm. Future research will explore ways to utilize various satellite data to improve ship detection accuracy.

Acknowledgments

This research was funded Korea Institute of Marine Science & Technology Promotion (KIMST) and by the Ministry of Oceans and Fisheries (RS-2024-00415504).

Conflict of Interest

No potential conflict of interest relevant to this article was reported.

Fig 1.

Figure 1.RGB values: (a) gray and (b) parula colormaps.
Korean Journal of Remote Sensing 2024; 40: 1141-1148https://doi.org/10.7780/kjrs.2024.40.6.1.21

Fig 2.

Figure 2.SAR images of ships captured using different polarization modes, expressed as (a) VV gray, (b) VV parula, (c) VH gray, and (d) VH parula colormaps.
Korean Journal of Remote Sensing 2024; 40: 1141-1148https://doi.org/10.7780/kjrs.2024.40.6.1.21

Fig 3.

Figure 3.Flowchart of the late fusion method.
Korean Journal of Remote Sensing 2024; 40: 1141-1148https://doi.org/10.7780/kjrs.2024.40.6.1.21

Fig 4.

Figure 4.Part of ship detection results for (a–c) single and (d–f) multimodal models.
Korean Journal of Remote Sensing 2024; 40: 1141-1148https://doi.org/10.7780/kjrs.2024.40.6.1.21

Table 1 . Description of sub-image data in this studyDescription of sub-image data in this study.

Sub-imageConditionPrerequisiteNo. of sub-images
400s × 400sRandomly crop the sub-image from the original image so that both sides have lengths between 400 and 499Contains at least three ships322
500s × 500sRandomly crop the sub-image from the original image so that both sides have lengths between 500 and 599
600s × 600sRandomly crop the sub-image from the original image so that both sides have lengths between 600 and 699
700s × 700sRandomly crop the sub-image from the original image so that both sides have lengths between 700 and 799

Table 2 . Comparison of single model accuracy based on polarization and colormaps.

Input sizePolarizationColormapIoU threshold
0.30.40.5
400s × 400sVVGray73.6547.8621.65
VH89.7072.5746.54
VVParula91.3973.8744.37
VH93.2677.4646.73
500s × 500sVVGray89.5067.2636.49
VH94.3984.3061.38
VVParula94.3384.7864.76
VH96.1287.3468.55
600s × 600sVVGray92.2774.8050.88
VH97.1986.3067.04
VVParula95.6488.6672.36
VH96.8385.4871.42
700s × 700sVVGray94.2779.5948.57
VH96.2487.6074.61
VVParula97.3189.5275.19
VH96.1488.4574.09

Table 3 . Ship detection accuracy comparison according to multimodal types and IoU thresholds (Input size: 700s×700s / Multimodal threshold: 0.4).

Multimodal thresholdMultimodal inputIoU threshold
0.30.40.5
0.4VV grayVH gray95.8886.7369.79
VH grayVV parula96.7491.5978.81
VH grayVH parula97.8388.9175.97
VV grayVH parula95.7988.7068.97
VV grayVV parula94.8286.8064.88
VH parulaVV parula97.9290.8678.87

Table 4 . Comparison of results produced by two multimodal models with high detection accuracy and their corresponding single model predictions.

Predict resultInput
VH grayVH parulaVV parulaVH grayVH parula
VV parulaVV parula
Ground Truth (GT)731731731731731
Detected (Det)801776860840823
True Positive (TP)657662671678684
False Positive (FP)144114189162139

References

  1. Chang, Y. L., Anagaw, A., Chang, L., Wang, Y. C., Hsiao, C. Y., and Lee, W. H., 2019. Ship detection based on YOLOv2 for SAR imagery. Remote Sensing, 11(7), 786. https://doi.org/10.3390/rs11070786
  2. Cheng, H. D., Jiang, X. H., Sun, Y., and Wang, J., 2001. Color image segmentation: Advances and prospects. Pattern Recognition, 34(12), 2259-2281. https://doi.org/10.1016/S0031-3203(00)00149-7
  3. Farahnakian, F., and Heikkonen, J., 2020. Deep learning based multi-modal fusion architectures for maritime vessel detection. Remote Sensing, 12(16), 2509. https://doi.org/10.3390/rs12162509
  4. Guo, Y., Chen, S., Zhan, R., Wang, W., and Zhang, J., 2022. LMSD-YOLO: A lightweight YOLO algorithm for multi-scale SAR ship detection. Remote Sensing, 14(19), 4801. https://doi.org/10.3390/rs14194801
  5. Im, Y., Youn, Y., Kang, J., Kim, S., Jeong, Y., and Choi, S., et al, 2023. Ship detection from SAR images using YOLO: Model constructions and accuracy characteristics according to polarization. Korean Journal of Remote Sensing, 39(5-3), 997-1008. https://doi.org/10.7780/kjrs.2023.39.5.3.9
  6. Kim, J., and Cho, J., 2021. A set of single YOLO modalities to detect occluded entities via viewpoint conversion. Applied Sciences, 11(13), 6016. https://doi.org/10.3390/app11136016
  7. Kim, J., Kim, J., and Cho, J., 2019 December. An advanced object classification strategy using YOLO through camera and LiDAR sensor fusion. In In Proceedings of the 2019 13th International Conference on Signal Processing and Communication Systems (ICSPCS), Gold Coast, Australia, Dec. 16−18, pp. 1-5. https://doi.org/10.1109/ICSPCS47537.2019.9008742
  8. Kim, J. Y., 2023. A comparative study on the characteristics of each version of object detection model YOLO. In Proceedings of the 2023 Korean Society of Computer Information Conference, Jeju, Korea, July 12-15, pp. 75-78.
  9. Li, J., Xu, C., Su, H., Gao, L., and Wang, T., 2022. Deep learning for SAR ship detection: Past, present and future. Remote Sensing, 14(11), 2712. https://doi.org/10.3390/rs14112712
  10. Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C. Y., and Berg, A. C., 2016. SSD: Single Shot MultiBox Detector. In: Leibe, B., Matas, J., Sebe, N., Welling, M., (eds.), Computer Vision -ECCV 2016, Springer, pp. 21-37. https://doi.org/10.1007/978-3-319-46448-0_2
  11. Lucchese, L., and Mitra, S. K., 2001. Color image segmentation: A state-of-the-art survey. Proceedings-Indian National Science Academy Part A, 67(2), 207-222.
  12. Miao, T., Zeng, H., Yang, W., Chu, B., Zou, F., Ren, W., and Chen, J., 2022. An improved lightweight RetinaNet for ship detection in SAR images. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 15, 4667-4679. https://doi.org/10.1109/JSTARS.2022.3180159
  13. Nie, X., Duan, M., Ding, H., Hu, B., and Wong, E. K., 2020. Attention mask R-CNN for ship detection and segmentation from remote sensing images. IEEE Access, 8, 9325-9334. https://doi.org/10.1109/ACCESS.2020.2964540
  14. Pelich, R., Chini, M., Hostache, R., Matgen, P., Lopez-Martinez, C., and Nuevo, M., et al, 2019. Large-scale automatic vessel monitoring based on dual-polarization Sentinel-1 and AIS data. Remote Sensing, 11(9), 1078. https://doi.org/10.3390/rs11091078
  15. Qin, P., Cai, Y., Liu, J., Fan, P., and Sun, M., 2021. Multilayer feature extraction network for military ship detection from high-resolution optical remote sensing images. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 14, 11058-11069. https://doi.org/10.1109/JSTARS.2021.3123080
  16. Qing, Y., Liu, W., Feng, L., and Gao, W., 2021. Improved Yolo network for free-angle remote sensing target detection. Remote Sensing, 13(11), 2171. https://doi.org/10.3390/rs13112171
  17. Redmon, J., 2016. You only look once: Unified, real-time object detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, June 27-30, pp. 257-276. https://doi.org/10.1109/JPROC.2023.3238524
  18. Song, J., Kim, D. J., and Kang, K. M., 2020. Automated procurement of training data for machine learning algorithm on ship detection using AIS information. Remote Sensing, 12(9), 1443. https://doi.org/10.3390/rs12091443
  19. Sun, Z., Leng, X., Lei, Y., Xiong, B., Ji, K., and Kuang, G., 2021. BiFA-YOLO: A novel YOLO-based method for arbitrary-oriented ship detection in high-resolution SAR images. Remote Sensing, 13(21), 4209. https://doi.org/10.3390/rs13214209
  20. Tang, G., Zhuge, Y., Claramunt, C., and Men, S., 2021. N-Yolo: A SAR ship detection using noise-classifying and complete-target extraction. Remote Sensing, 13(5), 871. https://doi.org/10.3390/rs13050871
  21. Tian, Y., Liu, J., Zhu, S., Xu, F., Bai, G., and Liu, C., 2022. Ship detection in visible remote sensing image based on saliency extraction and modified channel features. Remote Sensing, 14(14), 3347. https://doi.org/10.3390/rs14143347
  22. Wang, J., Lin, Y., Guo, J., and Zhuang, L., 2021. SSS-YOLO: Towards more accurate detection for small ships in SAR image. Remote Sensing Letters, 12(2), 93-102. https://doi.org/10.1080/2150704X.2020.1837988
  23. Wang, S., Gao, S., Zhou, L., Liu, R., Zhang, H., and Liu, J., et al, 2022. YOLO-SD: Small ship detection in SAR images by multiscale convolution and feature transformer module. Remote Sensing, 14(20), 5268. https://doi.org/10.3390/rs14205268
  24. Wang, Y., Wang, C., Zhang, H., Dong, Y., and Wei, S., 2019. A SAR dataset of ship detection for deep learning under complex backgrounds. Remote Sensing, 11(7), 765. https://doi.org/10.3390/rs11070765
  25. Zhang, T., Zhang, X., Liu, C., Shi, J., Wei, S., and Ahmad, I., et al, 2021a. Balance learning for ship detection from synthetic aperture radar remote sensing imagery. ISPRS Journal of Photogrammetry and Remote Sensing, 182, 190-207. https://doi.org/10.1016/j.isprsjprs.2021.10.010
  26. Zhang, Y., Fjeld, M., Fratarcangeli, M., Said, A., and Zhao, S., 2021b. Affective colormap design for accurate visual comprehension in industrial tomography. Sensors, 21(14), 4766. https://doi.org/10.3390/s21144766
  27. Zhang, Z., Zhang, L., Wang, Y., Feng, P., and He, R., 2021c. ShipRSImageNet: A large-scale fine-grained dataset for ship detection in high-resolution optical remote sensing images. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 14, 8458-8472. https://doi.org/ 10.1109/JSTARS.2021.3104230
  28. Zhao, J., Wang, Y., Cao, Y., Guo, M., Huang, X., and Zhang, R., et al, 2021. The fusion strategy of 2D and 3DD information based on deep learning: A review. Remote Sensing, 13(20), 4029. https://doi.org/10.3390/rs13204029
  29. Zhou, X., Jiang, L., Hu, C., Lei, S., Zhang, T., and Mou, X., 2022. YOLO-SASE: An improved YOLO algorithm for the small targets detection in complex backgrounds. Sensors, 22(12), 4600. https://doi.org/10.3390/s22124600
  30. Zhu, H., Xie, Y., Huang, H., Jing, C., Rong, Y., and Wang, C., 2021. DB-YOLO: A duplicate bilateral YOLO network for multiscale ship detection in SAR images. Sensors, 21(23), 8146. https://doi.org/10.3390/s21238146

Korean Journal of Remote Sensing