1. Introduction
Natural disasters had always profound and far-reaching impacts on humanity. In recent years, we observe a climate change leading to escalating weather phenomena, which in turn facilitate natural disasters. Floods occur in the middle of the summer season due to sudden enormous amounts of rain, and dry weather conditions in combination with strong, out-of-season winds open the door for catastrophic, non-controllable wildfires. Earthquakes, volcano eruptions, and hurricanes appear with immense magnitude. All these disasters lead to loss of life and properties, disruption of services such as water supply, electricity, and transportation with further health risks, and have an enormous economic as well as psychological impact on the population.
Efforts to mitigate the impact of natural disasters include early warning systems, improved infrastructure resilience, disaster preparedness education, and international cooperation for humanitarian assistance. Preparedness and response strategies are crucial to minimizing the human toll and facilitating a quicker recovery from such events. Natural disaster detection systems contribute to early warnings, risk reduction, efficient resource allocation, and community preparedness. By leveraging technology and global cooperation, these systems play a vital role in minimizing the impact of disasters on both human populations and the environment.
Technological advancements and collaborative technologies contribute to the sharing of disaster information benefitting from different types of media. Deep Learning (DL) algorithms show promise in extracting knowledge from diverse data modalities, but their application in disaster response tasks remains largely academic. Systematic reviews evaluated the successes, challenges and future opportunities of using DL for disaster response and management, while also examining Machine Learning (ML) approaches, offering guidance for future research to maximize benefits in disaster response efforts [
1,
2]. In this work, we are particularly focused on floods. A concise summary of the research conducted is shown in
Table 1.
DL methods are increasingly applied to remote sensing imagery to address the limitations of traditional flood mapping techniques. Convolutional layer-based models offer improved accuracy in capturing spatial characteristics of flooding events, while fully connected layer-based models show promise when coupled with statistical approaches. Remote sensing analysis, multi-criteria decision analysis, and numerical methods are replaced with DL models for flood mapping in which flood extent or inundation maps, susceptibility maps, and flood hazard maps determine, categorize, and characterize the disaster, respectively [
3]. Furthermore, in a recent review, current DL approaches for flood forecasting and management are critically assessed, pointing out their advantages and disadvantages. Challenges with data availability and potential future research directions are examined. The current state of DL applications in this matter is comprehensively evaluated, showing that they are a powerful tool to improve flood prediction and control [
4].
Convolutional Neural Networks (CNNs) proved to be effective in flood detection using satellite imagery. High-quality flood maps are generated with the help of temporal differences from various sensors after CNNs identify changes between permanent and flooded water areas utilizing Synthetic Aperture Radar (SAR) and multispectral images [
6,
7]. Furthermore, the efficacy of CNNs in semantically segmenting water bodies in highly detailed satellite and aerial images from various sensors, with a focus on flood emergency response applications, is assessed by combining different CNN architectures with encoder backbones to delineate inundated areas under diverse environmental conditions and data availability scenarios. A U-Net model with a MobileNet-V3 backbone pre-trained on ImageNet consistently performed the best in all scenarios tested, while the integration of additional spectral bands, slope information from digital elevation models, augmentation techniques during training, and the inclusion of noisy data from online sources further improved model performance [
23]. In addition, Bayesian convolutional neural networks (BCNN) have been recommended to quantify the uncertainties associated with SAR-based water segmentation, because of their greater flexibility to learn the mean and the spread of the parameter posterior [
12]. Finally, a CNN employed to automatically detect inundation extents using the Deep Earth Learning, Tools, and Analysis (DELTA) framework demonstrated high precision and recall for water segmentation despite a diverse training dataset. Supplementary, the effects of surface obstruction due to the inability of optical remote sensing data to observe floods under clouds or flooded vegetation are quantified, suggesting the integration of flood models to improve segmentation accuracy [
21].
Since rapid damage analysis and fast coordination of humanitarian response during extreme weather events are crucial, flood detection, building footprint detection, and road network extraction have been integrated into an inaugural remote sensing dataset called SpaceNet 8 and a homonym challenge has been launched [
8]. The provided satellite imagery posed real-world challenges such as varying resolutions, misalignment, cloud cover, and lighting conditions and top performing DL approaches focusing on multi-class segmentation showed that swiftly identifying flooded infrastructures such as buildings and roads can significantly shorten response times. The contestants found that simple U-Net architectures yielded the best balance of accuracy, robustness, and efficiency, with strategies such as pre-training and data augmentation proving crucial to improve model performance [
9].
U-Nets and their variations were widely used to tackle the problems of water bodies segmentation and flood extend extraction. In [
27], a ResNet was used to replace the contracting path of the U-Net, due to its ability to solve vanishing or exploding gradient issues caused by error backpropagation through skip connections of the residual module. The normalized difference water index (NDWI) was employed to create pseudo-labels for training, resulting in an unsupervised DL approach. In [
15], another adjusted U-Net was proposed. With carefully selected parameters and training with pre-processed for three-category classification Sentinel-1 images, the proposed method was able to distinguish flood pixels from permanent water and background.
Transformers have also been successfully applied for semantic segmentation in remote sensing images. A novel transformer based scheme employing the Swin Transformer as the backbone to better capture context information and a densely connected feature aggregation module (DCFAM) serving as a novel decoder to restore resolution and generate accurate segmentation maps, proved to be effective in the ISPRS Vaihingen and Potsdam datasets [
22]. An improved transformer-based multiclass flood detection model capable of predicting flood events while distinguishing between roads and buildings was introduced, which, with an additional novel loss function and a road noise removal algorithm, achieved superior performance, particularly in road evaluation metrics such as APLS [
18]. Finally, the Bitemporal image Transformer (BiT) model scored highest in a change detection approach capturing the changed region better [
7].
A multiscale attentive decoder-based network (ADNet) designed for automatic flood identification using Sentinel-1 images outperforms recent DL and threshold-based methods when validated on the Sen1floods11 benchmark dataset. Through detailed experimentation on various dataset settings, ADNet demonstrates effective delineation of permanent water, flood water and all water pixels using both co-polarization (VV) and cross-polarization (VH) inputs from Sentinel-1 images [
5].
Dilated or atrous convolutions, which on the one hand increase the network’s receptive field while on the other reduce the number of trained parameters needed [
30], are utilized in an effort to speed up search and rescue operations after natural disasters such as floods, high tides, and tsunamis. FASegNet, a novel CNN-based semantic segmentation model, was introduced specifically designed for flood and tsunami area detection. FASegNet utilizes encoder and decoder networks with an encoder-decoder-residual (EDR) block to effectively extract local and contextual information. An Encoder-Decoder High-Accuracy Activation Cropping (EHAAC) module minimizes information loss at the bottleneck, and skip connections transfer information between encoder and decoder networks, outperforming other segmentation models [
20].
A novel weak training data generation strategy and an end-to-end weakly supervised semantic segmentation (WSSS) method called TFCSD, challenges urban flood mapping [
10]. By decoupling the acquisition of positive and negative samples, the weak label generation strategy significantly reduces the burden of data labeling, enabling prompt flood mapping in emergencies. Additionally, the proposed TFCSD method improves edge delineation accuracy and algorithm stability compared to other methods, especially in emergency scenarios where pre-disaster river data is accessible, or when using the SAM ([
31]) assisted interactive labeling method when such data is unavailable.
Satellites such as Sentinel-1 and Sentinel-2 play a key role in flood mapping due to their rapid data acquisition capabilities. Their effectiveness in mapping floods across Europe was evaluated in a study in which the results indicate that observation capabilities vary based on catchment area size and suggest that employing multiple satellite constellations significantly increases flood mapping coverage [
32]. The urgent need for real-time flood management systems by developing an automated imaging system using Unmanned Aerial Vehicles (UAVs) to detect inundated areas promptly is addressed, so that emergency relief efforts will not be hindered by current satellite-based imaging systems which suffer from low accuracy and delayed response. By employing the Haar cascade classifier and DL algorithms, a hybrid flood detection model combining landmark-based feature selection with a CNN demonstrated improved performance over traditional classifiers [
17].
Specially designed datasets are introduced to address the lack of high-resolution (HR) imagery relevant to disaster scenarios. In [
19], FloodNet, a high resolution (HR) UAV imagery dataset, capturing post-flood damage, aims to detect flooded roads and buildings and distinguish between natural and flooded water. Baseline methods for image classification, semantic segmentation, and visual question are evaluated, highlighting its significance for analyzing disaster impacts with various DL algorithms, such as XceptionNet and ENet. In [
14], an improved Efficient Neural Network architecture was also the choice to segment the UAV video of flood disaster. The proposed method consists of atrous separable convolution as the encoder and depth-wise separable convolution as the decoder.
To facilitate efficient processing of disaster images captured by UAVs, an AI-based pipeline was proposed enabling semantic segmentation with optimized deep neural networks (DNNs) for real-time flood area detection directly on UAVs, minimizing infrastructure dependency and resource consumption of the network. Experimental results confirm the feasibility of performing sophisticated real-time image processing on UAVs using GPU-based edge computing platforms [
11].
It becomes clear that DL methods offer improved segmentation by creating adaptive mapping relationships based on contextual semantic information. However, these methods require extensive manual labeling of large datasets and lack interpretability, suggesting the need to address these limitations for further progress. Traditional ML methods, on the other hand, rely on manually designed mappings. Systematic reviews of water body segmentation over the past 30 years examine the application and optimization of DL methods and outline traditional methods at both the pixel and the image levels [
33]. Evaluating the strengths and weaknesses of both approaches prompts a discussion of the importance of maintaining knowledge of classical computer vision techniques. There remains value in understanding and utilizing these older techniques. The knowledge gained from traditional Computer Vision (CV) methods can complement DL, expanding the available solutions. There also exist scenarios in which traditional CV techniques can outperform DL or be integrated into hybrid approaches for improved performance. Furthermore, traditional CV techniques have been shown to have benefits such as reducing training time, processing, and data requirements compared to DL applications [
34].
A near decade ago, a method for automatically monitoring flood events in specific areas was proposed using remote cyber-surveillance systems and image-processing techniques. When floods are treated as possible intrusion objects, the intrusion detection mode is utilized to detect and verify flood objects, enabling automatic and unattended flood risk level monitoring and urban inundation detection. Compared to large-area forecasting methods, this approach offered practical benefits, such as flexibility in location selection, no requirement for real-world scale conversion, and a wider field of view, facilitating more accurate and effective disaster warning actions in small areas [
16]. Real-time methods to detect flash floods using stationary surveillance cameras, suitable for both rural and urban environments, became quite popular. Another method used background subtraction to detect changes in the scene, followed by morphological closing to unite pixels belonging to the same objects. Additionally, small separate objects are removed, and the color probability is calculated for the foreground pixels, filtering out components with low probability values. The results are refined using edge density and boundary roughness [
25].
Unsupervised object-based clustering was also used for flood mapping in SAR images. The framework segments the region of interest into objects, converts them into a SAR optical feature space, and clusters them using K-means, with the resulting clusters classified based on centroids and refined by region growing. The results showed improved performance compared to pixel and object-based benchmarks, with additional SAR and optical features enhancing accuracy and post-processing refinement reducing sensitivity to parameter choice even in difficult cases, including areas with flooded vegetation [
26]. The same techniques were also proposed for flood detection purposes in UAV-captured images. Employing RGB and HSI color models and two segmentation methods: K-means clustering and region growing in a semi-supervised scheme showed potential for accurate flood detection [
13].
There is also a datacube-based flood mapping algorithm that uses Sentinel-1 data repetition and predefined probability parameters for flood and non-flood conditions [
24]. The algorithm autonomously classifies flood areas and estimates uncertainty values, demonstrating robustness and near-real-time operational suitability. It also contributed to the Global Flood Monitoring component of the Copernicus Emergency Management Service.
Contextual filtering on multi-temporal SAR imagery resulted in an automated method for mapping non-urban flood extents [
28]. Using tile-based histogram thresholding and refined with post-processing filters, including multitemporal and contextual filters, the method achieved high accuracy. Additionally, confidence information was provided for each flood polygon, enabling stable and systematic inter-annual flood extent comparisons at gauged and ungauged sites.
Finally, in [
29] an unsupervised graph-based image segmentation method has been proposed that aims to achieve user-defined and application-specific segmentation goals. This method utilizes a graph structure over the input image and employs a propagation algorithm to assign costs to pixels based on similarity and connectivity to reference seeds. Subsequently, a statistical model is estimated for each region, and the segmentation problem is formulated within a Bayesian framework using probabilistic Markov random field (MRF) modeling. Final segmentation is achieved through minimizing an energy function using graph cuts and the alpha-beta swap algorithm, resulting in segmentation based on the maximum a posteriori decision rule. In particular, the method does not rely on extensive prior knowledge and demonstrates robustness and versatility in experimental validation with different modalities, indicating its potential applicability across different domains. It was also successfully applied on SAR images for flood mapping.
From our survey, it becomes clear that supervised methodologies are preferred nowadays, as they outnumber unsupervised approaches (see
Table 1). Of the unsupervised ones, we found only one that deals with RGB images, but relies on change detection, thus is in need of the pre-disaster image as well.
In this paper, we propose a novel unsupervised method for flood segmentation utilizing color images acquired from UAVs. Without the need of large datasets, extensive labeling, augmentation, and training, the segmentation can be performed directly on the UAV deployed over the disaster area. Therefore, relief efforts can be swiftly directed to damaged sites avoiding time loss, which can be crucial in saving lives and properties. Initially, we employ parameter-free calculated masks over each component of the LAB colorspace utilizing as well an RGB vegetation index and the detected edges of the original image in order to provide an initial segmentation. Next, unsupervised image analysis techniques, such as distance transform, are adapted to the flood detection problem, producing a probability map for the location of flooded areas. Then, the hysteresis thresholding segmentation method is applied, resulting in the final segmentation. The main contributions of our work can be summarized as follows:
To our knowledge, this is the first fully unsupervised method for flood area segmentation in color images captured by UAVs. The current work faces for the first time the problem of flood segmentation based on parameter-free calculated masks and unsupervised image analysis techniques.
The flood areas are given as solutions of a probability optimization problem based on the evolution of an isocontour starting from the high confidence areas and gradually growing according to the hysteresis thresholding method.
The proposed formulation yields a robust unsupervised algorithm that is simple and effective for the flood segmentation problem.
The proposed framework is suitable for on-board execution on UAVs, enabling real-time processing of data and decision making during flight, since the processing time per image is about 0.5 sec without the need of substantial computational resources or specialized GPU capabilities.
The proposed system has been tested and compared with other variants and supervised methods on the Flood Area dataset introduced in [
35], consisting of 290 color images, yielding high-performance results. Furthermore, experimental results of the proposed method are also reported on the Flood Semantic Segmentation dataset [
36], which consists of 663 color images.
The rest of this paper is organized as follows.
Section 2 introduces the datasets used for this article.
Section 3 presents our proposed unsupervised methodology. The experimental results and a comprehensive discussion are given in
Section 4. Finally, conclusions and future work are provided in
Section 5.
5. Conclusions and Future Work
Overall, we presented a fully unsupervised approach for flood detection in color images acquired by flying vehicles such as UAVs and helicopters. The method progressively eliminates image regions identified as non-flood using binary masks generated from color and edge data. Our method operates in a fully unsupervised manner, with no need for training and ground truth labeling. We iteratively apply the same algorithmic steps to each color component. Subsequently, a binary map is generated for each component, discarding regions identified as non-flood and producing a final mask of potential flood areas (PFAs), refined through basic morphological operations. By weighting the pixels within the PFAs, we calculate an estimation of the dominant color of the flood, and a hysteresis thresholding technique is employed to achieve the final segmentation through probabilistic region growing of an isocontour. To the best of our knowledge, it is the first unsupervised approach to tackle this problem.
In this work, we showed that the following simple features suffice to accurately solve the problem of unsupervised flood detection. First of all, the flood’s color is similar wherever it appears within the image, and this color differs from the background. Almost always, the flood’s color is not green, assuming tree-like vegetation to be covered with water is extreme. Finally, in the LAB colorspace, the flooded area exhibits a higher value in at least one of the color components than the background. Color quality and camera rotation angle of the captured image contribute to the solidity of our observations, and thus a good amount of control over the flying vehicle while capturing the images will support the aforementioned inferences.
Experimental results confirmed that our proposed approach is robust, performs well in metrics, and is comparable to recent DL approaches, although not outperforming them. Furthermore, we introduced a categorization of the dataset according to the depicted scenery and camera rotation angle, into rural and urban/per-urban, and no sky and with sky, respectively. We showed that our approach performs well in all categories, it is slightly excels in segmenting floods in rural environments and is better suited for acquired images that do not contain sky, which is a controllable factor when maneuvering the UAV. The inference time is about half a second and it does not require GPU core processing capabilities. The method is suitable for on-board execution and the flood segmentation provided can be used to better guide relief efforts preventing loss of lives and mitigating the flood’s impact on the infrastructure.
In future research, we plan on to extend this work, detecting flooded buildings and roads. This will lead to a refinement of already existing flood segmentations, and a correction of erroneous segmentations, which the method now produces when the observations it relies on do not apply in the image. Combined with suitable methodologies which identify buildings and roads, such as [
47], and cross-correlating the results, we will be able to (a) avoid missclassifications of rooftops, building, and road pixels which have a color extremely similar to the flood, thus anticipating to attain improved outcomes and elevated scores in accurately segmenting the flood event, and (b) identify damaged buildings, when most of their circumference is adjacent to the flood, and flooded roads, when there exist discontinuities in the structure. These will help to even better assess the situation in the flood hit area and more accurately guide disaster assistance, evacuation, and recovery efforts. Furthermore, we plan to exploit the gained knowledge, in order to construct a specialized DL architectures, directing the network’s attention towards the flood and even incorporating our classical computer vision approach into hybrid deep learning frameworks, tackling the problem.
Figure 1.
Sample images from the Flood Area dataset (top) and their corresponding ground truths (bottom).
Figure 1.
Sample images from the Flood Area dataset (top) and their corresponding ground truths (bottom).
Figure 2.
Sample images from the Flood Semantic Segmentation dataset (top) and their corresponding ground truths (bottom).
Figure 2.
Sample images from the Flood Semantic Segmentation dataset (top) and their corresponding ground truths (bottom).
Figure 3.
Graphical abstract of our proposed approach.
Figure 3.
Graphical abstract of our proposed approach.
Figure 4.
The original images (from the Flood Area dataset) and the corresponding RGBVI masks on their right side. The masks show detected greenery with dim gray color. Examples are presented for (a) urban areas, (b) rural areas, and (c) poor or failed greenery detection. The remaining potential flood areas are shown in cyan.
Figure 4.
The original images (from the Flood Area dataset) and the corresponding RGBVI masks on their right side. The masks show detected greenery with dim gray color. Examples are presented for (a) urban areas, (b) rural areas, and (c) poor or failed greenery detection. The remaining potential flood areas are shown in cyan.
Figure 5.
Blue and red curves correspond on the average value of (a) L, (b) A and (c) B color components computed on flood and background pixels respectively, for each image of the Flood Area dataset, sorted in ascending order. The yellow curves show the corresponding threshold.
Figure 5.
Blue and red curves correspond on the average value of (a) L, (b) A and (c) B color components computed on flood and background pixels respectively, for each image of the Flood Area dataset, sorted in ascending order. The yellow curves show the corresponding threshold.
Figure 6.
Original images from the Flood area dataset and their corresponding LAB components masks from left to right. Note that the edges in are dilated for illustration purposes. Non-flood areas are depicted with dim gray color, whereas remaining potential flood areas are shown in cyan.
Figure 6.
Original images from the Flood area dataset and their corresponding LAB components masks from left to right. Note that the edges in are dilated for illustration purposes. Non-flood areas are depicted with dim gray color, whereas remaining potential flood areas are shown in cyan.
Figure 7.
Probability maps (column 4) obtained using potential flood areas of (column 2), weight maps (column 3), as generated by the distance transform and the corresponding images from the Flood Area dataset (column 1). Potential flood area is shown in cyan, and non-flood area in dim gray color. The weights and probabilities range from 0 (dark blue color) to 1 (red color).
Figure 7.
Probability maps (column 4) obtained using potential flood areas of (column 2), weight maps (column 3), as generated by the distance transform and the corresponding images from the Flood Area dataset (column 1). Potential flood area is shown in cyan, and non-flood area in dim gray color. The weights and probabilities range from 0 (dark blue color) to 1 (red color).
Figure 8.
(a) Original image from the Flood Area dataset, (b) the applied hysteresis thresholding on the decisive probability map of the potential flood area and (c) the final segmentation mask. (b) In red and blue are the pixels with and respectively. Cyan colored pixels are with , they do not surpass the lower threshold, and are subsequently classified as background. The non-flood areas, according to the mask, are colored with dim gray pixels. (c) The last column shows the final segmentation obtained from our proposed method, where the flood is in blue and the background is in dim gray color.
Figure 8.
(a) Original image from the Flood Area dataset, (b) the applied hysteresis thresholding on the decisive probability map of the potential flood area and (c) the final segmentation mask. (b) In red and blue are the pixels with and respectively. Cyan colored pixels are with , they do not surpass the lower threshold, and are subsequently classified as background. The non-flood areas, according to the mask, are colored with dim gray pixels. (c) The last column shows the final segmentation obtained from our proposed method, where the flood is in blue and the background is in dim gray color.
Figure 9.
High performance results of the proposed flood segmentation method from the Flood Area dataset. Original images, ground truth, and the final segmentation of our proposed method (UFS-HT-REM).
Figure 9.
High performance results of the proposed flood segmentation method from the Flood Area dataset. Original images, ground truth, and the final segmentation of our proposed method (UFS-HT-REM).
Figure 10.
Satisfactory results of the proposed flood segmentation method from the Flood Area dataset. Original images, ground truth, and our proposed method’s (UFS-HT-REM) final segmentation.
Figure 10.
Satisfactory results of the proposed flood segmentation method from the Flood Area dataset. Original images, ground truth, and our proposed method’s (UFS-HT-REM) final segmentation.
Figure 11.
Poor segmentations resulting from the proposed methodology (UFS-HT-REM) from the Flood Area dataset. Original images, ground truth and the final segmentation of our proposed method.
Figure 11.
Poor segmentations resulting from the proposed methodology (UFS-HT-REM) from the Flood Area dataset. Original images, ground truth and the final segmentation of our proposed method.
Figure 12.
The average values of , , , and computed on the Flood Area dataset for different values of (a) (with ) and (b) (with ).
Figure 12.
The average values of , , , and computed on the Flood Area dataset for different values of (a) (with ) and (b) (with ).
Figure 13.
Representative results of the proposed methodology (UFS-HT-REM) from the Flood Semantic Segmentation dataset. Original images, ground truth, and the final segmentation of the proposed method are shown from left to right.
Figure 13.
Representative results of the proposed methodology (UFS-HT-REM) from the Flood Semantic Segmentation dataset. Original images, ground truth, and the final segmentation of the proposed method are shown from left to right.
Table 1.
A brief overview of the research for this article depicting approach (supervised, unsupervised), modality, and method.
Table 1.
A brief overview of the research for this article depicting approach (supervised, unsupervised), modality, and method.
Authors |
Year |
Approach |
Imagery |
Method |
Chouhan, A. et al. [5] |
2023 |
Supervised |
Sentinel-1 |
Multi-scale ADNet |
Drakonakis, G.I. et al. [6] |
2022 |
Supervised |
Sentinel-1, 2 |
CNN change detection |
Dong, Z. et al. [7] |
2023 |
Supervised |
Sentinel-1 |
STANets, SNUNet, BiT |
Hänsch, R. et al. [8,9] |
2022 |
Supervised |
HR satelite RGB |
U-Net |
He, Y. et al. [10] |
2024 |
Weakly-
supervised |
HR aerial RGB |
End-to-end WSSS framework
structure constraints and self-distillation |
Hernández, D. et al. [11] |
2021 |
Supervised |
UAV RGB |
Optimized DNN |
Hertel, V. et al. [12] |
2023 |
Supervised |
SAR |
BCNN |
Ibrahim, N. et al. [13] |
2021 |
Semi-
supervised |
UAV RGB |
RGB and HSI color models,
k-means clustering, region growing |
Inthizami, N.S. et al. [14] |
2022 |
Supervised |
UAV video |
Improved ENet |
Li, Z. et al. [15] |
2023 |
Supervised |
Sentinel-1 |
U-Net |
Lo, S.W. et al. [16] |
2015 |
Semi-
supervised |
RGB (Surveillance
camera) |
HSV color model,
seeded region growing |
Munawar, H.S. et al. [17] |
2021 |
Supervised |
UAV RGB |
Landmark-based feature selection,
CNN hybrid |
Park, J.C. et al. [18] |
2023 |
Supervised |
HR satelite RGB |
Swin transformer in a Siamese-UNet |
Rahnemoonfar, M. et al. [19] |
2021 |
Supervised |
UAV RGB |
InceptionNetv3, ResNet50, XceptionNet,
PSPNet, ENet, DeepLabv3+ |
Şener, A. et al. [20] |
2024 |
Supervised |
UAV RGB |
ED network with EDR block and
atrous convolutions (FASegNet) |
Shastry, A. et al. [21] |
2023 |
Supervised |
WorldView 2, 3
multispectral |
CNN with atrous convolutions |
Wang, L. et al. [22] |
2022 |
Supervised |
True Orthophoto
(near infrared), DSM |
Swin transformer and DCFAM |
Wieland, M. et al. [23] |
2023 |
Supervised |
Satelite and aerial |
U-Net model with MobileNet-V3
backbone pre-trained on ImageNet |
Bauer-Marschallinger,
B. et al. [24] |
2022 |
Unsupervised |
SAR |
Datacube, time series-based
detection, Bayes classifier |
Filonenko, A. et al. [25] |
2015 |
Unsupervised |
RGB (surveillance
camera) |
Change detection, color
probability calculation |
Landuyt, L. et al. [26] |
2020 |
Unsupervised |
Sentinel-1, 2 |
K-means clustering, region growing |
Li, J. et al. [27] |
2022 |
Unsupervised |
Sentinel-2
Landsat |
NDWI, U-Net |
McCormack, T. et al. [28] |
2022 |
Unsupervised |
Sentinel-1 |
Histogram thresholding,
multi-temporal and contextual filters |
Trombini, M. et al. [29] |
2023 |
Unsupervised |
SAR |
Graph-based MRF segmentation |
Table 2.
Results for the categories of images existing in the Flood Area dataset. The images were divided according to the environmental zone into 1.(a) rural, and 1.(b) urban/peri-urban, and according to the camera rotation angle into 2.(a) no sky (low angle), and 2.(b) with sky (high angle).
Table 2.
Results for the categories of images existing in the Flood Area dataset. The images were divided according to the environmental zone into 1.(a) rural, and 1.(b) urban/peri-urban, and according to the camera rotation angle into 2.(a) no sky (low angle), and 2.(b) with sky (high angle).
Category |
|
|
|
|
|
1.(a) Rural |
83.6% |
82.3% |
78.4% |
80.3% |
78.2% |
1.(b) Urban/peri-urban |
85.4% |
78.4% |
78.7% |
78.5% |
76.9% |
2.(a) No sky |
85.2% |
81.6% |
78.0% |
79.8% |
77.9% |
2.(b) With sky |
84.4% |
76.1% |
79.5% |
77.7% |
76.1% |
All |
84.9% |
79.5% |
78.6% |
79.1% |
77.3% |
Table 4.
Comparison of our proposed approach with selected DL approaches on the Flood Area dataset. The metrics used are Accuracy (
), Precision (
), Recall (
), and F1-Score (
) (calculated as in Eq.
13) expressed in percentage, and Trainable Parameters (Tr. Par.) expressed in millions (M).
Table 4.
Comparison of our proposed approach with selected DL approaches on the Flood Area dataset. The metrics used are Accuracy (
), Precision (
), Recall (
), and F1-Score (
) (calculated as in Eq.
13) expressed in percentage, and Trainable Parameters (Tr. Par.) expressed in millions (M).
Method |
|
|
|
|
Tr. Par. |
FASegNet |
91.5% |
91.4% |
90.3% |
90.9% |
0.64 M |
UNet |
90.7% |
90.0% |
90.1% |
90.0% |
31.05 M |
HRNet |
88.6% |
84.8% |
92.0% |
88.3% |
28.60 M |
Ours |
84.9% |
79.5% |
78.6% |
79.1% |
0 M |
Table 5.
Quantitative findings for the Flood Semantic Segmentation dataset (FSSD) in comparison with the Flood Area dataset (FAD) and for the union of the two datasets (calculated as the weighted average due to the varying number of images). The number of images for each dataset (
), Accuracy (
), Precision (
), Recall (
), and F1-Score (
) (calculated as in Eq.
13) expressed in percentage are reported.
Table 5.
Quantitative findings for the Flood Semantic Segmentation dataset (FSSD) in comparison with the Flood Area dataset (FAD) and for the union of the two datasets (calculated as the weighted average due to the varying number of images). The number of images for each dataset (
), Accuracy (
), Precision (
), Recall (
), and F1-Score (
) (calculated as in Eq.
13) expressed in percentage are reported.
Dataset |
Images |
|
|
|
|
|
FSSD |
663 |
88.5% |
79.8% |
83.7% |
81.7% |
79.4% |
FAD |
290 |
84.9% |
79.5% |
78.6% |
79.1% |
77.3% |
FSSD ∪ FAD |
953 |
87.4% |
79.7% |
82.2% |
80.9% |
78.8% |