The changes on the Earth stem from both natural hazards (e.g., floods and earthquakes) and human activities (e.g., urban development) [
1]. Therefore, Change Detection (CD) algorithms are crucial tools for disaster and resource management. There are several methods of landscape CD, one of the most important of which is Remote Sensing (RS) [
2,
3]. RS data measure the changes between objects in a specific region over time [
4]. It is a valuable source of data with different advantages, including frequent coverage, the ability to monitor large-scale areas, and low cost [
5]. RS data can be used for various CD applications, including fire monitoring [
6,
7], climate change [
8,
9,
10,
11,
12,
13], and flood mapping [
14,
15,
16,
17].
One type of RS imagery that provides better spectral resolution is Hyperspectral RS imagery (HIS) [
18,
19,
20,
21]. HIS improves the process of CD for similar targets due to its high number of spectral bands [
22,
23], compared to multispectral imagery. The hyperspectral sensors can be divided into two categories: (1) airborne (e.g., Airborne Visible/Infrared Imaging Spectrometer (AVIRIS)); and (2) space-borne (e.g., Recursore IperSpettrale della Missione Applicativa (PRISMA), Enmap). In the near future, new space-borne sensors will be deployed (HyspIRI, SHALOM, and HypXIM) [
24]. Multiple studies have so far used HIS for CD [
24,
25,
26,
27]. The specific nature of HIS has made extracting multi-temporal imagery a significant challenge [
28]. As a result, this remains a dynamic and challenging area of study. Atmosphere status, noise levels, and data overload are among the most challenging factors affecting the results of HCD [
24]. Numerous methodologies have so far been proposed for HCD. For example, Ertürk et al. [
29] suggested a CD technique by applying sparse spectral unmixing to bi-temporal hyperspectral images. First, this method predicts the changed areas using the spectral unmixing method, then creates a binary change map by thresholding the abundance maps. Ertürk [
30] also designed an HCD framework based on a fuzzy fusion strategy; similarity measures indices, the spectral angle mapper (SAM) algorithm, and change vector analysis to predict changed areas. The fuzzy inference fusion strategy was used to fuse the magnitude and angle measurements obtained by the change vector analysis (CVA) and SAM algorithms, respectively. Additionally, López-Fandiño et al. [
26] proposed a two-step HCD framework for performing binary and multi-CD. They first generated a binary change map based on segmentation and thresholding and using the SAM algorithm. Then, the image differencing algorithm was used to combine multi-temporal images. The Stacked Auto-encoders algorithm was then employed to reduce the dimensionality of HIS. Finally, the binary change map and the reduced HIS were used to produce the multi-class change map. In recent work, Ghasemian and Shah-Hosseini [
31] also designed an HCD framework for multiple and binary CD based on several steps: (1) stacking the bi-temporal dataset and generating sample data based on the peak density clustering algorithm, (2) implementing target detection methods based on the produced sample data, (3) generating a binary change map based on the Otsu thresholding, (4) utilizing the sparse coding algorithm and the support vector domain description (SVDD) for generating multiple maps. Saha et al. Furthermore, [
32] proposed an HCD framework based on an untrained deep model for HCD. This method extracts deep features for the first and second times of hyperspectral images using the untrained model and measures the similarity of the deep features through the Euclidean norm. The Otus algorithm is used to threshold the predicted deep features, resulting in a binary change map. Tong et al. [
33] also proposed a framework for HCD by analyzing and transfer learning of uncertain areas. This method is applied in four main steps: (1) generating a binary change map according to the uncertain area analysis using K-Means clustering, CVA, and rule-based methods, (2) classifying the source image based on an active learning framework, (3) second-time image classification based on improved transfer learning and a support vector machine (SVM) classifier, and (4) utilizing post-classification analysis for multiple change map detection. Moreover, Seydi and Hasanlou [
34] designed an HCD method based on a 3D convolutional neural network (3D-CNN) and an image differencing algorithm. This framework utilized the image differencing procedure to predict change and no-change areas and then employed the 3D-CNN to classify the change areas to generate a binary change map. Finally, Borsoi et al.[
35] proposed a fast spectral unmixing method for HCD based on the high temporal correlation between the abundances. This method detects abrupt changes by considering the residuals of end-member selection.
Although the current HCD methods have shown promising results, they usually have several limitations, including the following items:
Given these limitations, a novel method has been proposed in this study to minimize the challenges and to improve HCD results. This study introduces a new framework for HCD based on double-stream CNNs called the HCD-Net. The HCD-Net uses multiscale 3D/2D convolution layers and 3D/2D attention blocks. The advantages of the HCD-Net are: (1) using the multiscale multi-dimensional kernel, (2) utilizing 3D/2D attention blocks, (3) high efficiency, and providing robust results in HCD. The key contributions of this study are: