Although the spectral curves of the same objects under different conditions are somewhat different, they have the same feature points (upward, downward, maximum, and minimum) at the same spectral positions (
Figure 1). The WFCRMCA could detect abrupt signals through band-pass wavelet transform, such as crossing zero and extreme points. But crossing the zero point cannot be ensured a pulse signal, and perhaps is a smoothly changed signal, so the extreme points between adjacent zero points are much more critical. The signs in spectral vector format are classified according to the priority of importance from low to high: downward, upward, protruding, and concave (
Figure 2).
Wavelet feature clustering algorithms only analyze minutia data by detecting and determining the positions of abrupt signals. Using a fast binary Mallet wavelet algorithm [
22] in Equation 2 to extract wavelet coefficients, WFCRMCA can mark the upward-maximal points (
Figure 2a/a') and downward-minimal points (
Figure 2b/b') along the spectrum. WFCRMCA will overlook the weak signals if T
peak is large enough, leading to a failure in identifying some valuable signs among hidden objects.
The WFCRMCA uses
rij (correlation coefficient, CR), which works like a distance but not Euclidean distance as clustering criteria, to evaluate the difference between two spectral vectors on partial minutia. Equation 3 uses
Scale2-scale minutia of
Scale-scale wavelet coefficients to cluster,
ti,k is the
kth feature of
ith sampled vector,
N(·) is the number of feature positions that match criteria. WFCRMCA could use binary values to mark whether the position is valuable enough to attend clustering. When
, the bit number attending clustering comparison is
.
2.2. Markov Chain Clustering in Wavelet Feature Space
Wavelet-feature Markov clustering algorithm, i.e., WFCRMCA, first denoises the original data to make the spectral features more accurate, then uses a band-pass wavelet filter to detect all dot vectors for sharp points, including upward-maximal and downward-minimal points. As a result, simulated annealing Markov chain decomposition in state space, formed by evenly spatially sampled data, could realize the best centers at each temperature and sub-finest centers on the whole scope.
According to the peculiarity of simulated annealing Markov clustering, each clustering center is one state, and the space is a definite Markov state chain. If two classes (or states) merge, according to CR, it has nothing to do with other states. For example, for Markov chain in definite state space , if any two states communicate, they must be in the same class. Thus, the whole state space (pixels) could be separated into a few isolated classes according to transferred communication. T, which is defined as a threshold value of CR rij, is used as an annealing temperature to control the clustering process.
Def. 1 : If for states i and j, they have one-step transferred communication denoted as .
Theorem 1
: Communication can be transferred. If and (), .
According to Chapman-Kolmogorov equation:
Def. 2 : If feature i in wavelet characteristic space has pij=1, then i is an absorptive state, forming a single-dot set {i}.
Theorem 2. After Markov clustering, all wavelet features in wavelet feature space are frequently returned states.
i). A single-dot set is an absorptive, frequently returned state.
ii). As simulated annealing clustering causes T to be reduced gradually, the k+1th clustering iteration is supposed to create a non-single dot set. For instance, m pixels {1, 2,…, m} are absorbed into one class. Tk is the CR threshold of the kth iteration, and c1,…, i, … j, …, and cn are the created clustering centers of the kth iteration. Thus, .
During the k+1th iteration, , where Tstep is the depressed step of T at each iteration. If , then and , so , then i and j are merged together.
If
Tstep is small enough (i.e., the temperature is reduced slowly), and
i,
j,
l are absorbed in
k+1
th iteration,
so that it could be supposed as
Figure 3 that
.
As
m states communicate with each other, other
m-1 states could be seen as one state
j, let
pii=
pjj=
x, pij=1-
x,thus
So state i is frequently returned state. As m states communicate, the merged m states are frequently returned.
Theorem 3:
The sufficient and necessary condition of closed set C is that, for arbitrary elements ,
there exists (referring [23,24]).
Theorem 4 : Definite states of Markov chain in wavelet feature space can be uniquely decomposed without overlap into a definite number of frequently returned states, including closed sets C1, …, Cm and single dot sets Cm+1, …, Cn, existing:
Therefore, every state is frequently returned in the wavelet feature space at each temperature, and the number of isolated closed sets equals the number of classes. Then, the whole wavelet Markov chain feature state space has a decomposable expression that consists of several closed sets without overlap.
2.3. Adjustment of Clustering Centers
When two classes are merged whose correlation ratio rij is bigger than T, the numbers of each feature (including crossing zero part) are separately added up at the corresponding position ([0, b-1], the number of wavelet coefficients is approximately b). In addition, their sample numbers are also added up separately.
Similar to the traditional clustering method, reasonable adjustment of clustering centers is based on the statistic of intra-class features. For each position, the feature that occurs most frequently is chosen as the common feature of the new class, and then
b common features will be created. If several features come up at the same frequency, the feature with the highest priority (for example, downward-minimal or concave point > upward-maximal or protruding point) is chosen. Then, among all the class centers merged into one new class at this iteration, choose one pixel with the biggest CR with common features as a new class center. According to Equations 4 and 5,
is the statistic number of feature
k on the
lth position in class
i,
is the feature of the
lth position, which can be downward (0), upward (1), protruding (2), and concave (3).
R(
c1,
c2) is the correlation ratio between vector
c1 and
c2,
is the set of class centers absorbed by class
i, and
is the common features of class
i.
During the clustering process, many pixels with high similarity are merged, causing the number of class centers that will attend the following iterative clustering comparison to decrease sharply. As only newly created centers follow next-cycle clustering, WFCRMCA has a high clustering speed.
2.4. Wavelet-Feature Markov Clustering Algorithm
Based on the preceding theoretical analysis, the WFCRMCA uses a simulated annealing technique to gradually decrease CR threshold
T through Markov chain decomposition in wavelet feature space at each temperature, obtaining the best clustering centers of the whole space. Supposed that
ci is the class center of class
i, S
ci is the pixel set of class
i,
C is the set of all classes,
Zci is the set of class centers absorbed by class
i at the current temperature,
Nc is class number,
Ns is the number of sampled pixels (initial class centers are sampled pixels
),
R(
c1,
c2) is the CR between
c1 and
c2,
is the number of class centers absorbed by class
i,
is the pixel number in class
i,
Tstart is the initial value of CR
T, and
Tend is the lowest CR threshold. The detailed process of WFCRMCA is provided in the flow chart in
Figure 4. The simulated annealing Markov chain decomposition clustering in wavelet feature space is listed as follows:
Stepx and Stepy are the sampling distances along horizontal or vertical directions;
b, m, n are separately the band number, column number, and row number of original remotely sensed images;
Scale is wavelet transform scale;
Scale2 is the number of minutia scale attending clustering (i.e., Scale-Scale2 ~ Scale minutia sections)
- 2
Data preprocessing. Delete bands primarily affected by noise and atmosphere, such as the 1-6th, 33rd, 107-114th,153-168th, and 222-224th bands of AVIRIS. Multi-spectral images need to expand bands.
- 3
Apply band-pass Scale-scale wavelet filter (for example, Equation 6, [
25,
26]) to all pixels, search extreme points above noise threshold Tpeak between neighbor crossing zero points on each minutia section, and mark upward-maximal point as 1 and downward-minimal point as 2 at the corresponding position.
- 4
According to Stepx*Stepy sampling distance, sample the pixels and create Ns sampled pixels evenly.
- 5
Apply simulated annealing Markov state decomposition clustering to Scale-Scale2~Scale scale minutia sections of sampled data.
- a)
Set initial temperature T as Tstart, the clustering signal standard Tsignal (ratio of intra-class sampled pixel number over the number of total sampled pixels) is 1.0, and each pixel is one class center (beginning with Ns class centers). In the end, according to step b-e, apply Markov chain decomposition in state space to the wavelet features of sampled pixels by gradually depressing signal size.
- b)
Make judgments to all present class centers. If class i is a significant signal in which the number of pixels is more prominent than , move to the next class. Otherwise, search forward one by one for another class j whose size is smaller than , and make clustering judgments between class j and i.
- c)
According to Equation 8, if the CR between centers of two classes (i and j) meets the condition Pij=rij-T>0, then class j is absorbed into class i. Continuing this process b) until the last class is detected.
- d)
According to Equations 4 and 5, re-adjust the newly created centers: among all the class centers merged into one new class at this iteration, choose one pixel with the biggest CR with common features as a new class center.
- e)
Let T=T-Tstep decrease clustering temperature, and Tsignal= Tsignal /2 to reduce clustering size. Repeat steps a)-d) until T is reduced to the appointed small signal threshold Tend or set class number is reached.
- 6
According to the clustering centers created by 5), each pixel is clustered into one class whose center has the maximal CR.