Preprint
Article

Efficient Large-Scale Urban Parking Prediction: Graph Coarsening Based on Real-Time Parking Service Capability

Altmetrics

Downloads

49

Views

26

Comments

0

Submitted:

05 September 2024

Posted:

05 September 2024

You are already at the latest version

Alerts
Abstract
With the sharp increase in the number of vehicles, the issue of parking difficulties has emerged as an urgent challenge that many cities need to address promptly. In the task of predicting large-scale urban parking data, existing research often lacks effective deep learning models and strategies. To tackle this challenge, this paper proposes an innovative framework for predicting large-scale urban parking graphs leveraging real-time service capabilities, aimed at improving the accuracy and efficiency of parking predictions. Specifically, we introduce a graph attention mechanism that assesses the real-time service capabilities of parking lots to construct a dynamic parking graph that accurately reflects real preferences in parking behavior. To effectively handle large-scale parking data, this study combines graph coarsening techniques with temporal convolutional autoencoders to achieve unified dimension reduction of the complex urban parking graph structure and features. Subsequently, we use a spatio-temporal graph convolutional model to make predictions based on the coarsened graph, and a pre-trained autoencoder-decoder module restores the predicted results to their original data dimensions, completing the task. Our methodology has been rigorously tested on a real dataset from parking lots in Shenzhen. The experimental results indicate that compared to traditional parking prediction models, our framework achieves improvements of 46.8% and 30.5% in accuracy and efficiency, respectively. Remarkably, with the expansion of the graph’s scale, our framework’s advantages become even more apparent, showcasing its substantial potential for solving complex urban parking dilemmas in practical scenarios.
Keywords: 
Subject: Engineering  -   Transportation Science and Technology

1. Introduction

Urban transportation systems are facing serious challenges due to the continuous increase in the number of vehicles. Recent statistics [1] reveal that vehicle ownership in China escalated to 417 million in 2022, complemented by a striking addition of 34.78 million new registrations within the year. Particularly at the urban scale, as many as 13 cities saw an increase in car ownership by more than 3 million vehicles, including metropolises like Beijing, Shanghai, and Shenzhen. This surge in vehicle numbers not only leads to traffic congestion but also severely exacerbates parking issues. The expansion of urban parking facilities and spaces cannot keep pace with the growth in vehicle numbers, making it exceedingly difficult to find parking in city centers and commercial areas.
To alleviate the parking difficulty issue, urban-level smart parking platforms [2] have emerged. By leveraging cutting-edge technologies such as digital twins, the Internet of Things, and cloud computing [3], these platforms have successfully integrated urban parking facility resources, effectively solving the problem of information asymmetry between users and available parking spaces. Taking the Wuhan smart parking project [4] as an example, by the end of 2022, the platform had integrated more than 4,000 parking lots, including approximately 480,000 parking space data. With its portability, the platform offers convenience to citizens, who can easily check real-time parking availability in different areas through their mobile phones and use services such as parking reservation [5] and parking navigation [6,7,8] to achieve precise parking. In these platforms, the key to providing smart parking services lies in parking prediction technology [9]. Indeed, parking prediction has always been a hot research topic in urban parking technology [10,11]. By effectively predicting the availability of parking spaces in future time periods, it can not only enhance user experience but also improve parking turnover, optimize drivers’ parking behavior, and thereby help to allocate urban parking resources rationally.
Recent studies have framed such parking prediction tasks as spatio-temporal graph modeling problems [12]. On one hand, this is because parking lots are connected through urban road networks, forming unique parking graphs [13], for which graph convolutional networks can be utilized to capture the spatial characteristics of the parking graph [14]. On the other hand, given the tidal characteristics of parking flow, recurrent neural networks can be employed to fit the time series regression of parking data [15]. This approach, which considers both spatial and temporal dependencies, is known as spatio-temporal Graph Convolutional Networks (GCNs).
However, for the vehicle-dense cities mentioned above, with the rapid development of wireless sensors and the widespread use of vehicles [16], a massive influx of real-time parking data continuously flows into urban-level parking platforms. This leads to a dramatic increase in the volume of parking data managed by existing platforms and the coverage range of the parking Internet of Things network [17]. The rapid influx of such information necessitates an unsustainable amount of time to train spatio-temporal GCNs at the required scale for large-scale urban parking prediction.
To reduce the resources and time costs involved in training the aforementioned models, scholars have been dedicated to finding efficient graph dimensionality reduction techniques [18]. Among various studies, graph coarsening [19] stands out as one of the mainstream methods that can achieve this aim. Its core concept involves merging certain nodes within the graph into a hypernode and then using these hypernodes to construct a coarsened graph, thereby enhancing the learning efficiency and scalability of large-scale graphs [20]. However, a prevailing issue is that existing coarsening methods struggle to be directly applied to complex parking scenarios:
1.
The quality of the input graph significantly determines the performance of downstream tasks [21]. For parking graphs, a high-quality parking graph is not only an abstraction of spatial topology but also a adequate representation of parking behavior and decision-making. Neglecting either aspect will limit the understanding of parking scenarios, thereby negatively affecting the effectiveness of subsequent coarsening tasks as well as the accuracy and robustness of prediction tasks.
2.
Most of the existing coarsening methods only consider reducing the size of the graph structure [22], and the node features of the coarsened graph still being obtained by concatenating the original data. This means that the amount of data has not changed. If the model is trained directly based on the coarsened parking graph, the training overhead is not significantly reduced. Moreover, since the coarsened parking graph loses the original topology, the merged data corresponding to the hypernodes also lose their spatial features, leading to a decrease in prediction accuracy.
Graph attention networks (GAT) have attracted much attention in the transportation field [23] over recent years. By selectively focusing on specific nodes and discerning the importance among various nodes, these mechanisms help models to capture deeper hidden information in the traffic scenarios [24], thereby elevating predictive accuracy. The ParkingRank algorithm [25], which assesses the real-time service capability of parking lots using their static attributes, has been widely applied in parking recommendations [26], parking guidance [2], and other related tasks. We posit that this methodology can be interpreted as a preference-based attention mechanism, adept at capturing the service capability variations across different parking lots by evaluating their significance. This nuanced attention mechanism, by thoroughly considering node characteristics alongside topological connections, offers enriched guidance for the coarsening process, enhancing the model’s overall utility and effectiveness.
Meanwhile, autoencoders (AEs) have been widely used in traffic data compression and reconstruction tasks due to their theoretical ability to achieve lossless compression [27]. Specifically, the encoding module is able to learn the high-dimensional sparse information in traffic data and successfully encode it into a low-dimensional dense tensor [28] for more efficient traffic prediction, while the decoding module is tasked with reconstructing the condensed representations back to their original data form [29]. In this framework, we argue that the introduction of an AE can compensate for the lack of compression of node features in the graph coarsening process. Through the synergy of the encoding and decoding modules, a unified dimensionality reduction of the parking graph structure and features can be achieved, thus providing a more efficient and accurate data representation for the downstream parking prediction task. In summary, the main contributions of this paper can be summarized as follows:
  • We have proposed a method for constructing parking graphs that leverages a ParkingRank graph attention mechanism. This method intricately integrates the real-time service capacity assessment of parking lots into a graph attention network, creating a parking graph that accurately mirrors real-world parking behavior preferences. This graph is adept at capturing the complexity of parking scenarios, while also remaining flexible enough to adapt to its dynamic shifts, laying the foundation for downstream coarsening as well as prediction tasks.
  • We introduce a novel framework for parking prediction that employs a graph coarsening techniques and temporal convolutional autoencoder [30], designed to diminish the resource and time expenditures associated with urban parking prediction models. The scheme can make up for the shortcomings of traditional coarsening methods that neglect the dimensionality reduction of node features, and realize the unified dimensionality reduction of urban parking graph structure and features. Moreover, by incorporating temporal convolutional networks in the encoding-decoding phase, our approach not only significantly enhances the dimensionality reduction and reconstruction capabilities for parking time series data but also ensures the accuracy of the parking prediction task. Additionally, the compact data volume within each hypernode allows for the parallel processing of encoding-decoding operations across different sets of parking time series data, further boosting the overall training efficiency of the parking prediction task.
This paper is structured as follows: Section 2 provides a concise overview of spatio-temporal GCNs for parking prediction, techniques for dimensionality reduction in large-scale graphs, and the applications of Autoencoders (AE). Section 3 delves into the specifics of our innovative ParkingRank graph attention model and outlines our parking prediction methodology, which leverages TCN-AE for coarsening. Section 4 details the experimental framework and analysis employed to assess the effectiveness of our proposed approach. We conclude by summarizing the key insights and contributions of this study.

2. Related WORK

2.1. Spatio-Temporal Graph Convolutional Models

Spatio-temporal GCNs [31] have emerged as a powerful and prevalent approach within the realm of traffic prediction, adept at navigating the intricacies of traffic graphs by integrating both spatial and temporal dependencies. Unlike relying solely on GCNs or RNNs, spatio-temporal GCNs excel at dissecting time-series data embedded in graph structures [32], thereby yielding more precise prediction. For instance, Zhao et al. [33] introduced the T-GCN model, ingeniously combining GCN with GRU to understand the intricate topology of traffic networks and the dynamic shifts in traffic data, enhancing the capture of spatio-temporal characteristics more effectively. In a similar vein, Sun et al. [34] unveiled DDSTGCN, leveraging TCN to track variations in traffic states alongside GCN for topology learning, markedly boosting traffic prediction accuracy. In recent years, scholars have suggested employing attention mechanisms to bolster the extraction of spatio-temporal features, thereby extending the model’s interpretability. Zheng et al. [35] proposed a novel strategy that incorporates spatial and temporal attention layers to elucidate the spatio-temporal correlations within heterogeneous traffic flows, thereby not only elevating predictive performance but also the interpretability of the findings. Furthermore, Li et al. [36] developed AST-GAT, incorporating multi-head graph attention and attention-based LSTM modules to intricately map the spatio-temporal interdependencies among road segments, aiming to refine the precision of traffic prediction.
Due to the inherent structural and functional similarities between urban transportation networks and urban parking networks, both involve complex spatial relationships and dynamically changing temporal characteristics. Leveraging advanced spatio-temporal graph convolutional models from the field of traffic prediction, we believe can effectively enhance the accuracy of parking prediction. However, in the domain of parking prediction, the industrial sector continues to prioritize concerns regarding time costs and resource consumption, with a current lack of effective solutions. These challenges will limit the feasibility and effectiveness of the aforementioned spatio-temporal graph convolutional models in practical industrial applications.

2.2. Techniques for Dimensionality Reduction in Large-Scale Graphs

As the scale of graphs continues to expand, finding a universal method that can simplify the structure while preserving key attributes becomes increasingly important. Simplified graph representations not only facilitate storage but also offer efficiency in approximate algorithms [20]. There are mainly two methods for simplifying graphs: The first method, known as graph sparsification [37], approximates the maintenance of distance relationships between node pairs by removing edges from the graph. However, for highly structured parking graphs, this method might lose the original graph’s connectivity. An improved method is the K-neighbors Sparsifier [38], which performs local sparsification on each node by setting a threshold, but still introduces high computational complexity. With the growing popularity of deep learning, Wu et al. [39] proposed GSGAN, which generates new graphs through GANs to preserve the community structure of the original graph, but it introduces edges that do not exist in the original graph and contradicts reality.
The second approach is known as graph coarsening [40], which aims to reduce the number of nodes to a subset of the original nodes while maintaining the original graph properties. The key to graph coarsening [21] lies in the ability to accurately measure the variations before and after coarsening. The GraClus [41] technique, and the MCCA [42] algorithm all employ the idea of greediness by constantly compacting the connected regions of the original graph. However, these approaches lack guarantees for global optimization. Recently, spectral graph theory [22] proved that graphs with similar spectra are usually considered to have similar topologies. Therefore, the idea of minimizing the spectral distance can be introduced into the coarsening process, using the eigenvalues of the graphs to measure the structural similarity and ensuring that the generated coarsened graphs can maintain the properties of the original graphs.
However, we have observed that current graph coarsening methods primarily focus on reducing the complexity of the graph structure while often neglecting the dimensionality of node features. This approach does not effectively reduce the actual volume of parking data in large-scale urban parking prediction tasks. Although the parking network structure has been somewhat simplified through graph coarsening techniques, the training overhead for downstream parking prediction models has not decreased as significantly as anticipated.

2.3. Applications of Autoencoders

AEs have demonstrated remarkable proficiency in managing intricate spatio-temporal data relevant to transportation systems [35]. Typically employing symmetrically structured encoders and decoders for unsupervised learning, these AEs are adept not only at distilling input data into compact representations of potential vectors [27], but also at restoring the data to its original dimensions. They achieve this by optimizing the reconstruction of the data’s objective, thereby minimizing loss [29]. For instance, the TGAE introduced by Wang et al. [43] adeptly captures the underlying patterns of traffic flow through a spatio-temporal AE, facilitating the prediction of future traffic information. Similarly, the TCN-AE devised by Mo et al. [44] leverages TCNs within its encoding module to extract features from temporal data, subsequently utilizing the decoding module to map latent representations back to the original data space, thus enabling anomaly detection.
Inspired by the capabilities of AEs in dimensionality reduction and key feature extraction of spatio-temporal data, we propose the integration of AEs into large-scale urban parking prediction tasks, aiming to address the limitations of traditional coarsening methods in handling parking lot node features. Additionally, by leveraging the advantages of TCN in processing complex temporal data, AEs can more accurately capture and analyze the parking patterns within different parking lots, thereby significantly enhancing the performance and depth of understanding of downstream parking prediction models.

3. Methodology

3.1. Problem Definition

In this paper, our goal is to effectively reduce the time and resource costs of parking prediction models by achieving a unified reduction in the dimensions of the structure and feature data of urban parking graphs, while ensuring the accuracy of parking prediction tasks. The urban parking graph can be represented as G = ( V , E , A ) , where V denotes the set of parking nodes, E denotes the set of edges, and A R N × N denotes the directed adjacency matrix, N = | V | .
The parking prediction problem [12] can be interpreted as: given a parking graph G and historical parking data X ( t T + 1 ) : t for T time slices before time t, learning a function F to predict future parking data Y ^ ( t + 1 ) : ( t + T ) for T time slices after time t, as shown in Equation 1.
Y ^ ( t + 1 ) : ( t + T ) = F ( X ( t T + 1 ) : t , G ) ,
where the parking data X = { x 1 , x 2 , , x T } R T × N × F represents the parking flow information observed at T time slices for all parking lots. Each x t corresponds to a time slice and includes N parking lot nodes, each with F dimensional features, including the longitude, latitude, openness to the public, charging situation, and real-time occupancy rate of the parking lot.

3.2. ParkingRank Graph Attention

Graph coarsening, as the current mainstream graph dimensionality reduction technique, aims to overcome the huge computational obstacles faced by large-scale graph data when processing, extracting and analyzing. Typically, the input for graph coarsening is the graph adjacency matrix [21]. When dealing with complex structures such as urban parking graphs, the traditional adjacency matrix constructed from topological relationships is obviously difficult to adequately capture the complexity and diversity of the real parking landscape. In fact, drivers do not only consider the proximity of parking lots in the region when making parking decisions, but also comprehensively compare the distance of regional parking lots, real-time space occupancy rate, openness to the public, and charging situation, and other factors. Therefore, before performing the task of urban parking graph coarsening, we need a high-quality parking graph that can truly reflect the preference of parking behavior and adapt to the dynamic change of parking demand.
GAT [45] is an advanced spatial-based graph neural network methodology, whose primary advantage lies in redefining the aggregation of node features. Unlike GCN that typically assign equal weights to all neighboring nodes for information aggregation, GAT introduces an attention mechanism. This allows the model to dynamically allocate varying weights based on the importance of each neighboring node. Such a strategy enables GAT to more effectively capture spatial correlations within road networks and selectively emphasize those nodes that are crucial for the current task. In the application of urban parking network prediction, this feature of GAT is particularly significant. It enables the network to flexibly capture important relationships between parking nodes, focusing on those nodes that are critical for the current prediction task. We believe that this not only helps in more accurately describing the correlations between different parking lots but also aids in constructing an efficient parking network that reflects the relevancy of parking spaces. The overall framework of our proposed PRGAT is shown in Figure 1:
Given the complexity of the parking data X directly inputting the high-dimensional data x t into GAT would lead to considerable computational redundancy, thereby increasing the model’s time cost. To address this, we adopt a dimensionality reduction strategy that involves extracting and abstracting certain features of the parking lots. This effectively reduces the computational burden and enhances the model’s computational efficiency.
The ParkingRank model [25] is an algorithm to quantitatively assess the service capability of different parking lots. By integrating complex parking information such as the total number of parking spaces in the parking lot, the degree of openness to the public, and the price of parking, it can help drivers to understand the actual situation of the parking lot in a more comprehensive way, so that they can make informed choices in the parking decision-making process. Specifically, the algorithm primarily focuses on three aspects to describe the real-time service capacity of parking lots, as shown in Equation 2:
  • Parking lot service range: This aspect considers which types of vehicles are allowed to park in the parking lot. For example, parking lots at shopping centers may be open to all vehicles, while those in residential areas may only serve residents. Therefore, parking lots with a broader service scope generally have stronger service capabilities.
  • Total number of parking spaces: The more internal parking spaces a parking lot has, the stronger its service capacity usually is.
  • Price of parking: Higher parking prices may reduce the number of vehicles able to afford parking fees. Thus, expensive prices may lower the service capacity of the parking lot.
P R i = exp ( s i ) y i / y 1 + z i / z ,
where P R i represents the service capacity of each parking lot; s i [ 0 , 1 ] denotes the service range, which is used to measure the degree of openness of the parking lot, from the public parking lot ( s i = 1 ) to the private parking lot ( s i = 0 ); y i and z i represent the total number of parking spaces and the price of parking, respectively. y and z are the first-order paradigms for the y- and z-column vectors.
We believe that the process of quantifying the service capability of a parking lot is essentially a process of extracting the redundant parts of the parking lot features x t , and the low-dimensional evaluation results can be used as new abstract features to replace the original high-dimensional parking data x t .
Moreover, while GAT typically utilize Multi-Layer Perceptrons (MLP) or cosine similarity to ascertain the degree of association between nodes, such conventional methods may fall short in terms of interpretability, particularly when applied to entities like parking lots with distinct attributes. These methods often fail to offer a clear insight into the dynamics of parking lot interactions. To mitigate the black-box issue encountered by GATs in analyzing parking graphs, we integrate the parking spatio-temporal transfer matrix [2] into the computation of attention coefficients within GATs. This matrix meticulously accounts for the spatio-temporal evolution of parking lot features, encompassing real-time occupancy rates, service capabilities, and spatial connections, thereby facilitating a dynamic representation of parking cruising behavior. This nuanced incorporation allows for a richer, more detailed understanding of parking dynamics, effectively translating the raw data into actionable insights. The formulation of this transfer matrix A t r f is presented in Equation 3.
A trf = q i , when i = j w i , j 1 q j P R i , when i j ,
where each element α i , j trf of A trf R N × N represents the probability of a vehicle heading to another parking lot j when it finds parking lot i is full. q i represents the real-time occupancy rate of parking lot i, which also indicates the probability of a vehicle choosing to stay in parking lot i based on the parking situation. w i , j is the reciprocal of the normalized distance A i , j between parking lot i and parking lot j. P R i is the normalized service capacity of parking lot i.
Therefore, in this paper, the input to PRGAT consists of the original adjacency matrix A based on Euclidean distance and the new parking data X low R T × N × F low , which is composed of the service capacity P R i , the real-time occupancy rate q i , and latitude and longitude coordinates. The output is the updated parking data X re R T × N × F re . Here, F low represents the new feature dimensions after quantification and extraction, while F re denotes the new feature dimensions relearned by PRGAT.
The algorithm initially applies a linear transformation to each parking lot node’s features x t low ( i ) using a learnable weight matrix W R F re × F low to enhance the node’s expressive capability. Subsequently, it employs an attention mechanism T on the set of nodes to compute the attention coefficient e i , j between node i and node j. This procedure can be encapsulated by Formula 4, wherein the attention mechanism might be a function that signifies the correlation between two objects, like cosine similarity or a Multilayer Perceptron (MLP), and ∥ represents the vector concatenation operation.
e i , j = T W x t low ( i ) W x t low ( j ) ,
To capture local topological information and enhance computational efficiency, a masking mechanism and normalization operation have been introduced. The attention coefficients e i , j are confined within the first-order neighborhood of the nodes, enabling each node to concentrate solely on its directly connected neighboring nodes and disregard the other nodes in the graph, ultimately yielding the attention matrix A attn R N × N . See Formula 5 for details, where LeakyReLU represents the activation function.
A attn ( i , j ) = α i , j attn = exp LeakyReLU e i , j k N i exp LeakyReLU e i , k ,
Moreover, the masking mechanism ingeniously reflects the factors influencing the parking decision-making process. Drivers, when choosing a parking lot, tend to compare a specific parking lot with its adjacent ones, rather than conducting pairwise comparisons among all parking lots. Such a masking mechanism not only enhances the model’s sensitivity to local associations but also aligns more closely with the behavioral patterns in the actual parking decision-making process.
We input both the parking spatio-temporal transfer matrix A trf and the aforementioned attention matrix A attn into the softmax activation function simultaneously to derive the ParkingRank attention matrix A combined R N × N , as indicated in Equation 6.
A combined ( i , j ) = α i , j combined = softmax α i , j trf + α i , j attn ,
This ParkingRank attention matrix not only reflects the dynamic characteristics of the parking graph, capturing the flow trajectories and behavioral patterns of vehicles in urban parking scenarios, but also overcomes the shortcomings of insufficient interpretability of the parking lot relevance matrix computed by traditional graph attention methods.
After obtaining the normalized ParkingRank attention coefficients α i , j combined , GAT conducts a weighted aggregation of the features of each node i with its neighbors N i , thereby producing the final output for each node. This procedure is depicted in Formula 7, where σ denotes the sigmoid activation function.
x t r e ( i ) = σ j N i α i , j combined W x t low ( j ) ,
The PRGAT we designed is fundamentally a pre-trained model, its primary purpose being to serve as a front-end graph construction module within the overall framework for urban parking prediction. Its loss function is described in Formula 8, and the pseudocode for the graph is provided in Algorithm 1.
L G A T X low , X re = 1 T i = t T + 1 t X low ( i ) X re ( i ) 2 ,
Algorithm 1: ParkingRank Graph Attention (PRGAT)
  • Input:  A R N × N , X R T × N × F
  • Output:  A combined R N × N , X re R T × N × F re
1:
W MLP ( X , F )  
2:
a Correlation ( 2 × F , F )  
3:
A attn Zeros ( N , N )  
4:
A trf Zeros ( N , N )  
5:
for i 1 to N do 
6:
    P R i ParkingServiceCapacity ( X ( i ) )  
7:
    X low ( i ) ParkingReconstruction ( P R i , X ( i ) )  
8:
end for
9:
for i 1 to N do 
10:
    q i ParkingOccupancyRate ( X low ( i ) )  
11:
   for  j 1 to N  do 
12:
      s i , j Normalized ( A ( i , j ) )  
13:
      α i , j trf s i , j * ( 1 q i ) * P R i  
14:
      e i , j Concat ( W X low ( i ) , W X low ( j ) )  
15:
      α i , j ReLU LeakyReLU ( a T * e i , j )  
16:
      α i , j Masked MaskedFill ( α i , j ReLU , A ( i , j ) )  
17:
      α i , j attn Normalized ( α i , j Masked )  
18:
      A combined ( i , j ) softmax ( α i , j attn + α i , j trf )  
19:
   end for 
20:
end for 
21:
X re σ ( A combined W X low )  
22:
return A combined , X re  

3.3. Parking Graph Coarsening

Define the coarsened parking graph G c = ( V c , E c , A c ) is the result of coarsening the original adjacency matrix A, which is a smaller weighted graph. V c denotes a set of disjoint hypernodes, which covers all nodes in the original graph V, where each hypernode is formed by the aggregation of some of the nodes in the original graph, i.e., v i c = { v 1 , v 2 , , v n } R N . A c R N c × N c is the adjacency matrix of the coarsened graph, N c = | V c | .
This document employs the SGC algorithm [22], which utilizes spectral distance (SD) to demonstrate that the coarsened network retains spatial features similar to those of the original network. This concept is illustrated in the equation below:
S D G , G c = λ λ c 1 = i = 1 N λ ( i ) λ c ( i ) ,
where the vectors λ and λ c represent the eigenvalues of the Laplacian matrices of the original graph G and the coarsened graph G c , respectively. The spectral distance S D ( G , G c ) is considered sufficiently small if it is less than δ (where δ is a very small number). Only when this condition is met do the eigenvalues and eigenvectors of the two graphs exhibit similarity in the spectral domain. The calculated spectral distance can then substantiate that the coarsened graph significantly preserves the attributes of the original graph during the coarsening process. Thus, it can be inferred that the spatial structure of the coarsened network remains similar to that of the original network. The key steps of the SGC algorithm are introduced below.
The algorithm initiates by inputting the ParkingRank attention matrix A combined and the coarsening ratio n = N c N . It computes the Laplacian matrix L and, through matrix decomposition, obtains the first N c and the last N N c + 1 Laplacian eigenvalues λ and Laplacian eigenvectors u, for capturing both the local and global information of the parking graph. This procedure is illustrated in Formulas 10 and 11, wherein L signifies the normalized Laplacian matrix, with I N and D being the identity matrix and the degree matrix, respectively.
L = I N D 1 / 2 A combined D 1 / 2 ,
L u = λ D u ,
Then, the algorithm starts to iterate different eigenvalues intervals, performs parking clustering on the internal eigenvectors u respectively, divides the parking lots with similar features into a parking hypernode, and generates a preliminary coarsened graph G c based on the clustering results, the clustering principle is shown in Equation 12.
d c ( i , j ) = i = 1 , j = 1 n u i u j i = 1 n u i 2 j = 1 n u j 2 ,
The Laplacian eigenvectors u computed from the ParkingRank attention matrix A c o m b i n e d has rich parking graph topology information as well as parking lot node feature information, which can provide more accurate and intuitive parking lot node characterization for the coarsening process.
The process continuously calculates the distance S D between the Laplacian eigenvectors of the coarsened graph G c and those of the original graph, opting for the coarsening outcome with the minimum error. Finally, the coarsened graph G c = ( V c , E c , A c ) along with the corresponding index matrix i n d e x are returned. This procedure is depicted in Formula 13, with the coarsening specifics provided in Algorithm 2.
S D A combined , A c = d combined ( i ) d c ( i ) 1 ,
Algorithm 2: Parking graph Coarsening
  • Input:  A combined R N × N , n
  • Output:  A c R N c × N c , index  
1:
N c CoarsenedNodesNum ( n , N )  
2:
L NormalizedLaplacian ( A combined )  
3:
λ , u EigenValueAndVector ( L , N c )  
4:
for i 1 to N do 
5:
   for  j 1 to Neighbor ( i )  do 
6:
      A c temp , index ParkingSubZoneClustering ( u i , u j )  
7:
   end for 
8:
end for 
9:
for i 1 to N do 
10:
   for  j 1 to N  do 
11:
      A c arg min SD ( A combined ( i , j ) , A c ( i , j ) )  
12:
   end for 
13:
end for 
14:
return A c , index

3.4. Prediction Framework Based on Coarsened Parking Graphs

Since the coarsening process of the parking graph described above does not downscale the parking lot node features, the feature data corresponding to the parking hypernodes X combined R T × N × F low remains a splicing operation between the merged parking lot data, which is formalized as follows:
x t combined ( i ) = j v i c x t low ( i ) x t low ( j ) ,
If it is simply used as the feature input for the parking prediction models, it will not significantly improve the overall framework training efficiency, and even lose the intrinsic connection between the nodes. In this regard, we consider adopting the symmetric TCN-AE, which is because:
  • TCN is able to mine the intrinsic laws behind the parking time-series data itself [44], such as the tidal characteristics, which helps in the compression and reconstruction of the parking data.
  • The encoder is able to embed the high-dimensional sparse parking data into the low-dimensional dense tensor form, which reduces the computational overhead of the training model.
  • The decoder is able to achieve an approximate lossless reduction and can reconstruct the spatial structure of the original parking graph.
To maintain the consistency of the prediction results between the coarsened and the original parking graphs, we train a set of symmetric TCN-based AEs for each parking hypernode, utilizing the index matrix generated by the graph coarsening module. TCNs typically consist of multiple layers of dilated causal convolution, with each layer performing one-dimensional convolution. Moreover, to enhance the network’s expressive capacity, residual connections are often added following multiple convolutional layers, organizing these layers into several residual blocks. Each residual block primarily includes two dilated causal convolution layers, interspersed with normalization, activation, and dropout operations, before the output is added to the input of the next residual block. This design not only optimizes the flow of information but also enhances the model’s ability to capture features of parking data effectively.
The procedure is detailed as follows: As depicted in Figure 2, firstly, the parking data X low R T × N × F low undergoes encoding with the residual blocks within the encoder [44], as indicated in Formula 15. Owing to the reduction in the number of features as the quantity of residual blocks increases, it allows the feature data X c R T × N c × F low corresponding to the coarsened graph to be represented within a low-dimensional feature space.
X c ( i ) = X low * f ( i ) = j = 0 k 1 f ( j ) X low ( i j ) ,
where j = 0 k 1 f ( j ) denotes the filter f of length k and * denotes the convolution operation. This formula ensures that only information prior to the current time step is considered and future information is not computed.
Next, we input the adjacency matrix A c of coarsened graph and feature matrix X c into a spatio-temporal graph convolutional model. This paper takes T-GCN [33] as an example, which captures the spatio-temporal dependencies of the parking coarsened graph through 2 layers of GCN and 1 layer of GRU, finally obtaining the coarsened graph’s predicted results Y c R T × N c × F low through a linear transformation, as seen in Formulas 16, 17, 18, 19, and 20, where W and b represent the weights and biases of T-GCN, respectively, and u t , r t , c t , h t denote the reset gate, update gate, candidate hidden state, and current hidden state at time t,respectively. g c · denotes the graph convolution operation, and tanh is the activation function.
u t = σ W u · g c X c , A c , h t 1 + b u ,
r t = σ W r · g c X c , A c , h t 1 + b r ,
c t = tanh W c · g c X c , A c , r t , h t 1 + b c ,
h t = u t * h t 1 + 1 u t * c t ,
Y c = softmax W y · h t + b y ,
Due to the fact that the dimensions of the coarsened graph’s predicted results Y c do not align with the size of the original parking graph, these results are not inherently interpretable. To address this, decoding is necessary. Based on the prediction length, we opt for a particular pre-trained decoder, inputting the prediction of each hypernode to reconstruct the original parking graph’s prediction results Y ^ R T × N × F low , thereby fulfilling the prediction task for the entire urban parking graph. The decoding procedure is depicted in Formula 21. f 1 denotes the filters within the decoder, distinct from the encoding phase, as the feature count in the decoder’s residual blocks grows with the increase in residual blocks, aiming to reconstruct parking data within a high-dimensional feature space.
Y ^ ( i ) = Y c * f 1 ( i ) = s = 0 T 1 i = 0 k 1 f 1 ( i ) Y c ( i ) ,
The loss functions for the pre-trained TCN-AE and the parking prediction model utilize Mean Squared Error (MSE) and Huber Loss, respectively, as detailed in Formulas 22 and 23. In this context, X ^ R T × N × F low represents the outcome post-reconstruction by the autoencoder, Y signifies the real observed data, and θ represents the threshold parameter that controls MSE.
L A E X low , X ^ = 1 T i = t T + 1 t X low ( i ) X ^ ( i ) 2 ,
L predict ( Y , Y ^ ) = 1 2 ( Y Y ^ ) 2 , | Y Y ^ | θ θ | Y Y ^ | 1 2 θ 2 , otherwise ,

4. Experiments

4.1. Experiment Setup

This paper selected 9000 parking lots within five districts including Bao’an and Luohu in Shenzhen as research subjects. The parking data recorded for six months starting from June 1, 2016, was sampled at 15-minute intervals. These datasets encompassed key information such as the geographic coordinates, open status, pricing details, and real-time occupancy of each parking lot. To ensure data quality, we conducted deduplication and visualization-based outlier detection operations on the raw parking data. Additionally, to maintain consistency in data processing, all datasets underwent normalization before analysis. Finally, we partitioned the dataset chronologically into training set (70%), validation set (20%), and test set (10%) to ensure the validity and reliability of the experiments. Specifically, the distribution of parking lots in Bao’an District was visually presented through Figure 3, and specific details of the parking lot data are referred to in Table 1.
To evaluate the performance of our method, we selected several commonly used evaluation metrics in regression tasks: MAE, RMSE, and MAPE. The expressions for these metrics are as follows:
MAE = 1 T N t = 1 T i = 1 N | y t i y ^ t i | RMSE = 1 T N t = 1 T i = 1 N ( y t i y ^ t i ) 2 MAPE = 100 % T N t = 1 T i = 1 N y t i y ^ t i y t i
where y ^ t i and y t i respectively represent the predicted result and the ground truth for the i-th parking lot at time t. T denotes the length of the time sample, and N is the total number of parking lot nodes.
In the experimental parameter design, we use data from the past 12 time steps to predict data for the next 1, 2, and 4 steps, which corresponds to using the past 3 hours of parking traffic to predict the parking traffic for the next 15 minutes, 30 minutes, and 60 minutes, respectively. Additionally, to evaluate the predictive performance, we selected three widely used baseline models in the parking prediction field for comparison. Additionally, to explore the effectiveness of graph coarsening techniques, we also employed the recently popular graph sparsification methods as a benchmark:
  • T-GCN [33]: A classical model for traffic prediction that combines Graph Convolutional Networks (GCNs) and Gated Recurrent Units (GRU) to establish spatio-temporal correlations among traffic data.
  • STGCN [46]: Utilizes multiple ST-Conv blocks to model multi-scale traffic graphs, proven to effectively capture comprehensive spatio-temporal correlations.
  • STSGCN [47]: By stacking multiple STSGCL blocks for synchronous spatio-temporal modeling, it can effectively capture complex local spatio-temporal traffic correlations.
  • SparRL [37]: A universal and effective graph sparsification framework implemented through deep reinforcement learning, capable of flexibly adapting to various sparsification objectives.
For the training of various models, we utilized V100 graphics cards and implemented early stopping and dynamic learning rate adjustment mechanisms to optimize the training process. To ensure completeness and transparency of information, we have listed the configuration details of the models along with all training-related hyperparameters in Table 2.

4.2. Modeling Real Scenes

To demonstrate the graph construction process of the PRGAT method, we conducted an experiment using real parking lot data from Bao’an District in Shenzhen. In Bao’an District, the number of parking lots exceeds 1,000, leading to the number of reachable edges between parking lots reaching the order of millions. Therefore, to simplify the analysis while ensuring the practicality of the results, we selected a more central CBD area within the district as the study area, which contains over 600 parking lots. Additionally, we set a spatial distance threshold of 500 meters; distances between any parking lots exceeding this threshold were considered unreachable to align with real parking scenarios.
Shown in Figure 4 is a visualization of the parking graph constructed using the PRGAT model for the Baoan core area around evening during the week, where the different sizes of circles represent the service capacity of the parking lots, with larger circles implying stronger service capacity and more popularity among drivers. Unlike the traditional spatial adjacency matrix, the dynamic adjacency matrix generated by PRGAT can intuitively reflect the real-time importance of each parking lot, which helps the downstream prediction model to accurately determine the key parking lots in each time period, thus effectively identifying the places with strong or weak parking demand. Meanwhile, the graph composed of these parking lots with strong service capacity essentially depicts the parking backbone network of the city in different time periods, and this dynamic parking backbone network reflects the core areas and hotspots of the city’s parking demand, which provides an important perspective for understanding the city’s parking behavior pattern.
Figure 5 depicts the distribution of parking lots in the core area of Bao’an under different coarsening ratio (coarsened vertices/original graph vertices). By adjusting the coarsening ratio, we can clearly see significant changes in the structure and density of the parking graph: parking lots with stronger service capacities gradually absorb those with weaker capacities, thus forming larger parking hypernodes with enhanced comprehensive service capabilities. As the ratio of coarsening is progressively reduced, the backbone of the urban parking graph also becomes clearer. However, this coarsening process also poses certain challenges: a simplified parking graph means the loss of some important parking information, which could affect the downstream model’s ability to capture details, especially in areas highly sensitive to parking demand or where demand changes rapidly.

4.3. Selection of Coarsening Ratio

Considering the dual role of the coarsening ratio in the speed of model convergence and model prediction accuracy, we set up control experiments with coarsening ratio of 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, and without coarsening (coarsening ratio of 1), to study the impact of different coarsening ratio on the prediction of parking occupation rate during the evening rush hour. The prediction results are shown in the Figure 8, where dashed and solid lines respectively represent the impact of different coarsening ratio on model prediction accuracy and convergence speed.
As seen in Figure 6, as the coarsening ratio decreases from 1 to 0.6, the error growth rate of the model predictions remains relatively stable. This indicates that within this range of coarsening ratio, the model primarily loses redundant and secondary parking information, the absence of which has limited impact on overall prediction accuracy. However, as the coarsening ratio is further reduced, the prediction error of the models rapidly increases, particularly when the coarsening ratio is adjusted to 0.2, introducing calculation errors of 77.2%, 64.2%, and 61.8%, respectively. This suggests that crucial parking information (such as the highly volatile characteristics of parking lots) has been stripped away in the coarsening process.
Figure 6. Impact of Different Coarsening Ratio on Prediction Performance.
Figure 6. Impact of Different Coarsening Ratio on Prediction Performance.
Preprints 117338 g006
It’s noteworthy that the curve of model training duration presents a different trend. With the reduction of the coarsening ratio, the initial training time decreases rapidly, showing that coarsening can significantly reduce the computational burden of the model. However, after reaching a coarsening ratio of 0.6, the trend of reduction in training duration becomes more gradual, indicating that further coarsening does not significantly enhance efficiency. Specifically, in the coarsened graph, nodes of parking lots with significantly different characteristics may be merged, leading to a loss of crucial detail information that was originally observable. Consequently, the model’s ability to distinguish these differences is impaired, thereby affecting its predictive and generalization capabilities.
The two sets of experiments indicate that the ideal coarsening ratio should be within the range of 0.5 to 0.6. Coarsening within this interval manages to strike a balance between simplifying model complexity and retaining key feature information, which is particularly crucial when dealing with large-scale urban parking graphs, as these often contain a vast number of parking lot nodes and edges. Moreover, this phenomenon mirrors the real-world situation where approximately 50%-60% of parking lots make the primary contribution to urban parking demand. These parking lots are typically the ones with the highest usage frequency, optimal locations, and best service capabilities. However, it should be noted that the coarsening rate range we provide is merely idealized. The specific choice of coarsening rate must be dynamically adjusted based on the structure of the urban parking network and characteristics of parking data (such as parking frequency, peak periods) using adaptive deep learning models.

4.4. Quantitative Results

To explore the predictive performance of the urban parking prediction framework we proposed, we conducted predictions on the future 15 minutes, 30 minutes, and 60 minutes parking occupancy rates of various parking lots in Bao’an District during weekday evening rush hours using three baseline models, with the coarsening ratio set to 0.6. The corresponding prediction results are presented in Table 3.
Our experimental results demonstrate that, whether for short-term or long-term predictions, models using PRGAT as a pre-construction unit for graphs almost always outperform traditional parking prediction methods across all metrics. Specifically, taking TGCN as an example, compared to the default adjacency matrix, the ParkingRank attention matrix created by PRGAT achieves an average reduction of about 46.8% in MAE. Furthermore, compared to the traditional GAT mechanism, it also reduces RMSE by about 20.5%. This significant improvement in performance is attributed to the comprehensive considerations PRGAT takes in constructing dynamic parking graphs. Unlike traditional methods that focus solely on the spatial relationships between parking lots or on a single feature like available parking spaces, PRGAT also incorporates other critical information including real-time service capabilities such as the available capacity, level of accessibility, and cost of parking lots. This approach results in a graph that is not just a simple representation of spatial proximity but a richer, more accurate depiction of actual parking behavior patterns, greatly enhancing the model’s understanding and predictive ability in complex parking scenarios.
In terms of enhancing model training efficiency, compared to baseline methods, our framework achieves an average improvement of approximately 30.5%, further confirming the efficiency of our framework in handling complex parking prediction tasks. Furthermore, in the choice of graph dimensionality reduction techniques, graph coarsening outperforms graph sparsification in reducing model training duration. Graph sparsification primarily reduces the scale of the graph by decreasing the number of edges, but since the dimensionality of node features remains unchanged, its impact on training efficiency is not significant. Additionally, the loss of crucial connectivity information between parking lots can severely affect the predictive accuracy of downstream models. In contrast, graph coarsening, by merging nodes, not only reduces the scale of the parking network but also simplifies the model’s computational complexity while preserving key structural information, thereby aiding in maintaining the accuracy of the model’s predictions.
Notably, we optimized the AE by incorporating TCN to improve the processing of parking time series data, achieving near-lossless compression and restoration of large-scale parking data. Compared to previous work [48], this approach effectively keeps the MAPE within 12.6%, further increasing the accuracy of predictions.
It is important to recognize that although the application of coarsening techniques can improve model training efficiency, it inevitably introduces some level of error. Hence, our aim is to maximize training efficiency while striving to preserve prediction accuracy as much as possible. As the graph scale expands, our framework not only demonstrates a distinct advantage in prediction accuracy but also reveals increasingly significant differences in efficiency. This is especially evident when handling large-scale, dynamically changing complex urban parking graphs, showcasing the exceptional performance and practical value of our framework.

4.5. Scalability Results

To verify whether our proposed framework can adapt to parking graphs of different scales, we simulate the expansion of the parking graph scale by gradually increasing the number of parking lots. In the experiments, the coarsening ratio is set to 0.6, with the prediction of parking occupancy rates for the next 15 minutes during the evening rush hour as the evaluation task. This approach allows us to observe changes in the model’s prediction accuracy and convergence speed on parking graphs of various sizes. Synthesizing Figure 7a,b, it can be observed that when the scale of the parking graph is small, our framework shows relatively conservative improvements in terms of accuracy and efficiency, with an average accuracy improvement of about 11.1% and a training efficiency improvement of about 23.7%. Importantly, as the scale of the graph expands, the performance advantages of our prediction framework become particularly pronounced. Unlike the baseline models, which exhibit a higher error growth rate and exponential increase in iteration costs, our framework demonstrates a stable growth trend, significantly surpassing traditional methods in terms of error control and training efficiency.
Figure 7. Comparison of Prediction Performance for Future 15-Minute Parking Lot Occupancy Rates Across Parking Graphs of Various Scales (a) MAPE (b) Convergence Situation.
Figure 7. Comparison of Prediction Performance for Future 15-Minute Parking Lot Occupancy Rates Across Parking Graphs of Various Scales (a) MAPE (b) Convergence Situation.
Preprints 117338 g007
This performance advantage is mainly attributed to the efficient algorithms and model structures adopted within our framework, such as the PRGAT and TCN-AE. The former effectively captures key features of parking data, while the use of the latter significantly enhances the model’s capability to extract and analyze parking lot features. This ensures that the prediction framework can retain critical information from the original parking graph to a great extent, even after coarsening processing. In contrast, traditional parking prediction techniques, even when combined with graph sparsification, still face issues with suboptimal prediction performance and training speeds. This is due to graph sparsification often losing significant parking detail information as the network scale increases, severely impacting predictive performance. Additionally, since this technique does not involve dimension reduction of node features, the growth in training efficiency is similar to baseline models that do not employ any graph dimension reduction techniques.
In addressing the challenges of large-scale urban parking prediction, our framework optimizes data flow and information processing, effectively managing complex and voluminous parking data. This not only ensures high accuracy in model predictions but also significantly reduces training times, thereby meeting the needs for timeliness and accuracy. Such capabilities enable our framework to quickly and accurately respond to dynamic changes in urban parking, providing real-time predictive results.

4.6. Ablation Results

To further validate the rationality of each component within the framework, we conducted an ablation study based on the task of predicting the occupancy rate of parking lots in the Bao’an District for the next 15 minutes during the evening rush hour. The ablation experiments were divided into two parts, separately assessing the impact of PRGAT and TCN-AE on three sets of baseline models, with the coarsening ratio set to 0.6 in the experiments.
Figure 8(a) shows the experimental results of parking predictions using various types of adjacency matrices. Clearly, among all baseline methods, the ParkingRank attention matrix constructed by PRGAT exhibits the best performance, achieving an average reduction in relative error of about 38.3%. PRGAT, by integrating the attention coefficient matrix with the parking spatio-temporal transition matrix, delves deep into the feature and spatial dependencies between parking lots, effectively restoring real parking demand scenarios. In contrast, traditional static adjacency matrices, built solely on spatial location information, overlook the intrinsic characteristics of parking lots and their interactions with the surrounding environment. Often, they fail to fully capture the complex interactions and dynamic changes between parking lots, limiting the authenticity and dynamism of the scenarios captured by the model, and thus performing poorly in downstream prediction tasks. Moreover, compared to the traditional GAT mechanism, another advantage of PRGAT is that it not only offers a more detailed and dynamic method to represent parking graphs, but it also intuitively reflects which parking lots have stronger comprehensive service capabilities at specific moments, thus becoming a key basis for making parking decisions.
However, a limitation of the PRGAT is that it aggregates node feature information considering only the 2nd-order neighborhood, neglecting the potential connections between distant nodes. This may lead to inadequate capturing of the interactions between remote parking lots during holidays or major events. Therefore, future research exploring and implementing graph convolutional networks that include higher-order neighborhood information will be crucial. This will enable the model to capture a broader range of node interactions, thereby enhancing the accuracy and practicality of the predictions.
Figure 8(b) illustrates the effect of the AE on the prediction framework. From the figure, it can be found that not adding the AE leads to an average decrease in prediction accuracy of about 30.1%. This is because the coarsened parking graph changes its original spatial relationship, and simply splicing a set of vertex features corresponding to each hypernode will lead to redundancy of hypernode information and dimensionality inflation, thus limiting the model’s generalization ability. In addition, the loss of the intrinsic connection between nodes will also lead to the model’s inability to fully exploit the spatial dependency of the coarsened graph. On the contrary, the approach using autoencoder can selectively extract and reconstruct the key features inside the hypernodes, thus effectively adapting to the spatial changes inside the hypernodes. Meanwhile, we additionally introduce TCN in the encoding-decoding process, which enhances the model’s ability to learn the laws behind the dynamic changes of the parking lot view from the time-series data, and is more helpful for the compression and reconstruction of the parking data, which achieves an average 20% accuracy improvement.
Figure 8. Ablation Study on Future 15-Minute Parking Prediction (a) Graph Construction (b) Autoencoder.
Figure 8. Ablation Study on Future 15-Minute Parking Prediction (a) Graph Construction (b) Autoencoder.
Preprints 117338 g008

5. Conclusion

In this study, we propose an innovative large-scale urban parking prediction framework based on real-time service capabilities. Specifically, we first introduce a GAT mechanism based on real-time service capabilities, constructing a dynamic parking graph that reflects real-time parking behavior preferences. This aids spatio-temporal GCNs models in more accurately capturing the interactions and trends of change between parking lots. Subsequently, we designed a large-scale urban parking prediction framework based on coarsening and TCN-AE, aimed at effectively reducing the training time required for models to process complex parking data. Finally, we conducted experiments on a real dataset from Shenzhen’s parking lots to validate the performance of our proposed prediction framework. The experimental results show that our framework significantly improves both accuracy and efficiency compared to traditional parking prediction methods. Furthermore, as the graph scale expands, our framework’s performance exhibits even more pronounced advantages, underscoring its immense potential in addressing complex urban parking problems in practical applications.
While our research has achieved certain results, future work still needs to focus on improving and further exploring the following issues:
  • Selection of Coarsening Ratio: The coarsening dimension utilized in this study was determined through experimentation with a grid search strategy, yielding a general range that may limit its applicability to the specific dataset used and not be accurate enough for others. Therefore, in our subsequent efforts, we aim to employ deep learning methods to automatically learn and ascertain the optimal coarsening dimension.
  • Integration of Multi-source Data: Given that parking demand is influenced by a wide array of factors, including nearby traffic flow, weather conditions, and more, we intend to incorporate multiple data sources for feature fusion in the future. This approach will enhance the model’s comprehension of parking scenarios, providing a richer understanding of the complex dynamics at play.
  • Global Information Aggregation: The PRGAT algorithm proposed in this paper aggregates node feature information considering only the 2nd-order neighborhood, thus overlooking the influence of distant global nodes. In our future research, we plan to incorporate higher-order global information by employing concepts from fractal theory.

Author Contributions

Conceptualization, Y.W., Z.C. and L.P.; methodology, Y.W., Z.C. and L.P.; software, Y.W.; validation, Y.W. and Z.C.; formal analysis, Y.W.; investigation, Y.W. K.Z. and Y.C; resources, Z.C; data curation, K.Z., Z.C. and Y.C.; writing—original draft preparation, Y.W.; writing—review and editing, K.Z., Y.C. and L.P.; visualization, Y.W.; supervision, L.P.; project administration, L.P.; funding acquisition, Z.C. and L.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Key Areas R&D Program of Guangdong Province (Grant No. 2024B0101020004) and Major Program of Science and Technology of Shenzhen (Grant No. KJZD20231023100304010, KJZD20231023100509018).

Data Availability Statement

The dataset used in this study is provided by the Shenzhen Municipal Government and Shenzhen Smart City Group. It includes data from 9,000 parking lots across five districts in Shenzhen, covering geographic information, entry and exit timestamps, fee details, and internal facility information. Due to privacy and ethical considerations, this dataset is not publicly available.

Acknowledgments

The authors would like to thank the editor and the anonymous reviewers for their valuable comments.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Xin, H. China registers 415 million motor vehicles, 500 million drivers. Website, 2024. https://english.www.gov.cn/archive/statistics/202212/08/content_WS6391cafcc6d0a757729e41bc.html.
  2. Zou, W.; Sun, Y.; Zhou, Y.; Lu, Q.; Nie, Y.; Sun, T.; Peng, L. Limited sensing and deep data mining: A new exploration of developing city-wide parking guidance systems. IEEE Intelligent Transportation Systems Magazine 2020, 14, 198–215. [Google Scholar] [CrossRef]
  3. Ke, R.; Zhuang, Y.; Pu, Z.; Wang, Y. A smart, efficient, and reliable parking surveillance system with edge artificial intelligence on IoT devices. IEEE Transactions on Intelligent Transportation Systems 2020, 22, 4962–4974. [Google Scholar] [CrossRef]
  4. Research.; ltd, M. China Smart Parking Industry Report, 2022; Research In China, 2022.
  5. Rizvi, S.R.; Zehra, S.; Olariu, S. Aspire: An agent-oriented smart parking recommendation system for smart cities. IEEE Intelligent Transportation Systems Magazine 2018, 11, 48–61. [Google Scholar] [CrossRef]
  6. Kotb, A.O.; Shen, Y.c.; Huang, Y. Smart parking guidance, monitoring and reservations: a review. IEEE Intelligent Transportation Systems Magazine 2017, 9, 6–16. [Google Scholar] [CrossRef]
  7. Yang, K.; Tang, X.; Qin, Y.; Huang, Y.; Wang, H.; Pu, H. Comparative study of trajectory tracking control for automated vehicles via model predictive control and robust H-infinity state feedback control. Chinese Journal of Mechanical Engineering 2021, 34, 1–14. [Google Scholar] [CrossRef]
  8. Tang, X.; Yang, Y.; Liu, T.; Lin, X.; Yang, K.; Li, S. Path planning and tracking control for parking via soft actor-critic under non-ideal scenarios. IEEE/CAA Journal of Automatica Sinica, 2023. [Google Scholar]
  9. Fan, J.; Hu, Q.; Xu, Y.; Tang, Z. Predicting vacant parking space availability: a long short-term memory approach. IEEE Intelligent Transportation Systems Magazine 2020, 14, 129–143. [Google Scholar] [CrossRef]
  10. Lin, T.; Rivano, H.; Le Mouël, F. A survey of smart parking solutions. IEEE Transactions on Intelligent Transportation Systems 2017, 18, 3229–3253. [Google Scholar] [CrossRef]
  11. Anand, D.; Singh, A.; Alsubhi, K.; Goyal, N.; Abdrabou, A.; Vidyarthi, A.; Rodrigues, J.J. A smart cloud and IoVT-based kernel adaptive filtering framework for parking prediction. IEEE Transactions on Intelligent Transportation Systems 2022, 24, 2737–2745. [Google Scholar] [CrossRef]
  12. Ku, Y.; Guo, C.; Zhang, K.; Cui, Y.; Shu, H.; Yang, Y.; Peng, L. Toward Directed Spatiotemporal Graph: A New Idea for Heterogeneous Traffic Prediction. IEEE Intelligent Transportation Systems Magazine 2024, 16, 70–87. [Google Scholar] [CrossRef]
  13. Li, J.; Qu, H.; You, L. An integrated approach for the near real-time parking occupancy prediction. IEEE Transactions on Intelligent Transportation Systems 2022, 24, 3769–3778. [Google Scholar] [CrossRef]
  14. Shuai, C.; Zhang, X.; Wang, Y.; He, M.; Yang, F.; Xu, G. Online car-hailing origin-destination forecast based on a temporal graph convolutional network. IEEE Intelligent Transportation Systems Magazine 2023. [Google Scholar] [CrossRef]
  15. Li, D.; Wang, J. A Fusion Deep Learning Model via Sequence-to-Sequence Structure for Multiple-Road-Segment Spot Speed Prediction. IEEE Intelligent Transportation Systems Magazine 2022, 15, 230–243. [Google Scholar] [CrossRef]
  16. Zhang, J.; Chen, Z.; Xu, Z.; Du, M.; Yang, W.; Guo, L. A distributed collaborative urban traffic big data system based on cloud computing. IEEE Intelligent Transportation Systems Magazine 2018, 11, 37–47. [Google Scholar] [CrossRef]
  17. Zhou, X.; Zhang, Y.; Li, Z.; Wang, X.; Zhao, J.; Zhang, Z. Large-scale cellular traffic prediction based on graph convolutional networks with transfer learning. Neural Computing and Applications 2022, 1–11. [Google Scholar] [CrossRef]
  18. Fahrbach, M.; Goranci, G.; Peng, R.; Sachdeva, S.; Wang, C. Faster graph embeddings via coarsening. In Proceedings of the international conference on machine learning. PMLR; 2020; pp. 2953–2963. [Google Scholar]
  19. Cai, C.; Wang, D.; Wang, Y. Graph coarsening with neural networks. arXiv 2021, arXiv:2102.01350. [Google Scholar]
  20. Loukas, A. Graph reduction with spectral and cut guarantees. Journal of Machine Learning Research 2019, 20, 1–42. [Google Scholar]
  21. Kumar, M.; Sharma, A.; Kumar, S. A unified framework for optimization-based graph coarsening. Journal of Machine Learning Research 2023, 24, 1–50. [Google Scholar]
  22. Jin, Y.; Loukas, A.; JaJa, J. Graph coarsening with preserved spectral properties. In Proceedings of the International Conference on Artificial Intelligence and Statistics. PMLR, 2020, pp. 4452–4462.
  23. Chen, Y.; Shu, T.; Zhou, X.; Zheng, X.; Kawai, A.; Fueda, K.; Yan, Z.; Liang, W.; Kevin, I.; Wang, K. Graph attention network with spatial-temporal clustering for traffic flow forecasting in intelligent transportation system. IEEE Transactions on Intelligent Transportation Systems 2022. [Google Scholar] [CrossRef]
  24. Lu, Z.; Lv, W.; Xie, Z.; Du, B.; Xiong, G.; Sun, L.; Wang, H. Graph sequence neural network with an attention mechanism for traffic speed prediction. ACM Transactions on Intelligent Systems and Technology (TIST) 2022, 13, 1–24. [Google Scholar] [CrossRef]
  25. Lu, Q.; Tang, Z.; Nie, Y.; Peng, L. ParkingRank-D: A Spatial-temporal Ranking Model of Urban Parking Lots in City-wide Parking Guidance System. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (iTSC). IEEE; 2019; pp. 388–393. [Google Scholar]
  26. Dong, S.; Chen, M.; Peng, L.; Li, H. Parking rank: A novel method of parking lots sorting and recommendation based on public information. In Proceedings of the 2018 IEEE International Conference on Industrial Technology (ICIT). IEEE; 2018; pp. 1381–1386. [Google Scholar]
  27. Huang, C.J.; Ma, H.; Yin, Q.; Tang, J.F.; Dong, D.; Chen, C.; Xiang, G.Y.; Li, C.F.; Guo, G.C. Realization of a quantum autoencoder for lossless compression of quantum data. Physical Review A 2020, 102, 032412. [Google Scholar] [CrossRef]
  28. Liu, M.; Zhu, T.; Ye, J.; Meng, Q.; Sun, L.; Du, B. Spatio-temporal autoencoder for traffic flow prediction. IEEE Transactions on Intelligent Transportation Systems 2023. [Google Scholar] [CrossRef]
  29. Fainstein, F.; Catoni, J.; Elemans, C.P.; Mindlin, G.B. The reconstruction of flows from spatiotemporal data by autoencoders. Chaos, Solitons & Fractals 2023, 176, 114115. [Google Scholar]
  30. Guo, G.; Yuan, W.; Liu, J.; Lv, Y.; Liu, W. Traffic forecasting via dilated temporal convolution with peak-sensitive loss. IEEE Intelligent Transportation Systems Magazine 2021, 15, 48–57. [Google Scholar] [CrossRef]
  31. Du, W.; Li, B.; Chen, J.; Lv, Y.; Li, Y. A spatiotemporal hybrid model for airspace complexity prediction. IEEE Intelligent Transportation Systems Magazine 2022, 15, 217–224. [Google Scholar] [CrossRef]
  32. Yao, Z.; Xia, S.; Li, Y.; Wu, G.; Zuo, L. Transfer learning with spatial–temporal graph convolutional network for traffic prediction. IEEE Transactions on Intelligent Transportation Systems 2023. [Google Scholar] [CrossRef]
  33. Zhao, L.; Song, Y.; Zhang, C.; Liu, Y.; Wang, P.; Lin, T.; Deng, M.; Li, H. T-gcn: A temporal graph convolutional network for traffic prediction. IEEE transactions on intelligent transportation systems 2019, 21, 3848–3858. [Google Scholar] [CrossRef]
  34. Sun, Y.; Jiang, X.; Hu, Y.; Duan, F.; Guo, K.; Wang, B.; Gao, J.; Yin, B. Dual dynamic spatial-temporal graph convolution network for traffic prediction. IEEE Transactions on Intelligent Transportation Systems 2022, 23, 23680–23693. [Google Scholar] [CrossRef]
  35. Zheng, Y.; Li, W.; Zheng, W.; Dong, C.; Wang, S.; Chen, Q. Lane-level heterogeneous traffic flow prediction: A spatiotemporal attention-based encoder–decoder model. IEEE Intelligent Transportation Systems Magazine 2022. [Google Scholar] [CrossRef]
  36. Li, D.; Lasenby, J. Spatiotemporal attention-based graph convolution network for segment-level traffic prediction. IEEE Transactions on Intelligent Transportation Systems 2021, 23, 8337–8345. [Google Scholar] [CrossRef]
  37. Wickman, R.; Zhang, X.; Li, W. A Generic Graph Sparsification Framework using Deep Reinforcement Learning. arXiv preprint arXiv:2112.01565, arXiv:2112.01565 2021.
  38. Sadhanala, V.; Wang, Y.X.; Tibshirani, R. Graph sparsification approaches for laplacian smoothing. In Proceedings of the Artificial Intelligence and Statistics. PMLR; 2016; pp. 1250–1259. [Google Scholar]
  39. Wu, H.Y.; Chen, Y.L. Graph sparsification with generative adversarial network. In Proceedings of the 2020 IEEE International Conference on Data Mining (ICDM). IEEE; 2020; pp. 1328–1333. [Google Scholar]
  40. Chen, J.; Saad, Y.; Zhang, Z. Graph coarsening: from scientific computing to machine learning. SeMA Journal 2022, 79, 187–223. [Google Scholar] [CrossRef]
  41. Dhillon, I.S.; Guan, Y.; Kulis, B. Weighted graph cuts without eigenvectors a multilevel approach. IEEE transactions on pattern analysis and machine intelligence 2007, 29, 1944–1957. [Google Scholar] [CrossRef] [PubMed]
  42. Rhouma, D.; Ben Romdhane, L. An efficient multilevel scheme for coarsening large scale social networks. Applied Intelligence 2018, 48, 3557–3576. [Google Scholar] [CrossRef]
  43. Wang, Q.; Jiang, H.; Qiu, M.; Liu, Y.; Ye, D. Tgae: Temporal graph autoencoder for travel forecasting. IEEE Transactions on Intelligent Transportation Systems 2022. [Google Scholar] [CrossRef]
  44. Mo, R.; Pei, Y.; Venkatarayalu, N.V.; Joseph, P.N.; Premkumar, A.B.; Sun, S.; Foo, S.K.K. Unsupervised TCN-AE-based outlier detection for time series with seasonality and trend for cellular networks. IEEE Transactions on Wireless Communications 2022, 22, 3114–3127. [Google Scholar] [CrossRef]
  45. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  46. Yu, B.; Yin, H.; Zhu, Z. Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting. arXiv 2017, arXiv:1709.04875. [Google Scholar]
  47. Song, C.; Lin, Y.; Guo, S.; Wan, H. Spatial-temporal synchronous graph convolutional networks: A new framework for spatial-temporal network data forecasting. In Proceedings of the Proceedings of the AAAI conference on artificial intelligence, 2020, Vol.; 2020; Volume 34, pp. 914–921. [Google Scholar]
  48. Wang, Y.; Ku, Y.; Liu, Q.; Yang, Y.; Peng, L. Large-Scale Parking Data Prediction: From A Graph Coarsening Perspective. In Proceedings of the 2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC). IEEE; 2023; pp. 1410–1415. [Google Scholar]
Figure 1. The structure of PRGAT.
Figure 1. The structure of PRGAT.
Preprints 117338 g001
Figure 2. The structure of Urban Parking Prediction Framework.
Figure 2. The structure of Urban Parking Prediction Framework.
Preprints 117338 g002
Figure 3. Distribution map of parking lots in Bao’an District.
Figure 3. Distribution map of parking lots in Bao’an District.
Preprints 117338 g003
Figure 4. The parking graph of the core area of Bao’an, encompassing over 600 parking facilities.
Figure 4. The parking graph of the core area of Bao’an, encompassing over 600 parking facilities.
Preprints 117338 g004
Figure 5. Distribution of parking lots in the core area of Bao’an under different coarsening granularities: (a) Coarsening ratio = 0.9 (b) Coarsening ratio = 0.7 (c) Coarsening ratio = 0.4 (d) Coarsening ratio = 0.1.
Figure 5. Distribution of parking lots in the core area of Bao’an under different coarsening granularities: (a) Coarsening ratio = 0.9 (b) Coarsening ratio = 0.7 (c) Coarsening ratio = 0.4 (d) Coarsening ratio = 0.1.
Preprints 117338 g005
Table 1. Description of Shenzhen Parking Lot Dataset
Table 1. Description of Shenzhen Parking Lot Dataset
District #Nodes Start Time Granularity Time Steps
Bao’an 1660 2016/6/1 15min 17,280
Luohu 1730 2016/6/1 15min 17,280
Futian 1806 2016/6/1 15min 17,280
Longgang 1473 2016/6/1 15min 17,280
Longhua 2531 2016/6/1 15min 17,280
Table 2. Description of Model Configurations and Training Parameters
Table 2. Description of Model Configurations and Training Parameters
Method Configuration Batch Size Learning Rate Optimizer Loss Function Weight Decay Patience (/Epoch)
PRGAT Number of Attention Heads Feature Dimension per Head 64 1e-4 Adam MSE 1e-4 100
8 128
SGC Threshhold λ - - - - - -
1e-8
SparRL Maximum number of neighbors to pay attention to 64 1e-4 Adam Huber Loss 1e-4 500
64
TGCN GRU Hidden Units 64 1e-5 Adam Huber Loss 1e-4 200
100
STGCN Graph Convolution Dimension Temporal Convolution Dimension 64 1e-5 Adam Huber Loss 1e-4 200
16 64
STSGCN GCNs per Module Spatio-temporal GCNs Layers(STSGCL) 64 1e-5 Adam Huber Loss 1e-4 200
3 4
TCN-AE TCN Dilation Rates Filter Count and Kernel Size 64 1e-4 Adam MSE 1e-4 100
(1,2,4,8,16) 20
Table 3. Comparison of Predicted Parking Lot Occupancy Rates in Bao’an District for the Next Hour
Table 3. Comparison of Predicted Parking Lot Occupancy Rates in Bao’an District for the Next Hour
Method MAE RMSE MAPE(%) Epoch
15min 30min 60min 15min 30min 60min 15min 30min 60min
Default+TGCN 7.933 8.7224 11.3431 11.2128 12.3413 15.1087 13.82 16.23 20.59 834
GAT+TGCN 7.7224 8.3431 10.2926 10.0689 11.4238 12.3675 12.74 14.41 19.82 839
PRGAT+TGCN 4.2631 4.719 5.8398 7.4103 8.7885 10.7526 9.89 10.38 13.09 844
Default+Coarsening+TGCN+AE 8.0153 8.9607 11.8285 11.4731 13.455 16.0598 14.31 17.25 20.77 655
Default+Sparsification+TGCN+AE 9.9726 11.1673 13.6201 13.4071 14.757 18.6382 14.22 17.54 22.96 827
PRGAT+Sparsification+TGCN+TCN-AE 7.851 9.0267 11.5006 12.4467 12.6782 15.483 13.36 15.28 18.64 846
PRGAT+Coarsening+TGCN+TCN-AE 4.6552 5.209 7.3209 7.8395 9.4337 10.9448 10.24 10.53 13.49 558
Default+STGCN 4.728 5.5534 9.598 10.7798 11.7946 14.9729 10.68 16.47 19.12 673
GAT+STGCN 3.7947 4.3721 7.6845 7.44 10.9601 12.4731 9.42 14.2 16.93 645
PRGAT+STGCN 2.159 2.7601 4.328 4.893 6.4853 10.4475 7.63 10.65 11.74 639
Default+Coarsening+STGCN+AE 5.4295 7.6965 10.2398 11.4307 12.9623 16.0431 11.74 17.83 20.01 568
Default+Sparsification+STGCN+AE 6.5246 7.9378 11.0445 12.6193 14.7207 17.8087 13.47 18.55 20.79 651
PRGAT+Sparsification+STGCN+TCN-AE 5.0214 6.5779 9.538 10.6814 12.5294 15.7111 11.43 14.84 17.17 662
PRGAT+Coarsening+STGCN+TCN-AE 2.1977 3.3915 5.3604 4.9548 7.8741 10.9849 8.65 11.6 12.31 476
Default+STSGCN 6.5194 6.5358 9.5476 10.8984 14.4152 17.1354 12.19 14.37 16.33 788
GAT+STSGCN 5.6569 6.1912 8.1355 9.8521 13.534 16.7885 10.71 11.75 12.49 752
PRGAT+STSGCN 4.1645 4.6743 7.6569 7.1038 9.4388 11.8712 7.85 8.22 8.89 769
Default+Coarsening+STSGCN+AE 7.3511 7.8178 11.0491 11.4534 14.6645 18.5779 12.67 15.03 16.82 661
Default+Sparsification+STSGCN+AE 8.1708 10.319 12.4703 13.3065 15.2423 19.1752 14.11 17.33 19.8 776
PRGAT+Sparsification+STSGCN+TCN-AE 6.6163 7.656 9.7815 11.8712 13.9123 18.7807 12.82 13.94 17.25 740
PRGAT+Coarsening+STSGCN+TCN-AE 4.6079 4.9813 8.093 7.9218 10.0797 12.9521 8.72 9.31 10.09 556
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated