With the rise of the Internet of Things (IOT), mobile devices and Location-Based Social Network (LBSN), an abundant amount of trajectory data has made research on location prediction more popular. The check-in data shared through LBSN hides information related to life patterns, obtaining this information is helpful for location prediction. However, the trajectory data recorded by mobile devices is different from check-in data that has semantic information. In order to obtain the user’s semantic, relevant studies match the stay point to the nearest Point of Interest (POI), but location error may lead to wrong semantic matching. Therefore, we propose A Self-Attention for Next Location Prediction based on Semantic Mining to predict the next location. When calculating the semantic feature of a stay point, the first step is to search for the k nearest POI, then use the reciprocal of the distance from the stay point to the k nearest POI and the number of categories as weights. Finally, we use the probability to express semantic without losing other important semantic information. Furthermore, this research combined with sequential pattern mining can bring richer semantic features. In order to better perceive the trajectory, temporal features learn periodicity of time series by sine function. In terms of location features, we build a directed weighted graph and regard the frequency of users visiting locations as the weight, so the location features are rich in contextual information. We then adopt the Self-Attention model to capture long-term dependencies in long trajectory sequences. Experiments in Geolife show that the semantic matching of this study improved by 45.78% in TOP@1 compared with the closest distance search for POI. Compared with the baseline, the model proposed in this study improved by 2.5% in TOP@1.
Keywords:
Subject: Computer Science and Mathematics - Artificial Intelligence and Machine Learning
1. Introduction
In recent years, with the rise of positioning devices, Internet-Of-Things (IOTs) and smart city concepts, mobile phone positioning data and social network data have provided a large number of continuous location trajectory data; research on trajectory prediction and analysis has become more and more popular, and related research topics have gradually received attention. By predicting the location people will visit in the future, advertising companies can immediately provide location-related ads [19], government departments can predict the flow of people for traffic planning in order to ease traffic congestion [6]. Platforms such as uber are also using next location prediction technology to better estimate customer’s travel needs and allocate resources accordingly [16]. In the past, the most commonly used method for location prediction was either Markov chain [5,8,18] or machine learning [17], but due to large amount of trajectory data, many related studies have begun using deep learning technology to predict the user's next location [1,6,13,14,15,19,23,25].
Many relevant studies often consider spatiotemporal features such as location features and temporal features when using deep learning to predict the user’s next location. However, the check-in data shared through LBSN hides semantic information related to life patterns, if the model considers semantic features, it can obtain more detailed characteristics about the user's life pattern and location preference. Therefore, semantic features can help enhance the performance of next location prediction. However, not all users like to share their check-in information on LBSN. The trajectory data recorded by mobile devices is not like the check-in data of social networks that have complete semantic annotation information. In order to understand the user’s semantic trajectory behavior, relevant studies [19,24] matched the trajectory stay point to the nearest Point of Interest (POI). However, location error may lead to incorrect semantic matching; incorrect semantic matching causes the model to learn wrong semantic behavior patterns.
Due to the motivation mentioned above, we propose a Self-Attention for Next Location Prediction based on Semantic Mining. In our model, we have location features, temporal features, semantic features and user features, while focusing on how to extract semantic features. For semantic features, the stay points of trajectory are matched to the k nearest POI, we then use the reciprocal of the distance from the stay points to the k nearest POI and the number of categories as weights. Finally, we use probability to express semantic features without losing other important semantic information. This method solves the wrong semantic matching problem caused by location error and also considers more spatial information of POIs. Furthermore, we combine sequential pattern mining to further infer user’s semantic behaviors, which can bring richer semantic features. For temporal features, we adopt Time2vec to learn periodic features from the time features. For location features, we adopt the Node2vec to extract contextual features of the trajectory sequence. Finally, we concatenate the features as input of the prediction mode and we adopt Self-Attention to capture long-term dependencies in sequences; Self-Attention uses fully connection to consider the information of each stage together, so it retains more long-term information. In the output of Self-Attention, we consider the user id feature to make the model more personal and obtain prediction results. As a result, we can predict the user's future location by inputting the user's past locations, timestamps, and semantics into Self-Attention.
The contributions of this paper are listed below:
We design semantic matching to effectively extract the semantic feature of each stay point. We then combine sematic matching with sequential pattern mining to bring richer semantic features;
We use Node2vec and Time2vec which consider the interaction of each location and the periodicity of the time series;
We adopt Self-Attention to predict the user's next trajectory for solving the problem with RNN (Recurrent Neural Network); the problem with RNN is that long-term memory is diminished with each transfer;
We use Geolife dataset for our experiments. Based on experiment results, our method is better than a number of state-of-the-art methods.
The following is the structure of this paper. First, we review relevant research on location prediction in Section 2. Next, Section 3 is the problem statement for explaining the numerous terms in location prediction. Then, Section 4 explains the method we proposed for location prediction. Following that, we evaluate our model's performance and compare its accuracy to that of other models in Section 5. In Section 6, we conclude our findings and briefly discuss future works.
2. Related Work
In this section, we discussed important literature regarding location prediction. In terms of location prediction, besides the design of the model itself, there are many different methods for feature extraction, such as temporal, location, and semantics. Therefore, we can separate the literature into two topics: location prediction and feature extraction.
2.1. Location Prediction
In the first part, we introduce the literature on location prediction. In 2018, Xia et al. considered both stay points and semantic information and used a decision tree for location prediction [17]. A Markov chain is often used to deal with location predictions. A Markov chain is a method with discrete random variables that capture the regularity of human movement. In 2018, Jiang et al. extended the first-order Markov chain to the k-order Markov chain to consider more important historical information [8]. In 2018, Xia et al. established a variable-order Markov chain to predict location based on the matching of the historical trajectory and the current trajectory [18]. In 2017, Fernandes et al. combined Naïve Bayesian Classification algorithm and the Markov model to predict next location [5]. However, the Markov chain transfer process is memoryless, which means that the result of the current state is only influenced by the previous state.
Recently, many studies have used deep learning to predict the user’s future location. RNN is a typical deep learning model for time series. Since it can transfer long-term memory to the previous state, it solves the Markov chain problem (the Markov chain is limited to the previous state). The SERM model proposed by Yao et al., it learns embeddings for multiple features (user, location, time, semantic), and uses Long Short-Term Memory (LSTM) to forecast the next location. [23]. In 2020, Su et al. considered both propagation directions of LSTM; they used Bi-directional Long Short-Term Memory (BiLSTM) to predict the user’s next location [13]. Furthermore, in 2020, Xu et al. combined BiLSTM with Similarity-based Markov Model (SMM) to predict the user’s next location [19]. The problem with RNN is that long-term memory is diminished with each transfer. The Attention mechanism [14] improves this because the attention mechanism does not pass the long-term memory through each stage. In other words, the attention mechanism uses fully connected layer to consider the information of each stage together, so it retains more long-term memory. In 2018, Al-Molegi et al. adopted the Attention mechanism in their neural network to capture long-term dependencies in trajectory sequences to predict the user’s future location [1]. In 2020, Zhang et al. designed a variant of LSTM with an Attention mechanism to better capture spatiotemporal dependencies [25]. Wang et al. designed a variant of the Attention mechanism; this variant focuses on the time interval and distance interval, thus enriching the continuity of spatiotemporal, which can better distinguish each location [15]. However, in 2020, Feng et al. proposed a unique attention mechanism application. This method uses the user's current trajectory with LSTM to predict the user’s future location, but also combines the historical trajectory with Attention to capture multi-level periodicity. In addition to considering location feature, time feature, semantics feature and user ID feature, they also consider user text feature [6]. In 2021, Wen et al. adopted LSTM to capture long-term and short-term spatiotemporal dependencies, and Attention mechanism is introduced to distinguish each location in different contexts [16]. In addition to the structure of the neural network, the feature extraction is also a key point that affects the next location prediction.
2.2. Feature Extraction
In the second section, we introduce feature extraction methods. Trajectories have three main features: location, time, and semantic. The location feature is usually a one hot vector whose dimension is determined by the number of locations. Due to the large number of locations, the curse of dimensionality occurs when training the model. In order to get around this problem, relevant studies use embedding layer (fully connection), which reduces a high-dimensional one-hot vector to a low-dimensional space, which eliminates the curse of dimensionality in the prediction model [8,23,25]. On the other hand, since trajectories have contextual features like words, some studies use Word2vec to extract location features [15,19]. The principle of Word2vec is to use the weights of the fully connected layer of a classifier as word vectors. In this classifier, the word’s one-hot vectors are input, then fed into a fully connected layer, and then fed to a softmax to get the probability of the contexts. Finally, the weight of the fully connected layer after training will be obtained, which is the word vectors of Word2vec. The difference between the embedding layer and Word2vec is explained below. Word2vec is an unsupervised learning method that uses context to learn word embedding; Word2vec is a pretrained model. In the embedding layer, the weight of fully connection is updated based on the learning of label information. In order to achieve a higher supervised learning effect, the embedding layer is used as a layer of the network to learn and adjust according to the target. But Word2vec and fully connection omit a lot of valid contexts which makes it difficult to contain more spatial information and distinguish between each location. Therefore, in the loc2vec embedding method proposed by Sassi et al., each location is encoded as a vector whereby the more often two locations co-occur in the trajectory sequences, the closer their vectors will be [12]. Furthermore, the graph embedding method of node2vec is based on Word2vec; its directed weighted graph carries more spatiotemporal context information [7]. In 2021, Wen et al. used Node2Vec to extract location feature, which encodes each location in all trajectories to construct a directed weighted graph. It then takes the number of visits to the location as the weight of the graph, which reflects the visit order preference and frequency [16]. Also inspired by Word2vec, in 2019 Xu et al. proposed an embedding model called Venue2Vec, which combines spatiotemporal context, semantic information, and sequence relationships; semantics of the same type and locations that are close or that users frequently visit will have similar vector spaces [20]. In addition to Word2vec-based methods, in 2020 Chen et al. proposed the Convolutional Embedding Model (CEM), which embeds the locations by using convolutions [3].
In terms of temporal features, some studies usually perform a series of preprocessing in timestamp to extract temporal features for the next location prediction, such as converting timestamps into hours, days, weeks, months, and seasons. However, temporal features are cyclical. Taking hours as an example, temporal features are usually represented by 1 to 24, although using hours to represent the temporal features can reflect the periodicity, but it may not be able to fully express the period of time. Therefore, Lu et al. considered the cyclic features of time, and designed the coordinates on a unit circle using the Sine and cosine function to represent the cyclic features of time [10]. Kazemi et al. proposed Time2vec. This method uses temporal features to learn the phase-shift and frequency of the sine function; it can learn periodic temporal features, which allows the neural network to better learn temporal features [9].
In terms of semantic feature extraction, trajectory data is divided into check-in data and GPS trajectory data. Check-in data such as Gowalla dataset and Foursquare dataset, usually has semantic information. On the other hand, GPS trajectory data such as Geolife dataset does not record semantic information. First, we introduce the semantic feature extraction of check-in data. In 2016, Xie et al. use graph structure to embed semantic features [21]. To encode user’ semantic trajectory in a vector space. In 2019, Cao et al. proposed habit2vec, which keeps the original user's living habits pattern [4]. In 2020, Zhang et al. proposed Sen2vec which calculates semantic vectors according to the frequency and principal component analysis [25]. As for GPS trajectory data, in order to extract semantic information, it is necessary to collect POI data and match the stay points. If the POI is not considered, the semantic information can be extracted from the trajectory sequence data alone. In 2009, Ye et al. used the sequential pattern mining to infer users' semantic behavior [22]. In 2016, Chen et al. proposed semantic trajectory patterns, which are moving patterns with location, time and semantic attributes. Given a user’s trajectory, their objective is to mine common semantic trajectory patterns [2].
3. Problem Statement
We explicitly define the notations used in the location prediction issue in this section. Our topic is to predict the future location from user’s trajectory. The notations concerning location prediction are listed in Table 1.
Definition1.Trajectory:
A time-ordered sequence of GPS points, a GPS trajectory , each GPS point is a 4-tuple , where the user ID, latitude, longitude, and time are denoted by , , and respectively.
Definition2.Staypoint:
Given a set of sequential GPS points satisfying, ,, , and . , a stay point is a 5-tuple , where and denote the average latitude and longitude of GPS point respectively, and are defined as (1), where and refers to the time of entering and leaving a stay point.
Definition3.POI:
POI (Point of Interest) refers to a landmark on the electronic map, which can be a restaurant, a bank, a school, a hospital, etc.,, which is composed of the type of poi , the latitude and the longitude.
Definition4.Semantic:
The semantic is defined as the purpose of the user visiting a stay point . The semantic is calculated by matching the stay point to the k nearest POI. The dimension of is the number of , each dimension corresponds to a different . Given a set of consecutive stay points , each 𝑠 will search for k nearest POI as semantic .
Definition5.Staygrid:
We use grids to represent the stay points , . where is the index of the gird, and we use time of entering stay point as the time of stay grid .
Definition6.TrajectorySequence:
According to time window , a user's stay grid sequence is divided into several subsequences , where 𝑚 is the index of the subsequence. Each contains several stay grids of in the time window , if , then belongs to (the time window can be 1 day, 1 week , 1 month or 1 year). We then convert into a trajectory sequence using sliding window. Given a subsequence , we use sliding window to get , , where is the length of the sliding window whereas is defined as a trajectory sequence.
ProblemDefinition:
For a user , given a trajectory sequence , our purpose is to predict the next location .
4. Proposed Method
In this section, we introduce our proposed method, which is split into three parts. Frist, we introduce the entire model architecture and explains the input and output of the model. then, we will introduce how we extract input features. Finally, we explain our proposed model in detail.
4.1. System Framework
Our framework is shown in Figure 1. The architecture is divided into three parts: input features (green), prediction model (blue) and predicted results (red). In the beginning, the required data (trajectory data and POI data) are processed to obtain four features: 1. User feature 2. Location feature 3. Temporal feature 4. Semantic feature. Then, we divided the data into training and testing sets. Next, the neural network model is fed the training data (the structure of our model will be discussed in the end of this section). The trained model will generate the anticipated future locations when we feed it the testing data after the training operation.
4.2. Input Features
We have four major features: user feature, location feature, temporal feature, and semantic feature. The first three features are obtained from trajectory data whereas the last feature (semantic feature) is obtained by matching the stay point to nearest POI. Therefore, we need two types of data: trajectory data and POI data.
Trajectory data: The main data for predicting the user's next location. The location is recorded every 1 to 5 seconds. Trajectory data records the living habits of users and provides clues for predicting the user's next location;
POI data: It is the POI data in the study area. Each poi ∈ POI contains poi.na, poi.ty, poi.lat, and poi.lng.
We will describe how to extract features from data and how to transform data into features in the sections that follow. And what methods are used to get important features. These features will be input into the model.
User Feature: User feature is the user ID . To personalize the prediction model, we consider the user feature ( is a one hot vector whose dimension is the number of users );
Location Feature: We use stay grid to represent a location. stay grid is one hot vector, and the dimension of is determined by the number of grids that cover the study area which users have visited;
Temporal Feature: Temporal feature reveals what time the information is in. Several methods usually perform a series of preprocessing on timestamps to extract temporal features for the next location prediction, such as converting timestamps into hour, day, week, month, and season. In this paper, we convert timestamps into hours (1 to 24) to obtain temporal features t;
Semantic Feature: Semantic feature is defined as the purpose of a user visiting a stay point. Some studies usually match a stay point to the nearest POI. In this paper, we introduce a semantic matching to extract semantic features; the architecture of semantic matching as shown in Figure 2.
We When we calculate the semantic vector of the stay point, we consider the nearest POI. In addition, inspired by Ye et al. [22], we combine sequential pattern mining to mine the user’s home and workplace. We then calculate the semantic vector considering the home and workplace.
We use the stay points to mine the user’s home and workplace. However, this sequence still cannot be directly applied to mining home and workplace, because no two stay points have the same latitudes and longitudes. For example, stay points at workplace have different coordinates even though they are very close to each other [22]. To solve this problem, we apply the OPTIC clustering algorithm to find out where the stay points are clustered as demonstrated in Figure 3. The stay points of a user are collected into a dataset and clustered into several geographic areas . There are two parameters: number of points and distance threshold; we set number of points at 2 and distance threshold at 45 meters. Accordingly, additional stay points will be added to the cluster if there are at least two within 45 meters of a clustered stay point. Then the clustering stay points will be the input to home and workplace mining.
Next, we adopt the Closet+ algorithm to mine the home and workplace in the clustering stay points. Figure 4 illustrates an example of home and workplace mining. The first table in Figure 4 means that on Day 1, the user went to , on Day 2 the user went to and so on. We then count the number of clusters in the database to obtain the second table. Next, we delete the clusters below the min support = 1 in the second table to produce the fourth table. Using the fourth table, we can find the most frequent clusters during the day and night, which we can use to mine the workplace and home clusters. In this paper, we only calculated the first stage of the Closet+ algorithm.
Next, we will introduce the semantic matching algorithm. The flow chart of the semantic matching algorithm is shown in Figure 5 where is Real estate, is Semantic vector (the dimension of Semantic vector is the number of different , in addition to home and workplace), is temporary register (the same dimension as ). The following explains how to calculate the semantic vector of the stay point :
First, the algorithm searches for the nearest POI, then starts to calculate the of each ;
If c = home and = R, then is changed to home. Next, we calculate the reciprocal of the distance between and ;
If c = work andR, then is changed to workplace (because we think that POI that are not R is likely to be workplace). Next, we calculate the reciprocal of the distance between and ;
If home and c workplace, we directly calculate the reciprocal of the distance between and ;
We do repeat , when the loop stops. We then normalize , ending the algorithm.
We introduce an example of semantic matching in Figure 6. Assuming that there are only 3 POI types, so the dimension of is 5 ([School, Shop, R, Home, Work]), and the k = 4 (The algorithm searches for the nearest 4 ). The blue dotted lines in Figure 6 are the distance between the stay point and the 4 . Once the algorithm starts, the stay point will search for the 4 nearest . We need to calculate the reciprocal of the distance between the 4 and the stay point, since the clustering stay point is neither the home cluster nor the workplace cluster. In this example, the semantic feature is calculated as (2).
4.3. Prediction Model
Our Prediction model architecture is shown in Figure 7, it uses the features mentioned in input features to obtain the predicted next location. In order to better learn temporal features and location features, we use Time2vec to learn the periodic features of the temporal features and use Node2vec to extract contextual features rich in trajectory sequences. Next, the temporal features, location features, and semantic features are concatenated as input for the prediction model. In order to predict the next location, we adopt Self-Attention to solve this issue. Self-Attention is a model that has been successfully used for time series forecasting challenges and can model temporal properties well. In order to personalize our model, we add the user ID features to the output of Self-Attention. Finally, we can obtain the final prediction result after softmax.
4.3.1. Temporal Features Extractor
The temporal feature t is a one-dimensional vector represented by hours (1-24). However, time is a periodic feature. Although using hours to represent the temporal features can reflect the periodicity, but it may not be able to fully express the period of time. Therefore, inspired by Kazemi et al. [9], we adopt Time2vec to extract the periodic temporal features. The formula of Time2vec is shown as (3), where is the dimension of the time2vec output . The first dimension learns original temporal features, the other dimensions learns periodic temporal features using the sine function. Where and are learnable parameters, referring to the phase-shift and frequency of the sine function.
4.3.2. Location Features Extractor
The location feature is a one hot vector (the dimension is composed of the number of grids covering the study area). Due to the large number of grids, the dimension of is very huge, which may lead to the curse of dimensionality, not only increasing the amount of computation, but also making the model easier to overfit. Therefore, we adopt a graph-based embedding method called Node2vec, which encodes each grid in all trajectories to construct a directed weighted graph. It then takes the number of visits to the location as the weight of the graph, which reflects the visit order preference and frequency; while using Breadth-First Search (BFS), Depth-First Search (DFS), and random walks to generate training data. Finally, using Word2vec embeddings to obtain more expressive features that fully consider the interaction between each location and its neighbors, it also can capture richer context spatial information.
In Node2vec, for all trajectories of all users, we create a directed weighted . Whereis a set of vertices (all grids will be a vertices), and is a set of edges. We define edge as (4).
Based on the trajectory movement of all users, we can get . The denotes the weight of , which is calculated from the total number of visits. Then, Node2vec combines the BFS and DFS to sample the graph. Given the current node , the transition probability of visiting the next node is shown as (5).
Before calculating the transition probability, node and node must have an edge link; is the probability of an unnormalized transition between node and node , whereas refers to the normalization constant. In order to control the random walk strategy, Node2vec uses two hyperparameters (p and q), which is defined as (6).
The random walk strategy is demonstrated in Figure 8. Where is the weight of edge , and is the shortest number of steps between the previous node and the next node . When node transfers to node , we have to consider the number of steps between the previous node and next node .
The parameter determines the probability that will visit the previous node , and only functions when = 0. The probability of accessing the previous node is larger if has a large value. The random walk's direction is determined by the parameter ; causes the random walk to tend nodes near node (BFS); only functions when = 2. The random walk has a propensity to travel to nodes far from node (DFS) if . The loss of Node2vec is calculated as (7), where is the set of neighboring nodes of node obtained by sampling strategy and is a mapping function that maps node to an embedding vector. For the current node , the optimization goal is to give each node the condition to maximize the probability of the adjacent node .
Then according to this loss , the skip-gram in word2vec is directly used to learn the embedding vector. The principle of Word2vec is to use the weights of the fully connected layer of a classifier as word vectors. In this classifier, the word’s one-hot vectors are input, then fed into a fully connected layer, and then fed to a softmax to get the probability of the contexts. We followed previous work [16] and set . The obtained location embedding vector is defined as (8), where is the dimension of Node2vec output that can be determined, and is the location feature extracted by Node2vec. We use Node2vec to obtain more expressive location features which fully considers the relationship between each node and its neighbor nodes.
4.3.3. Model Structure
We fed the features into the model to obtain the next location. The output of Time2vec is , and the output of Node2vec . The semantic and user id are fed into the fully connected layer to obtain and as (9), where and are trainable weight matrices, and are the bias parameter of the fully connected layer. Our prediction model input is shown as (10).
is not directly trained by the prediction model because does not have a time series characteristic; how is considered by the model will be described later. In order to capture long-term dependencies in trajectory sequences, we adopt the Self-Attention model. Self-Attention use fully connected layer to consider the information from each step collectively, which allows it to maintain more long-term memory than RNN, which communicates information at each stage. Self-Attention also has fewer parameters and a lower time complexity when compared to RNN. The formula for Self-Attention is shown as (11), where are weight matrices that can be trained. is fed to 3 full connections to get . ( = = ) is the dimension of the hidden state of Self-Attention. We then compute the dot product of with (calculate the features similarity of each location in a trajectory sequence), divide by (the purpose of dividing by is to make the gradient of softmax not too small), apply a softmax function (the features similarity of each stage is expressed by probability, and the higher the features similarity is, the more important the features of the stage is), then dot product (get a weighted score for each location, this determines which stage features are important) to obtain .
Finally, we only take the last stage as the output of Self-Attention, so the dimension of is . There are two reasons we only take the last stage : Each stage uses the fully connection to consider all spatiotemporal contexts of the sequence; if the output is stages, the number of network parameters will increase, which may easily cause the problem of overfitting.
In order to personalize the prediction model, we consider the user ID in the output of Self-Attention, and then feed it to softmax to get the final output. The formula is shown as (12), where are the weight matrices, and are the bias parameters of the fully connection. We feed to the fully connection to get the output vector . The trajectory patterns of each user are different, and the model may mix them together during training. In order to distinguish the trajectory sequence of each user, is added to the feature of the user ID to obtain , so that the model can distinguish the trajectory pattern of each user. Finally, is fed to the fully connection, so that the dimension of the output is the same as the number of grids . After feeding to softmax, the final output of the model can be obtained.
5. Experimental Evaluation
This section discusses the experiments and examines the results using our model. This chapter is divided into 4 parts. Our experimental data and setup are covered in great depth in part1. Part 2 to part 3 are experiments to evaluate the accuracy of our model under numerous parameters while comparing the performance of our model with other models. Part 4 visualizes the next location prediction results.
5.1. Experimental Data and Setting
We use two types of data, which are trajectory data and POI data. The first data is trajectory data, and we use Geolife dataset as the trajectory data [26]; the second type of data is POI data (Beijing city), and we collect it from Tencent Web Service API.
Trajectory data: Our experiments are performed on the Geolife trajectory dataset. The Geolife dataset was obtained in the Geolife project by 182 users in Beijing over a period of more than 5 years (April 2007 to August 2012). Geolife is characterized by a series of timestamps, each timestamp contains latitude and longitude, and the dataset contains records of 24,876,978 GPS points. In the stay point detection, the time threshold is 5 minutes, and the distance threshold is 200m; we then get a total of 43,442 stay points. We remove users whose stay point records are less than 200, so the number of users is reduced from 182 to 50, and we will get 35,960 stay points. We then build a virtual grid in Beijing and map the coordinates of each stay point to the corresponding grid. The size of the grid is 500 × 500, there are 41,080 grids covering Beijing, the number of grids that the users have visited is 2,211. When generating trajectory sequence, we set the time window to one week and the sliding window to 10. Finally, we get 23,775 Trajectory sequences. Further details of the preprocessed dataset is shown in Table 2;
Rental data: POI data: We obtain the POI data of Beijing from the Tencent Web Service API. It provides detailed POI information such as coordinates, address, ID, name, phone number, type, etc. The Tencent Web Service API divides POIs into 18 main categories, the statistical distribution of which is shown in Figure 9. Therefore, we can calculate the semantic feature vector of each stay point based on these 18 types.
As for the experiment settings, all experiments were performed on a computer with Intel Core i9-10900K CPU 3.70 GHz, NVIDIA GeForce RTX 3090, 64GB RAM under Microsoft Windows 10. We divide each trajectory sequence into training and test sets. For each user, we use the first 80% of trajectory sequences as training data and the remaining 20% as testing data. We use Python 3.7 and Keras to put our model into practice. We adopt categorical cross entropy to minimize the loss function, the Adam optimizer is used to train our model, and dropout is used to prevent overfitting. The hidden state 𝑑, dropout rate, epoch, and the dimensions are set as 64, 0.5, 45, and 20, respectively.
The location prediction is a classification problem; the number of classes is determined by the grid covering the study area (a total of 2,211 grids in our study area). Since there are many classes to classify, both prediction accuracy and improvement rate will be low and hard to improve. To evaluate the performance of each method for next location prediction, we evaluate our model with TOP K accuracy (TOP@K), which checks whether the ground-truth location is shown among the Top K result list, the units are percentage points (%), where K = {1, 5, 10, 15}. The evaluation of higher-ranking results is more helpful in practical applications. We train each method 10 times and compute the average accuracy and standard deviation to avoid obtaining good results by training only once.
5.2. Internal Experiment
We conducted seven different internal experiments with different sliding windows, location features, temporal features, semantic features, user features and prediction models. Settings and defaults are listed in Table 3.
5.2.1. Time Interval
In this experiment, we compared the trajectory sequence without time interval setting (Figure 10) to trajectory sequence with time interval setting (Figure 11). Trajectory sequence with time interval setting means that the time interval of each point of a continuous sequence is the same (we set the time interval to 1 hour). Trajectory sequence without time interval setting means that the time interval of each location is inconsistent.
In Table 4, the two methods differ in how the training data is processed, and the testing data in both methods do not have time interval setting. We can observe that the performance without the time interval setting is better, the main reason is that the time interval setting will generate redundant data; many records of the trajectory sequence are the same location, and it may delete important locations. Furthermore, if the time interval of a location is less than 1 hour, this location will be deleted to ensure the same time interval.
5.2.2. Sliding Window
Secondly, we performed two experiments to observe the effect of sliding window size on the model. In the first experiment, we do not set the time window on any of the trajectory sequences to observe the effect of different sliding window size on the performance. In the second experiment, we set the time window to observe the different sliding window size on the performance. In Figure 12 (Experiment 1), we can see that the accuracy decreases as the sliding window increases. This is because there is no time window setting in this experiment, therefore the trajectory sequences with long time intervals may occur. This makes it hard to predict the next location. In Figure 13 (Experiment 2), it can be noted that the accuracy increases with the increase of the sliding window; in this experiment, the amount of data decreases as the size of the sliding window increases. This is because if the trajectory sequence length is less than the sliding window size, the trajectory sequence will be deleted. Based on these two experiments, we set the sliding window size to 10, which not only retains a certain number of trajectory sequences, but also maintains a certain accuracy.
5.2.3. Location Feature
The third internal experiment is location feature setting. We first test the performance of the model using different location feature extraction methods. Then we compare FC (Fully Connection), Word2vecc and Node2vec. The experiments are shown in Table 5. The results of FC and Word2vec are very similar. Because FC and Word2vec are essentially the same, the difference is that Word2vec is a pre-trained model, and FC is trained in the model. On the other hand, the results of Node2vec are slightly better; Node2vec considers the visiting frequency of nodes while fully considering the relationship between each node and its neighbor nodes, therefore it allows the model to fully consider the spatiotemporal contexts.
5.2.4. Temporal Feature
The fourth internal experiment is temporal feature setting, we compare different temporal feature extraction methods, Table 6 lists the experiment results. The first method uses the numbers 1-24 (hour) to represent temporal features. The second method is Sin&cos, it was proposed by Lu et al. [10], and they designed the coordinates on a unit circle using the sine and cosine function to represent the cyclic features of time. In this experiment, using hour or Sin&cos does not perform well. Time2vecs performs better because it considers both the original time feature and the periodic feature, which better expresses the periodicity and aperiodicity of time.
5.2.5. Semantic Feature
The fifth internal experiment is semantic feature setting. Before discussing semantic feature setting, we discuss the impact of semantic matching k. In this experiment, the model only has the semantic feature as input. In Figure 14, we can observe that when k increases, the performance of the prediction model shows an upward trend reaching its peak when k = 73 (so we set k = 73). Based on this experiment result, if the POI distribution around the stay point is considered, the accuracy can be effectively improved. If the closest distance method is used to search for POI (without considering POI distribution), when calculating the semantic vector of the stay point, this method may match the user to a different POI in close proximity instead of the POI which the user has actually visited.
We perform two experiments to discuss semantic feature setting. In the first experiment, the prediction model only considers the semantic features (Table 7). In the second experiment, all features are used as input for the prediction model (Table 8). We compared the closest distance (CD), semantic matching without considering home and workplace (SM*), and semantic matching (SM). A comparison of these methods is discussed below:
Comparing SM* with CD: the improvement rate of SM* reaches 15% in TOP@1(the model only considers semantic features). This shows that if the POI distribution around the stay point is considered, the accuracy can be effectively improved;
Comparing SM with CD: the improvement rate of SM reaches 45% in TOP@1 (the model only considers semantic features). SM uses sequential pattern mining to compute semantic feature vectors, so that the prediction model can more accurately capture the intent of user activities and achieve better performance.
5.2.6. Prediction Model
The sixth internal experiment is prediction model setting. We compare LSTM, BiLSTM and Self-Attention. LSTM is a variant of the RNN model, which is widely used to handle sequential data; BiLSTM is a combination of backward LSTM and forward LSTM, it is commonly used to model contextual features. The results are listed in Table 9. We can observe that Self-Attention has an advantage in predicting the next location, this is because Self-Attention is not like RNN-based models, which transmits long-term memory through many stages, thus losing some memory. Instead, the Self-Attention uses full connection to consider each stage in parallel, thus retaining more long-term information.
5.2.7. User Feature
The last internal experiment is user feature setting. We compared removing user ID with having user ID. The experimental results are shown in Table 10, which shows that having User ID increases performance while simultaneously personalizing the model.
5.3. External Experiment
In external experiments, to evaluate the performance of our model, we compare the proposed model with two different models:
SERM [23]: SERM jointly learns the embedding of various features (location, time, semantic and user), and uses LSTM to predict the next location;
MSSRM [16]: MSSRM jointly learns various features (user, location and time), uses Node2vec embedding to learn location features, and uses Time2vec to learn time features. LSTM is adopted to capture long short-term spatiotemporal dependencies, and Self-Attention is introduced to distinguish each location in different contexts.
We then perform four external experiments; parameter settings as mentioned in Experimental Data and Setting. The first is a comparison of different methods; the second is a comparison of different grid sizes; the third is a comparison of different sliding window sizes; the final comparison compares the performance of different methods on weekdays and weekends.
5.3.1. Comparison of Different Methods
A comparison of the different methods is shown in Table 11. Based on the results, our method has the best performance amongst the three methods. SERM uses fully connected layers to embed features and does not fully consider information regarding visiting order and visiting frequency; it uses LSTM to predict locations which may result in information loss when passing on long-term memory due to longer trajectory sequences. The main reason why MSSRM performs worse than our method is because MSSRM does not consider semantic features, thus losing important behavior patterns and living habits information. Our proposed method on the other hand, effectively extracts semantic features and captures the semantic-awareness spatiotemporal transformation to improve the performance of location prediction.
5.3.2. Comparison of Different Methods
The comparison of different grid sizes is shown in Table 12. We compare 3 different grid sizes (300×300, 500×500 and 700×700). We can observe that the smaller the grid size, the lower the accuracy. This is because as the grid size becomes smaller, the number of locations covered in the study area increases, which makes it difficult to predict the next location. The overall performance: Our Method > MSSRM > SERM.
5.3.3. Comparison of Different Sliding Window Sizes
The comparison of different sliding window sizes is shown in Table 13. This experiment compares 3 different sliding window sizes (7,10 and 13). Based on the table below, when the sliding window size is too large or too small, it is detrimental to the performance. Furthermore, although sliding window size has an impact on the accuracy, it still does not affect the advantages of our method.
5.3.4. Comparison of Different Methods on Weekdays and Weekends
We compared the performance of different methods on weekdays and weekends. Some trajectories cross both weekdays and weekends, we also compared these trajectories together. From Table 14, we can observe that each model performs best on weekdays because most users follow a similar behavior pattern throughout weekdays, thus their daily life patterns are easier to predict. Our model can perform better than both SERM and MSSRM regardless of weekdays and weekends.
5.4. Visualization of Location Prediction Results
We visualize our location prediction model which we show in Figure 15. We use the user's previous ten locations to predict the next location. Each grid has a probability, the redder the grid indicates a higher probability of being the next location, and the yellow stars refer to the ground truth. The first example is on the top left of the figure. We can observe that the user is traveling around Kunming Lake (a well-known spot in Beijing). The second example is on the top right of the figure. The user's trajectory sequence 1-2 is in Beijing Forestry University, the user then passes through Jianqingyuan Community to reach the China Academy of Building Research (3-5). The user then heads towards the Chinese Academy of Sciences (6-10) and then travels back to the China Academy of Building Research (ground truth). The third example is on the bottom left of the figure. The user's trajectory sequence 1-4 is initially in Jianqingyuan community (home); after visiting several different places (5-7), the user visits the Chinese Academy of Sciences to study (8-10) and then returns home (ground truth). The last one is at the bottom right of the figure, the user's trajectories sequence 1-8 are in the Peking university campus, the trajectories sequence 9-10 and ground truth are in the Tiantongwan community. Based on this trajectory sequence, we can observe that the user returned home to rest after finishing his class at Peking University.
6. Conclusions and Future Work
In this paper, we propose a Self-Attention for Next Location Prediction based on Semantic Mining. We design a semantic matching method which considers k nearest POI, fully considers the spatial features around the stay point, and combines sequential pattern mining to bring richer semantic features. To extract spatiotemporal contexts, we use Node2vec and Time2vec which fully consider the interaction of each location and the periodicity of the time series. Finally, we adopt Self-Attention to capture the spatiotemporal dependencies to predict the next location. Experiments in Geolife show that the semantic matching of this study improved by 45.78% in TOP@1 compared with the closest distance search for POI. In terms of location features and temporal features extraction, TOP@1 has improved by 0.8% and 3.7%, respectively. Compared with the baseline, the model proposed in this paper improved by 2.5% in TOP@1, and improved by 5.6% and 5.0% in TOP@5 and TOP@10, respectively. It can be observed that the accuracy and the improvement rate of models are quite low, because the location prediction is a classification problem; the number of classes is determined by the grid covering the study area (a total of 2211 grids in our study area). Since there are many categories to classify, the prediction accuracy will be lower.
In the future, we plan to mine other semantic behaviors (not just home and workplace) and combine them with semantic matching to better understand user’s life patterns. Given the same reciprocal distance when calculating the semantics of each stay point, it can reflect the POI features around a stay point; however, some POI categories with a large number of POIs may occupy most of the weight. In the future, we plan to control the weight of each POI category through the number of POIs in each category. There are also some features that were not considered in this study, such as moving distance or dwell time, we plan to integrate these features into the model to make the model better learn contextual features. In addition, it is possible to find better neural network models to improve prediction accuracy, such as variants of Attention mechanism and variants of different prediction models (RNN or LSTM).
Author Contributions
Conceptualization, Eric Hsueh-Chan Lu; methodology, Eric Hsueh-Chan Lu and You-Ru Lin; software, You-Ru Lin; validation, Eric Hsueh-Chan Lu and You-Ru Lin; formal analysis, Eric Hsueh-Chan Lu and You-Ru Lin; investigation, You-Ru Lin; resources, Eric Hsueh-Chan Lu; data curation, You-Ru Lin; writing—original draft preparation, Eric Hsueh-Chan Lu and You-Ru Lin; writing—review & editing, Eric Hsueh-Chan Lu; visualization, You-Ru Lin; supervision, Eric Hsueh-Chan Lu; project administration, Eric Hsueh-Chan Lu; funding acquisition, Eric Hsueh-Chan Lu. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by Ministry of Science and Technology, Taiwan, R.O.C., grant number MOST 111-2121-M-006-009- and The APC was funded by Ministry of Science and Technology, Taiwan, R.O.C..
Data Availability Statement
Restrictions apply to the availability of these data. Data was obtained from Geolife dataset and are available Zheng, Y., Xie, X., & Ma, W. Y. with the permission of Geolife dataset.
Conflicts of Interest
The authors declare no conflict of interest.
References
Al-Molegi, A.; Jabreel, M.; Martínez-Ballesté, A. Move, Attend and Predict: An Attention-Based Neural Model for People’s Movement Prediction. Pattern Recognition Letters2018, 112, 34–40. [Google Scholar] [CrossRef]
Chen, C. C.; Chiang, M. F. Trajectory Pattern Mining: Exploring Semantic and Time Information. In Proceedings of the IEEE Conference on Technologies and Applications of Artificial Intelligence, Hsinchu, Taiwan, Nov. 25-27 2016; pp. 130–137. [Google Scholar]
Chen, M.; Zuo, Y.; Jia, X.; Liu, Y.; Yu, X.; Zheng, K. CEM: A Convolutional Embedding Model for Predicting Next Locations. IEEE Transactions on Intelligent Transportation Systems2020, 23, 3349–3358. [Google Scholar] [CrossRef]
Cao, H., Xu; F., Sankaranarayanan, J.; Li, Y.; Samet, H. Habit2vec: Trajectory Semantic Embedding for Living Pattern Recognition in Population. IEEE Transactions on Mobile Computing2019, 19, 1096–1108.
Fernandes, R.; GL, R. D. S. A New Approach to Predict User Mobility Using Semantic Analysis and Machine Learning. Journal of Medical Systems2017, 41, 1–12. [Google Scholar] [CrossRef] [PubMed]
Feng, J.; Li, Y.; Yang, Z.; Qiu, Q.; Jin, D. Predicting Human Mobility with Semantic Motivation via Multi-Task Attentional Recurrent Networks. IEEE Transactions on Knowledge and Data Engineering2022, 34, 2360–2374. [Google Scholar] [CrossRef]
Grover, A.; Leskovec, J. Node2vec: Scalable Feature Learning for Networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, California, USA, Aug. 13-17 2016; pp. 855–864. [Google Scholar]
Jiang, J.; Pan, C.; Liu, H.; Yang, G. Predicting Human Mobility Based on Location Data Modeled by Markov Chains. In Proceedings of the IEEE Fourth International Conference on Ubiquitous Positioning, Indoor Navigation and Location Based Services, Shanghai, China, Nov. 3-4 2016; pp. 145–151. [Google Scholar]
Kazemi, S. M.; Goel, R.; Eghbali, S.; Ramanan, J.; Sahota, J.; Thakur, S.; Wu, S.; Smyth, C.; Poupart, P.; Brubaker, M. Time2vec: Learning A Vector Representation of Time. arXiv preprint arXiv:1907.05321, 2019.
Lu, E. H. C.; Lin, Z. Q. Rental Prediction in Bicycle-Sharing System Using Recurrent Neural Network. IEEE Access2020, 8, 92262–92274. [Google Scholar] [CrossRef]
Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient Estimation of Word Representations in Vector Space. arXiv preprint arXiv:1301.3781, 2013.
Sassi, A.; Brahimi, M.; Bechkit, W.; Bachir, A. Location Embedding and Deep Convolutional Neural Networks for Next Location Prediction. In Proceedings of the IEEE 44th LCN Symposium on Emerging Topics in Networking, Osnabrück, Germany, Oct. 14-17 2019; pp. 149–157. [Google Scholar]
Su, L.; Li, L. Trajectory Prediction Based on Machine Learning. IOP Conference Series: Materials Science and Engineering2020, 790, 012032. [Google Scholar] [CrossRef]
Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, California, USA, Dec. 4-9 2017; pp. 5998–6008. [Google Scholar]
Wang, S.; Li, A.; Xie, S.; Li, W.; Wang, B.; Yao, S.; Asif, M. A Spatial-Temporal Self-Attention Network (STSAN) for Location Prediction. Complexity2021, 6692313. [Google Scholar] [CrossRef]
Wen, S.; Zhang, X.; Cao, R.; Li, B.; Li, Y. MSSRM: A Multi-Embedding Based Self-Attention Spatio-Temporal Recurrent Model for Human Mobility Prediction. Human-Centric Computing and Information Sciences2021, 11. [Google Scholar] [CrossRef]
Xia, L.; Huang, Q.; Wu, D. Decision Tree-based Contextual Location Prediction from Mobile Device Logs. Mobile Information Systems2018, 1852861. [Google Scholar] [CrossRef]
Xia, Y.; Gong, Y.; Zhang, X.; Bae, H. Y. Location Prediction Based on Variable-order Markov Model and User's Spatio-Temporal Rule. In Proceedings of the IEEE Conference on Information and Communication Technology Convergence, Jeju Island, Korea, Oct. 17-19 2018; pp. 37–40. [Google Scholar]
Xu, M.; Han, J. Next Location Recommendation Based on Semantic-Behavior Prediction. In Proceedings of the 5th International Conference on Big Data and Computing, Chengdu, China, May 28-30 2020; pp. 65–73. [Google Scholar]
Xu, S.; Cao, J.; Legg, P.; Liu, B.; Li, S. Venue2vec: An Efficient Embedding Model for Fine-Grained User Location Prediction in Geo-Social Networks. IEEE Systems Journal2019, 14, 1740–1751. [Google Scholar] [CrossRef]
Xie, M.; Yin, H.; Wang, H.; Xu, F.; Chen, W.; Wang, S. Learning Graph-based POI Embedding for Location-Based Recommendation. In Proceedings of the 25th ACM International Conference on Information and Knowledge Management, Indianapolis, IN, USA, Oct. 24-28 2016; pp. 15–24. [Google Scholar]
Ye, Y.; Zheng, Y.; Chen, Y.; Feng, J.; Xie, X. Mining Individual Life Pattern Based on Location History. In Proceedings of the IEEE Tenth International Conference on Mobile Data Management: Systems, Services, and Middleware, Taipei, Taiwan, May 18-20 2009; pp. 1–10. [Google Scholar]
Yao, D.; Zhang, C.; Huang, J.; Bi, J. SERM: A Recurrent Model for Next Location Prediction in Semantic Trajectories. In Proceedings of the ACM on Conference on Information and Knowledge Management, Singapore, Singapore, Nov. 6-10 2017; pp. 2411–2414. [Google Scholar]
Ying, J. J. C.; Lee, W. C.; Weng, T. C.; Tseng, V. S. Semantic Trajectory Mining for Location Prediction. In Proceedings of the ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, Chicago, Illinois, USA, Nov. 1-4 2011; pp. 34–43. [Google Scholar]
Zhang, X.; Li, B.; Song, C.; Huang, Z.; Li, Y. SASRM: A Semantic and Attention Spatio-Temporal Recurrent Model for Next Location Prediction. In Proceedings of the IEEE International Joint Conference on Neural Networks, Glasgow, United Kingdom, Jul. 19-24 2020; pp. 1–8. [Google Scholar]
Zheng, Y.; Xie, X.; Ma, W. Y. GeoLife: A Collaborative Social Networking Service Among User, Location and Trajectory. IEEE Data(base) Engineering Bulletin2010, 33, 32–39. [Google Scholar]
Figure 1.
The framework of our research. Source: Icon made by Smashicons, Freepik, & dDara from www.flaticon.com.
Figure 1.
The framework of our research. Source: Icon made by Smashicons, Freepik, & dDara from www.flaticon.com.
Table 4.
The TOP@K (%) of different time interval setting.
Table 4.
The TOP@K (%) of different time interval setting.
Method
Top@1
Top@5
Top@10
Top@15
Same time interval
0.17
0.45
0.57
0.56
Without time interval
0.45
0.55
0.42
0.52
Table 5.
The TOP@K (%) of each location feature setting.
Table 5.
The TOP@K (%) of each location feature setting.
Method
Top@1
Top@5
Top@10
Top@15
None
0.22
0.51
0.31
0.33
FC
0.42
0.55
0.48
0.38
Word2vec
0.52
0.56
0.37
0.40
Node2vec
0.45
0.55
0.42
0.52
Table 6.
The TOP@K (%) of each temporal feature setting.
Table 6.
The TOP@K (%) of each temporal feature setting.
Method
Top@1
Top@5
Top@10
Top@15
None
0.39
0.49
0.36
60.730.50
Hour
0.44
0.74
0.64
0.53
Sin&cos
0.43
0.55
0.53
0.56
Time2vec
0.45
0.55
0.42
0.52
Table 7.
The TOP@K (%) of predict location only considering semantic feature setting.
Table 7.
The TOP@K (%) of predict location only considering semantic feature setting.
Method
Top@1
Top@5
Top@10
Top@15
CD
0.32
0.27
0.33
35.150.28
SM*(Improvement Rate)
9.270.72(15.01%)
0.74(31.47%)
0.79(24.56%)
0.72(19.97%)
SM(Improvement Rate)
0.31 (45.78%)
30.780.61 (42.89%)
0.66 (31.15%)
0.66 (24.80%)
Table 8.
The TOP@K (%) of each semantic feature setting.
Table 8.
The TOP@K (%) of each semantic feature setting.
Method
Top@1
Top@5
Top@10
Top@15
None
0.52
0.45
0.47
61.240.54
CD
0.45
0.61
0.57
0.52
SM*
0.25
0.49
0.60
0.63
SM
0.45
0.55
0.42
0.52
Table 9.
The TOP@K (%) of each prediction model setting.
Table 9.
The TOP@K (%) of each prediction model setting.
Method
Top@1
Top@5
Top@10
Top@15
LSTM
0.51
0.50
0.48
61.650.47
BiLSTM
0.42
0.61
0.68
0.72
Attention
0.45
0.55
0.42
0.52
Table 10.
The TOP@K (%) of each user feature setting.
Table 10.
The TOP@K (%) of each user feature setting.
Method
Top@1
Top@5
Top@10
Top@15
Remove User ID
0.45
0.53
0.56
59.220.63
Has User ID
0.45
0.55
0.42
0.52
Table 11.
The TOP@K (%) of different methods.
Table 11.
The TOP@K (%) of different methods.
Method
Top@1
Top@5
Top@10
Top@15
SERM
0.42
0.50
0.48
59.020.36
MSSRM
0.43
0.41
0.30
0.35
Our Method
0.45
0.55
0.42
0.52
Table 12.
The TOP@K (%) of different grid sizes.
Table 12.
The TOP@K (%) of different grid sizes.
Grid Size (m)
Method
Top@1
Top@5
Top@10
Top@15
300×300
SERM
0.61
0.48
0.57
48.710.47
MSSRM
0.49
0.37
0.47
0.46
Our Method
0.37
0.35
0.51
0.30
500×500
SERM
0.42
0.50
0.48
59.020.36
MSSRM
0.43
0.41
0.30
0.35
Our Method
0.45
0.55
0.42
0.52
700×700
SERM
0.42
0.31
0.44
66.790.52
MSSRM
0.61
0.52
0.41
0.57
Our Method
0.50
0.39
0.47
0.45
Table 13.
The TOP@K (%) of different sliding window sizes.
Table 13.
The TOP@K (%) of different sliding window sizes.
Window Size
Method
Top@1
Top@5
Top@10
Top@15
7
SERM
0.45
0.66
0.65
0.71
MSSRM
0.34
0.50
0.34
0.44
Our Method
21.310.51
47.970.27
57.250.30
61.330.47
10
SERM
0.42
0.50
0.48
0.36
MSSRM
0.43
0.41
0.30
0.35
Our Method
21.560.45
49.590.55
58.930.42
63.050.52
13
SERM
0.63
0.75
0.72
0.88
MSSRM
0.59
0.54
0.58
0.58
Our Method
20.910.49
48.170.50
56.820.54
60.720.49
Table 14.
The TOP@K (%) of weekdays and weekends.
Table 14.
The TOP@K (%) of weekdays and weekends.
Date
Method
Top@1
Top@5
Top@10
Top@15
Weekday
SERM
0.45
0.72
0.60
0.61
MSSRM
0.59
0.84
0.78
0.77
Our Method
22.980.51
50.840.83
60.340.79
64.120.76
Weekend
SERM
1.19
1.54
1.19
0.94
MSSRM
1.36
1.18
1.19
1.08
Our Method
21.580.82
51.821.19
61.150.76
65.290.66
Cross
SERM
0.49
0.55
0.74
0.48
MSSRM
0.53
0.50
0.62
0.59
Our Method
19.660.78
46.260.84
55.460.95
59.890.70
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.