2. Literature Review
In this section, a review of recent research endeavors within this domain, specifically focusing on the analysis of social media conversations about COVID-19 and Long COVID, with Twitter as the central platform of interest, is presented. Shamrat et al. [
50] applied the k-nearest neighbors (kNN) algorithm to categorize COVID-19-related tweets into three distinct classes—’ positive,’ ‘negative,’ and ‘neutral’. The first step of their study involved filtering relevant tweets. Following this initial filtration process, the kNN algorithm was applied for an in-depth analysis of these tweets and their associated sentiments. Sontayasara et al. [
51] utilized the capabilities of the Support Vector Machine (SVM) classifier to develop an algorithm tailored for sentiment analysis. Their work involved analysis of tweets wherein individuals articulated their intentions to visit Bangkok during the COVID-19 pandemic. By applying this SVM-based classifier, they successfully segregated the tweets into three distinct sentiment categories—‘positive,’ ‘negative,’ and ‘neutral’. Asgari-Chenaghlu et al. [
52] uncovered and analyzed the trending topics and the primary concerns related to the COVID-19 pandemic, as conveyed on Twitter. Their approach was characterized by detecting the pulse of Twitter discourse, pinpointing emergent themes, and unraveling the overarching concerns articulated by users. Amen et al. [
53] proposed a framework that employed a directed acyclic graph model to analyze COVID-19-related tweets. This framework was specifically designed to detect anomalies in the context of events, as expressed on Twitter.
Lyu et al.’s work [
54] centered around the development of an approach tailored to the examination of COVID-19-related tweets intricately intertwined with the guidelines set forth by the Centers for Disease Control and Prevention (CDC). Their overarching goal was to uncover a spectrum of public perceptions, encompassing concerns, attention, expectations, and more, all orbiting around the CDC’s directives concerning COVID-19. The work of Al-Ramahi et al. [
55] involved filtering and scrutinizing tweets about COVID-19 spanning the period from January 1, 2020, to October 27, 2020. In their work, they analyzed tweets wherein individuals vocalized their divergent viewpoints on the efficacy of mask-wearing as a strategy to curb the spread of COVID-19. Jain et al. [
56] proposed a methodology engineered to analyze COVID-19-related tweets, with the aim of bestowing an influence score upon the users who posted these tweets. The crux of their study lay in the identification of Twitter’s influential figures who actively engaged with the COVID-19 discourse, thereby shaping the narrative and impacting public opinion. Meanwhile, Madani et al. [
57] delved into the realm of machine learning, crafting a robust classifier that used random forest-based approaches to analyze tweets. This classifier’s primary functionality was to sift through tweets pertaining to COVID-19 and flag those that propagated fake news or misinformation. The classifier achieved an accuracy of 79%, underscoring its effectiveness in combatting the spread of misinformation during the pandemic.
Shokoohyar et al. [
58] developed a comprehensive system designed to delve into the expansive realm of Twitter discourse, where individuals voiced their opinions concerning the COVID-19-induced lockdown measures in the United States. Their aim was to gain profound insights into the multifaceted array of sentiments, perspectives, and reactions related to this pivotal aspect of the pandemic response. Chehal et al. [
59] developed specialized software to analyze the collective mindset of the Indian population, as reflected in their tweets during the two nationwide lockdowns enforced by the Indian government in response to the COVID-19 crisis. Glowacki et al. [
60] developed a systematic approach geared towards the identification and examination of COVID-19-related tweets, where Twitter’s community engaged in dialogues surrounding issues of addiction. This initiative provided a unique window into the evolving discourse on addiction amidst the backdrop of a global pandemic. Selman et al.’s explorative study [
61] centered on studying tweets where individuals shared heart-wrenching accounts of their loved ones succumbing to COVID-19. Their particular focus honed in on those instances where patients faced their final moments in isolation. This research illuminated the profoundly emotional and isolating aspects of the pandemic’s impact. Koh et al. [
62] endeavored to study tweets containing specific keywords, spotlighting conversations among Twitter users grappling with profound feelings of loneliness during the COVID-19 era. Their analysis, encompassing a corpus of 4492 tweets, provided valuable insights into the emotional toll of the pandemic. Mackey et al. [
63] zeroed in on tweets where individuals voluntarily reported their symptoms, experiences with testing sites, and recovery status in relation to COVID-19. Their work served as means for understanding the public’s firsthand experiences and interactions with healthcare systems. In [
64], Leung et al. aimed to decode the intricate nuances of anxiety and panic-buying behaviors during the COVID-19 pandemic, with a specific emphasis on the unprecedented rush to acquire toilet paper. Their methodology involved the analysis of 4081 tweets. This unique angle shed light on the peculiar buying trends and emotional undercurrents that manifested during the pandemic.
The broad focus of the research conducted by Pokharel [
65] was to perform sentiment analysis of tweets about the COVID-19 pandemic. Specifically, their dataset was developed to comprise tweets from Twitter users who had shared their location as ‘Nepal’ and had actively engaged in tweeting about COVID-19 between May 21, 2020, and May 31, 2020. The methodology involved using Tweepy and TextBlob to analyze the sentiments of these tweets. The findings of this study revealed a multifaceted tapestry of emotions coursing through the Nepali Twitter community during that period. While the prevailing sentiment was positivity and hope, there were discernible instances where emotions of fear, sadness, and even disgust surfaced. Vijay et al. [
66] conducted an in-depth examination of the sentiments conveyed in tweets related to COVID-19 from November 2019 to May 2020 in India. This study involved the categorization of all tweets into three distinct categories, namely ‘Positive,’ ‘Negative,’ and ‘Neutral.’ The findings of this investigation showed that initially, the majority of individuals seemed inclined towards posting negative tweets, perhaps reflecting the uncertainties and anxieties that characterized the early stages of the pandemic. However, as time progressed, there emerged a noticeable shift in the sentiment landscape, with a growing trend towards the expression of positive and neutral tweets.
The study conducted by Shofiya et al. [
67] involved targeting Twitter data that was uniquely associated with Canada and centered around text containing keywords related to social distancing. The research employed a two-pronged approach involving the utilization of the SentiStrength tool in conjunction with the Support Vector Machine (SVM) classifier to gauge the sentiments of these Tweets. A dataset comprising 629 tweets was analyzed in this study. The findings showed that 40% of the analyzed tweets had a neutral emotion, 35% of the tweets had a negative emotion, and 25% of the tweets expressed a positive emotion. The primary objective of the investigation conducted by Sahir et al. [
68] centered on the examination of public sentiment surrounding online learning during the outbreak of the COVID-19 pandemic, specifically in October 2020. This methodology involved document-based text mining and implementation of the Naïve Bayes classifier to detect the sentiments. The results showed that 25% of the tweets expressed a positive sentiment, 74% of the tweets expressed a negative sentiment, and a mere 1% of the tweets had a neutral sentiment. In a similar work [
69], a dataset of Tweets about online learning during COVID-19 was developed.
The study conducted by Pristiyono et al. [
70] aimed to delve into the perspectives and viewpoints held by the Indonesian population concerning the COVID-19 vaccine in the specific timeframe of January 2021. This process involved the utilization of RapidMiner for data collection. The results of the sentiment analysis showed that the percentage of negative tweets, positive tweets, and neutral tweets were 56%, 39%, and 1%, respectively. In a similar study [
71], the researchers analyzed a corpus of tweets pertaining to COVID-19 from Ireland. This dataset spanned an entire year, commencing on January 1, 2020, and concluding on December 31, 2020. The primary objective of this study was to conduct sentiment analysis. The findings showed that a significant portion of the sentiments was marked by robust criticism, with particular emphasis placed on Ireland’s COVID tracker app. The critique was predominantly centered on concerns related to privacy and data security. However, the study also showed that a noteworthy segment of sentiments also expressed positivity. These positive sentiments were directed towards the collective efforts and resilience towards COVID-19. Awoyemi et al. [
72] used the Twarc2 tool to extract a corpus of 62,232 Tweets about Long COVID posted between March 25, 2022, and April 1, 2022. The methodology involved the application of different natural language processing-based approaches, including the Latent Dirichlet Allocation (LDA) algorithm. The outcome of this analysis highlighted the top three prevailing sentiments in the Twitter discourse. Notably, trust emerged as a prominent sentiment, constituting 11.68% of the emotional spectrum, underlining a notable level of faith and reliance within the online community. Fear also held a significant presence, encompassing 11.26% of the emotional landscape, reflecting concerns and apprehensions embedded in the discussions. Alongside fear, sadness played a substantial role, constituting 9.76% of the emotional spectrum, indicative of the emotional weight carried by certain aspects of the discourse. The primary aim of the work of Pitroda et al. [
73] was to delve into the collective societal responses and attitudes surrounding the phenomenon of Long Covid, as expressed on Twitter. A dataset comprising 98,386 tweets posted between December 11, 2021, and December 20, 2021, was used for this study. The methodology for performing sentiment analysis utilized the AFFIN lexicon model. The findings showed that 44% of the tweets were negative, 34% of the tweets were positive, and 23% of the tweets were neutral. To summarize, there has been a significant amount of research and development in this field since the beginning of COVID-19. However, these works have the following limitations:
Lack of focus on Long COVID: As discussed in this section, a wide range of research questions in the context of COVID-19, such as – tourism [
51], trending topics [
52], concerns [
52], event analysis [
53], views towards wearing masks [
55], analysis of behaviors of influencers [
56], misinformation detection [
57], analysis of addiction [
60], detection of loneliness [
62], and panic buying [
64], have been explored and investigated in the last few months by studying and analyzing relevant Tweets. However, none of these works [
51,
52,
53,
55,
56,
57,
60,
62,
64] focused on the analysis of Tweets about Long COVID.
Limitations in the few works that exist on Long COVID: While there have been a few works, such as [
72,
73], that focused on the analysis of Tweets about Long COVID, the primary limitation of such works is the limited time range (for example: March 25, 2022, to April 1, 2022 in [
72] and December 11, 2021, to December 20, 2021 in [
73]) of the Tweets that were analyzed. Such limited time ranges represented only a small fraction of the total time Long COVID has impacted the global population.
The works in this field (for example: [
50], [
51], [
58], [
65,
66,
67,
68,
70,
71]) that involved sentiment analysis of Tweets about COVID-19, did not investigate sentiments associated with tweets about Long COVID.
Lack of focus on self-reporting of Long COVID: Analysis of self-reporting of symptoms related to various health conditions on Twitter has attracted the attention of researchers from different disciplines, as is evident from the recent works related to Twitter data analysis that focused on self-diagnosis or self-reporting of mental health problems [
74], autism [
75], dementia [
76], depression [
77], breast cancer [
78], swine flu [
79], flu [
80], chronic stress [
81], post-traumatic stress disorder [
82], and dental issues [
83], just to name a few. Since the outbreak of COVID-19, works in this field (such as [
63]) have focused on developing approaches to collect and analyze tweets where people self-reported symptoms of COVID-19. However, none of the prior works in this field have focused on investigating tweets comprising self-reporting of Long COVID.
The work of this paper aims to address these limitations by analyzing a corpus of Tweets where people self-reported symptoms of Long COVID. The step-by-step methodology that was followed for this work is described in
Section 3.
3. Methodology
Sentiment Analysis, often referred to as Opinion Mining, represents the computational exploration of individuals’ sentiments, viewpoints, and emotional expressions regarding a given subject. This subject can encompass various entities, such as individuals, events, or topics [
84]. The terms Sentiment Analysis (SA) and Opinion Mining (OM) are commonly used interchangeably, denoting a shared essence. Nevertheless, some scholars have posited subtle distinctions between OM and SA [
85,
86]. Opinion Mining, in its essence, involves the extraction and examination of people’s opinions pertaining to a specific entity. In contrast, Sentiment Analysis seeks to identify the sentiment embedded within a text and subsequently analyze it. Consequently, SA endeavors to unearth opinions, discern the sentiments they convey, and classify these sentiments based on their polarity. This classification process can be envisioned as a three-tiered hierarchy comprising document-level, sentence-level, and aspect-level SA. At the document level, the primary objective is to categorize an entire opinion document as either expressing a positive or negative sentiment. Here, the document functions as the fundamental information unit typically focused on a single overarching topic or subject. In sentence-level SA, the aim is to classify the sentiment within each individual sentence. The initial step involves distinguishing between subjective and objective sentences. Subsequently, for subjective sentences, Sentence-level SA ascertains whether they convey positive or negative opinions [
87]. It’s worth noting that Wilson et al. [
88] highlighted that sentiment expressions may not always possess a subjective nature. However, the distinction between document and sentence-level classifications is not fundamentally profound since sentences can be regarded as concise documents [
86]. While document and sentence-level classifications offer valuable insights, they often fall short of providing the granular details necessary for evaluating opinions on various facets of the entity. To obtain this comprehensive understanding, Aspect-level SA comes into play. This level of analysis endeavors to categorize sentiments with respect to specific aspects or attributes associated with entities. The initial step involves the identification of these entities and their respective facets. Importantly, opinion holders can articulate diverse sentiments concerning distinct aspects of the same entity. In essence, SA or OM encompasses a multifaceted process that spans various levels of analysis, from overarching documents to individual sentences and, ultimately, the nuanced evaluation of specific aspects related to entities. This comprehensive approach to sentiment analysis is invaluable in unveiling the intricate tapestry of opinions and emotions expressed in text data, enabling a deeper understanding of public sentiment in various contexts.
The problem of Sentiment Analysis can be investigated using diverse approaches such as manual classification, Linguistic Inquiry and Word Count (LIWC), Affective Norms for English Words (ANEW), the General Inquirer (GI), SentiWordNet, and machine learning-oriented techniques-based algorithms like Naive Bayes, Maximum Entropy, and Support Vector Machine (SVM). However, the specific methodology employed in this study was VADER, an acronym for Valence Aware Dictionary for sEntiment Reasoning. The rationale behind choosing VADER stems from its demonstrated prowess, surpassing manual classification in performance, and its ability to surmount the limitations encountered by comparable sentiment analysis techniques. The following represents an overview of the distinct features and properties of VADER [
91]:
- a)
VADER sets itself apart from LIWC by exhibiting heightened sensitivity to sentiment expressions thriving in the context of analysis of social media texts.
- b)
The General Inquirer falls short in its coverage of sentiment-relevant lexical features that are commonplace in social text, a limitation that VADER adeptly addresses.
- c)
The ANEW lexicon proves less responsive to the sentiment-relevant lexical features typically encountered in social text. This isn’t a limitation of VADER.
- d)
The SentiWordNet lexicon exhibits substantial noise, with a significant portion of its synsets lacking positive or negative polarity. This isn’t a limitation of VADER.
- e)
The Naïve Bayes classifier hinges on a simplistic assumption of feature independence, a limitation circumvented by VADER’s more nuanced approach.
- f)
The Maximum Entropy approach, devoid of the conditional independence assumption between features, factors in information entropy through feature weightings.
- g)
Machine learning classifiers, though robust, often demand extensive training data—a hurdle also faced by validated sentiment lexicons. Furthermore, machine learning classifiers rely on the training set to represent an extensive array of features.
VADER stands out by employing a sparse rule-based model, crafting a computational sentiment analysis engine tailored for social media-style text. It excels in its adaptability across various domains and requires no dedicated training data, relying instead on a versatile, valence-based, human-curated gold standard sentiment lexicon. VADER boasts efficiency, accommodating real-time analysis of streaming data, all while maintaining its computational performance, the time complexity of which is O(N). To add to this, VADER is accessible to all, with no subscription or purchase costs associated. Beyond polarity detection (positive, negative, and neutral), VADER possesses the capability to gauge the intensity of sentiment expressed in texts.
To develop the system architecture for sentiment analysis of this research work, RapidMiner was used [
92]. RapidMiner, originally known as Yet Another Learning Environment (YALE) [130], stands as a versatile data science platform empowering the conception, deployment, and application of an extensive array of algorithms and models within the realms of Machine Learning, Data Science, Artificial Intelligence, and Big Data. Its utility spans the domains of both scholarly research and the creation of pragmatic business-oriented applications and solutions. The comprehensive RapidMiner ecosystem encompasses a suite of integrated development environments, including (1) RapidMiner Studio, (2) RapidMiner Auto Model, (3) RapidMiner Turbo Prep, (4) RapidMiner Go, (5) RapidMiner Server, and (6) RapidMiner Radoop. It is worth mentioning here that in this paper, the specific reference to “RapidMiner” pertains exclusively to “RapidMiner Studio” and not to any of the alternative development environments affiliated with this software tool. RapidMiner Studio is designed as an open-core model, comprising a robust Graphical User Interface (GUI) that empowers developers to craft an array of applications and workflows while facilitating algorithm development and implementation. Within the RapidMiner development environment, specific operations or functions are denoted as “operators,” and a coherent sequence of these “operators” (configured linearly, hierarchically, or through a blend of both approaches) directed towards a particular task or objective is termed a “process.” RapidMiner streamlines the creation of these “processes” by offering a diverse selection of pre-built “operators” that can be directly employed. Furthermore, a specific class of “operators” is available to modify the defining characteristics of other “operators,” adding a layer of flexibility and customization to the analytical process.
To develop a “process” in RapidMiner for addressing this research problem, at first, a relevant dataset had to be identified. The dataset used for this work is a dataset of 1,244,051 Tweets about Long COVID, posted on Twitter between May 25, 2020, and January 31, 2023 [
93]. This dataset contains a diverse range of topics represented in the Tweets about Long COVID, present in two different files. To develop the “process” (shown in
Figure 1), these files were imported into RapidMiner and merged using the Append “operator”. As the focus of this study is to analyze Tweets about Long COVID where individuals self-reported Long COVID on Twitter, a text processing-based approach was applied to identify such Tweets. By reviewing similar works in this field that studied Tweets involving self-reporting of mental health problems [
74], autism [
75], dementia [
76], depression [
77], breast cancer [
78], swine flu [
79], flu [
80], chronic stress [
81], post-traumatic stress disorder [
82], and dental issues [
83], a Bag of words was developed that contained specific phrases related to the self-reporting of Long COVID. These phrases were – “I have long COVID”, “I am experiencing long COVID”, “I have been experiencing long COVID”, “I am suffering from long COVID”, “I am going through long covid”, “I have been feeling symptoms of long covid”, “I have had long covid”, “I have been diagnosed with long covid”, “I had long covid”, “I am feeling symptoms of long covid”, “I am experiencing long covid”, “I felt symptoms of long covid”, “I experienced long covid”, and “I am diagnosed with long covid”. These Bag of Words were applied through a data filter to the Tweets to obtain a corpus of 7,348 Tweets where each Tweet involved self-reporting Long COVID. Thereafter, data preprocessing of these Tweets was performed. The data processing comprised the following steps:
- a)
Removal of characters that are not alphabets.
- b)
Removal of URLs.
- c)
Removal of hashtags.
- d)
Removal of user mentions.
- e)
Detection of English words using tokenization.
- f)
Stemming and Lemmatization.
- g)
Removal of stop words
- h)
Removal of numbers
- i)
Replacing missing values.
Multiple “operators” were developed in RapidMiner to perform the data preprocessing. Thereafter, the Extract Sentiment operator in RapidMiner was customized to implement the VADER sentiment analysis approach on this corpus of 7,348 Tweets. The VADER methodology computed a compound sentiment score in the range of +4 to -4 for each of these Tweets. A rule-based approach was developed to identify the number of positive, negative, and neutral Tweets. This rule-based method functioned by analyzing the compound sentiment score of each Tweet. If this score was greater than 0, that Tweet was classified as a positive Tweet; if this score was less than 0, that Tweet was classified as a negative Tweet. Finally, if this score was 0, that Tweet was classified as a neutral Tweet. As the data processing and this rule-based approach for classification of the sentiment of a Tweet involved multiple “operators” working in a sequential manner, these were developed as “sub-processes” in the main “process” that was developed in RapidMiner.
Figure 1 shows this main “process” that was developed in RapidMiner to perform the sentiment analysis of these Tweets. In this “process”, “Dataset – F1” and “Dataset – F2” represent the two different files of this dataset [
93]. This “process” consists of two “sub-processes” for data preprocessing and rule-based detection of sentiment classes, which are shown in
Figure 2a,b, respectively. The results of running this “process” on all the Tweets in the dataset are presented and discussed in
Section 4.
4. Results and Discussions
This section presents the results of performing sentiment analysis of Tweets (posted on Twitter between May 25, 2020, and January 31, 2023) where individuals self-reported Long COVID. As described in
Section 3, a text-processing-based approach that involved the development of a Bag of Words was utilized to develop a corpus of Tweets, each of which involved self-reporting of Long COVID.
Figure 3 shows a monthly variation of these Tweets during this time range, i.e., between May 25, 2020, and January 31, 2023, using a histogram. To analyze this monthly variation of Tweets, a total of 32 bins were used. As can be seen from this Figure, the average number of Tweets per month where individuals self-reported Long COVID on Twitter, has been considerably high in 2022 as compared to the average number of Tweets per month in 2021.
Figure 4 shows the results of the RapidMiner “process” shown in
Figure 1. As can be seen from
Figure 4, the percentage of Tweets with positive, negative, and neutral sentiment were 43.12%, 42.65%, and 14.22%, respectively.
Thereafter, a comprehensive analysis of these Tweets was also performed to identify the number of Tweets in different fine-grain sentiment classes to better understand the distribution of specific emotions expressed by Twitter users when posting positive or negative Tweets, where they self-reported Long COVID. These fine-grain sentiment classes were sadness, neutral, fear, anger, joy, surprise, and disgust. The result of this analysis is shown in
Figure 5. As can be seen from
Figure 5, the emotion of sadness was expressed in most of these Tweets. It was followed by the emotions of fear, neutral, surprise, anger, joy, and disgust, respectively.
5. Conclusions
In the last decade and a half, humanity has confronted several infectious disease outbreaks, each of which had a toll on both lives and the global economy. During those outbreaks, social media platforms emerged as invaluable resources of information, capable of providing profound insights into the nuances and dynamics of these outbreaks. Leveraging social media data for syndromic surveillance, with a dedicated emphasis on public health matters disseminated through web-based content, has been of paramount significance to researchers across different disciplines. The term “Long COVID” describes the persistence of symptoms of COVID-19 for several weeks or even months following the initial SARS-CoV-2 infection. Since the outbreak of COVID-19, social media platforms, such as Twitter, have experienced a tremendous increase in conversations related to Long COVID. Recent works in this field have focused on sentiment analysis of Tweets related to COVID-19 to unveil the multifaceted spectrum of emotions and viewpoints held by the Twitter community. However, most of these works did not focus on Long COVID, and the few works that focused on Long COVID analyzed Tweets that were posted on Twitter during a very limited time range. Furthermore, no prior work in this field has investigated Tweets where individuals self-reported Long COVID. The work presented in this paper aims to address these research challenges by presenting multiple novel findings from a comprehensive analysis of a dataset comprising 1,244,051 Tweets about Long COVID, posted on Twitter between May 25, 2020, and January 31, 2023. First, the analysis shows that the average number of Tweets per month where individuals self-reported Long COVID on Twitter, has been considerably high in 2022 as compared to the average number of Tweets per month in 2021. Second, findings of sentiment analysis using VADER show that the percentage of Tweets with positive, negative, and neutral sentiment were 43.12%, 42.65%, and 14.22%, respectively. Third, the analysis of sentiments associated with the Tweets also shows that the emotion of sadness was expressed in most of these Tweets. It was followed by the emotions of fear, neutral, surprise, anger, joy, and disgust, respectively. Future work in this area would involve performing topic modeling of these Tweets to understand the specific topics that were represented in these discussions on Twitter in the context of long COVID.