Published on in Vol 8 (2024)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/49562, first published .
Identifying X (Formerly Twitter) Posts Relevant to Dementia and COVID-19: Machine Learning Approach

Identifying X (Formerly Twitter) Posts Relevant to Dementia and COVID-19: Machine Learning Approach

Identifying X (Formerly Twitter) Posts Relevant to Dementia and COVID-19: Machine Learning Approach

Original Paper

Department of Computer Science, University of Saskatchewan, Saskatoon, SK, Canada

Corresponding Author:

Raymond J Spiteri, PhD

Department of Computer Science

University of Saskatchewan

S425 Thorvaldson Building, 110 Science Place

Saskatoon, SK, S7N5C9

Canada

Phone: 1 306 966 2909

Email: spiteri@cs.usask.ca


Background: During the pandemic, patients with dementia were identified as a vulnerable population. X (formerly Twitter) became an important source of information for people seeking updates on COVID-19, and, therefore, identifying posts (formerly tweets) relevant to dementia can be an important support for patients with dementia and their caregivers. However, mining and coding relevant posts can be daunting due to the sheer volume and high percentage of irrelevant posts.

Objective: The objective of this study was to automate the identification of posts relevant to dementia and COVID-19 using natural language processing and machine learning (ML) algorithms.

Methods: We used a combination of natural language processing and ML algorithms with manually annotated posts to identify posts relevant to dementia and COVID-19. We used 3 data sets containing more than 100,000 posts and assessed the capability of various algorithms in correctly identifying relevant posts.

Results: Our results showed that (pretrained) transfer learning algorithms outperformed traditional ML algorithms in identifying posts relevant to dementia and COVID-19. Among the algorithms tested, the transfer learning algorithm A Lite Bidirectional Encoder Representations from Transformers (ALBERT) achieved an accuracy of 82.92% and an area under the curve of 83.53%. ALBERT substantially outperformed the other algorithms tested, further emphasizing the superior performance of transfer learning algorithms in the classification of posts.

Conclusions: Transfer learning algorithms such as ALBERT are highly effective in identifying topic-specific posts, even when trained with limited or adjacent data, highlighting their superiority over other ML algorithms and applicability to other studies involving analysis of social media posts. Such an automated approach reduces the workload of manual coding of posts and facilitates their analysis for researchers and policy makers to support patients with dementia and their caregivers and other vulnerable populations.

JMIR Form Res 2024;8:e49562

doi:10.2196/49562

Keywords



Dementia is a group of progressive syndromes that cause impairments in high-level cognitive functions, such as memory, language, and thinking, as well as everyday functioning and social interactions [1]. Currently, approximately 50 million people worldwide are affected by dementia, and this number is projected to continue to rise rapidly [2]. As populations age worldwide, dementia is expected to continue to be a significant health care challenge, requiring considerable medical, social, and institutional care [3].

The emergence of the COVID-19 pandemic presented an additional challenge for patients with dementia and their caregivers. Older patients were at a higher risk of contracting COVID-19 and were more likely to experience severe symptoms and consequences due to age, vulnerability, frailty, and other health conditions commonly associated with dementia [4]. In the United Kingdom, half (50%) of the COVID-19–related deaths in care homes were among patients with dementia [5]. Patients with dementia face challenges in adhering to self-protection protocols, such as wearing masks, using hand hygiene, and practicing physical distancing. In addition, they may have difficulty understanding and remembering the risks associated with COVID-19 [6]. In addition, the caregivers of patients with dementia also face limitations imposed by COVID-19, such as social isolation, loss of support, and care-partner burnout, all of which can further exacerbate the unpleasant situation of patients with dementia by limiting their access to public services and the support provided by their caregivers [6].

Social media platforms have emerged as valuable sources of information for individuals seeking updates on various issues, performing scientific studies, providing support, and raising public awareness. X (formerly Twitter), established in 2006 with about 400 million users, is a popular social media platform for microblogging where users can publicly share their thoughts using short messages called posts (formerly tweets). Posts can provide insight into COVID-19–related experiences of patients with dementia and their caregivers, offering an opportunity to study the impact of the pandemic on them. By identifying and analyzing posts relevant to dementia and COVID-19, substantial opportunities to support effective decision-making and policy development and further research can be uncovered.

Manual analysis, mining, and coding of posts can be a challenging, time-consuming, and laborious task, hindering the ability of health care researchers and practitioners to gain insights into the impact of COVID-19 on patients with dementia and other vulnerable populations. As such, there is a need to improve existing methodologies to streamline and automate the process. Machine learning (ML) algorithms have become a popular approach for performing tasks such as thematic or semantic analysis, and the classification of posts with reliable accuracy, offering a promising solution to this challenge [7-11].

ML algorithms have shown promise in classifying posts based on different sentiments. For example, Roy and Ojha [12] trained 3 classifiers using automatically labeled data in their study. However, the inclusion of a significant amount of noise through the automated labeling of data resulted in poor performance of their classifiers. Chiroma et al [13] manually labeled posts and used ML algorithms to classify suicide-related posts. The authors applied the Bag of Word technique for feature extraction and evaluated the performance of 5 traditional ML algorithms. Although there was no significant difference in performance among different algorithms for most classifications, binary classification showed the most promising results.

Despite these results, recent studies have raised concerns about the effectiveness of traditional ML algorithms for identifying and classifying posts [14,15]. Therefore, this study aims to demonstrate an alternative method for identifying posts relevant to dementia and COVID-19 using a combination of natural language processing (NLP) and ML algorithms and explore the reliability and performance of these algorithms. By leveraging the power of more advanced algorithms and techniques, specifically transfer learning algorithms, we aim to develop a model that can effectively analyze large volumes of posts to identify relevant posts and potentially offer valuable insights to policy makers, health care professionals, and researchers to inform evidence-based decision-making and support patients with dementia and their caregivers. Furthermore, the methodology described applies more generally to identify relevant social media posts to help facilitate further analysis. The overview and workflow of our study are shown in Figure 1.

Figure 1. The workflow of this study. (1) Data collection involved gathering posts containing search keywords from 3 distinct data sets: the First Wave data set consisting of 5063 posts from February 15, 2020, to September 7, 2020; the Longitudinal data set consisting of 110,528 posts from September 8, 2020, to December 8, 2021; and the Alzheimer’s Awareness Month data set consisting of 1289 posts from January 1, 2022, to January 31, 2022; (2) data preprocessing, and (3) classification/evaluation.

Data Collection

In this study, we aimed to analyze posts and identify those relevant to dementia and COVID-19. We used 3 data sets for this purpose. The first data set, referred to as the First Wave data set, consisted of 5063 posts collected between February 15, 2020, and September 7, 2020, using the search terms “dementia” or “Alzheimer” in combination with “COVID-19,” “COVID,” or “Corona” [16]. This data set captured the experiences of people impacted by Alzheimer disease or dementia and COVID-19 in the early stages of the pandemic. To capture the later stages of the pandemic, we collected a second data set, referred to as the longitudinal data set, consisting of 110,528 posts between September 8, 2020, and December 8, 2021, using the same search terms as the first data set. Finally, to examine the X discourse on dementia during Alzheimer’s Awareness Month in Canada, we collected a third data set, referred to as the Alzheimer’s Awareness Month data set, comprising 1289 posts between January 1, 2022, and January 31, 2022. Table 1 summarizes the particulars of the data sets, including the number of instances, and search terms for each data set. Figure S1 in Multimedia Appendix 1 depicts the word cloud of the most frequently used word in each data set.

Table 1. The characteristics of different data setsa.
Data setPosts, nTime frameFirst search termsSecond search terms
First Wave5063February 15, 2020, to September 7, 2020Dementia, AlzheimerCOVID-19, COVID, Corona
Longitudinal110,528September 8, 2020, to December 8, 2021Dementia, AlzheimerCOVID-19, COVID, Corona
Alzheimer’s Awareness Month1289January 1, 2022, to January 31, 2022#AlzheimersAwarenessMonth, #AlzAwareness, #dementiawareness, dementia month, dementia awareness month, Alzheimer’s awareness month, Alzheimer’s month, January is Alzheimer’s Awareness month@alzCanada, @AlzheimerOnt, @AlzheimerSK, @DementiaAB\_NT, @AlzheimerNS, @AlzheimerNB, @AlzheimerPEI, @alzheimerMB, @AlzheimerBC,
@asnl2,
@FqsaAlzh

aThe data applied in the study consist of 3 distinct data sets including different search keywords.

Data Preprocessing

To construct data sets consistently, preprocessing was performed on all 3 data sets to remove noise. Preparation steps such as simple filtering (removing emoticons, misspelled words, duplicates, obviously irrelevant characters, posts, and replies), tokenization, lemmatization, lowercasing, and normalization (removing punctuation, possessive pronouns, and stop words) were applied. After this filtering, the remaining posts were manually labeled as either relevant or irrelevant by a team of 11 coders using thematic analysis and the codebook from Bacsu et al [16]. Intercoder reliability was assessed through random reviewing of 25% (1560/6243) of all labeled posts and an average intercoder reliability of 83.4% was achieved.

Feature Extraction

After the preprocessing steps, the tokenized words were then converted into vectors using the term frequency–inverse document frequency (TF-IDF) method [17-19]. TF-IDF calculates the relative frequency of words in a specific document compared with the inverse proportion of that word over the entire text document determining the importance of a given word in a document. Accordingly, words frequently used in a document have higher TF-IDF weights and are regarded as more representative. Words common to many documents, such as articles and prepositions, have lower TF-IDF weights and are considered less representative [18,19]. The TF-IDF hyperparameter, max-features, determines the number of the most representative words for the rest of the analysis [18]. We set the max-features threshold to 1250 based on experimentation and the scores of different ML classifiers. This approach helps select the most representative words for the highest performance.

ML Algorithms

Overview

Various traditional and transfer learning ML algorithms were applied to identify posts relevant to dementia and COVID-19.

Traditional ML Algorithms
Logistic Regression

Logistic regression (LR) is a statistical ML method used for analyzing the relationship between a binary dependent variable and 1 or more independent variables. It is commonly used for classification tasks, where the goal is to predict the probability of an event occurring based on a set of input variables. The output of a logistic regression model is a probability value between 0 and 1, which can be interpreted as the likelihood of the event occurring [20].

Naïve Bayes

Naïve Bayes (NB) is a probabilistic ML algorithm that is based on the Bayes rule. It uses the assumptions of strong independence between variables to construct a simple and fast algorithm [21].

Multinomial Naïve Bayes

Multinomial Naïve Bayes (MNB) is a probabilistic ML algorithm based on Bayes’ theorem that calculates the probability of a particular event occurring based on prior knowledge of the conditions that may affect that event [22,23]. In text classification, the algorithm learns the probability distribution of words in each class and then calculates the conditional probability of a new document belonging to each class given its word frequency distribution.

K-Nearest Neighbors

K-nearest neighbors (kNN) is a nonparametric ML algorithm known for its simplicity and effectiveness. kNN classifies data based on the closeness of training samples in a given region [24].

Support Vector Machine

Support vector machine (SVM) is a widely used supervised ML algorithm for classification and regression analysis. SVMs are based on finding the best hyperplane that separates data into different classes. The hyperplane is chosen such that it maximizes the distance between the 2 classes [25].

Decision Tree

Decision tree (DT) is a supervised ML algorithm that is designed to solve classification problems by learning a hierarchy of “if/else” questions and answers and by creating a tree representation that results in a decision [26]. The goal of the DT classifier is to get the right classification result by asking the least number of “if/else” questions.

Ensemble Algorithms

Overview

Ensemble algorithms are designed to improve the accuracy of classification by combining multiple base classifiers. Although any type of base classifier including DT, neural networks, or SVMs can be used to create ensemble algorithms, DT is a commonly used algorithm [27].

Random Forest

Random forest (RF) is a powerful ensemble algorithm that uses DTs as the base classifiers. In RF, a set of features is randomly selected to determine the best split at each node of the DT. To make a prediction for a new data point, RF first applies each DT in the forest and predicts the target. Then, it uses the majority vote of all the DT predictions to assign the target with the highest probability to the new data [27,28].

AdaBoost

Adaboost (Adaptive Boosting) is a boosting algorithm that combines weak classifiers to create a strong classifier. It works by iteratively adjusting the weights of misclassified samples and adding new weak classifiers based on the samples that were misclassified in the previous iteration. This makes the algorithm adaptive, enabling it to focus on the more difficult samples in subsequent iterations. The final classification is determined by a weighted combination of the weak classifiers [29].

XGBoost

Extreme Gradient Boosting (XGBoost) is a distributed boosting algorithm that uses regression trees as a base classifier. It is designed to improve the accuracy of gradient boosting and is known for its high predictive power and speed, capable of being many times faster than other boosting techniques due to its parallel and distributed computing capabilities [30]. In addition, XGBoost performs well in sparse feature spaces [31] and uses more accurate approximations to find the best tree model.

Transfer Learning Algorithms

Overview

Transfer learning refers to transferring knowledge from different but related source domains to the target model in target domains in order to improve the performance of the target model. Consequently, a target model can be constructed without having to rely on a large number of domain data [11,32].

Bidirectional Encoder Representations From Transformers

Bidirectional Encoder Representations from Transformers (BERT) is a transfer learning algorithm that uses bidirectional transformers to create contextualized embeddings for each word in a sentence or text. It has achieved state-of-the-art performance in a wide range of NLP tasks, including text classification, question answering, and language understanding [33]. BERT has 2 phases: pretraining and fine-tuning. In the pretraining phase, the model is trained on a large corpus of text to learn general language patterns. In the fine-tuning phase, the pretrained model is adapted to a specific task by training it on a smaller, task-specific data set [34].

A Lite BERT

A Lite BERT (ALBERT) is a variant of BERT that reduces the number of parameters in the model while maintaining the same performance. This is achieved by applying 2 parameter reduction techniques: factorized embedding parameterization and cross-layer parameter sharing. ALBERT has been shown to have faster training times and better scalability than BERT, making it a useful option for large-scale NLP tasks [35].

Evaluation

To evaluate the performance of compared algorithms, we randomly partitioned the First Wave data set into 2 subsets: a training set used for cross-validation (CV; 4050/5063, 80% labeled posts) and a test set (1013/5063, 20% labeled posts). The CV set was further divided into 10 subsets, allowing us to construct and train 10 distinct models. The CV set was further partitioned into 10 subsets, enabling the construction and training of 10 distinct models. Using unseen test data is essential to evaluate the models’ generalization capability. This assessment with unseen data minimizes the risk of overfitting and offers a more dependable estimate of the algorithm’s real-world performance. Then, we selected the best-performing classifier and evaluated its reliability on unseen data using the longitudinal data set. A reliable algorithm should accurately predict the class of a large portion of the unseen data. Finally, we trained and tested the best-performing algorithm on the Alzheimer’s Awareness Month data set, using a training set (129/1289, 10% labeled posts) and a test set (1160/1289, 90% labeled posts). To assess the performance of the used ML algorithms, evaluation metrics including accuracy, precision, sensitivity, specificity, F1-score, and area under the curve (AUC) were used. These metrics provide a comprehensive understanding of the performance of algorithms in terms of correctly identifying relevant posts and minimizing false positives and false negatives.

Parameter Sensitivity Analysis

To optimize the performance of competing algorithms, it is necessary to tune their parameters. However, there is no one-size-fits-all approach for parameter selection. Therefore, in this study, we explored different sets of parameter values for each algorithm to identify the optimal configuration. A summary of the ML algorithms and their respective parameters used in this study can be found in Table 2.

Table 2. Parameters of different MLa algorithmsb.
AlgorithmParameters
LRc
  • C: [0.1, 0.5, 1,2]
  • Multiclass: [ovr]
  • Solver: [Newton-CG, IBFGS, Liblinear, SAG, SAGA]
kNNd
  • Number of neighbors: [3, 5, 7, …, 51]
  • Metrics: [Euclidian, Manhattan]
SVMe
  • C: [0.1, 0.5, 1, 1.5, 2]
  • Kernel: [linear, poly, RBF, sigmoid]
DTf
  • Criterion: [Entropy, Gini]
  • Min sample leaf: [1,2,4,6,8]
  • Min sample split: [1, 2, 3, …, 10, …, 20]
RFg
  • N estimators: [100, 200, 300, 400, 500]
  • Criterion: [Entropy, Gini]
  • Max depth: [2, 4, 6, …, 32, …, 64]
Adaboosth
  • Base estimators: [LR, DT, SVM]
  • N estimators: [10, 20, 30, …100, …, 500]
  • Max depth: [1, 2, 3, 4, …, 20]
XGBoosti
  • Max depth: [1, 2, 3, 4, …, 19, 20]
  • Learning rate: [0.01, 0.015, 0.02, 0.025, …, 0.1]

aML: machine learning.

bEach algorithm has different parameters. The values in brackets represent values for each parameter.

cLR: logistic regression.

dKNN: k-nearest neighbor.

eSVM: support vector machine.

fDT: decision tree.

gRF: random forest.

hAdaboost: Adaptive Boosting.

iXGBoost: Extreme Gradient Boosting.

Implementation

Posts were obtained using TWINT, a powerful scraping tool that allows for the scraping of posts without the requirement for an X account and the use of X’s application programming interface, enhancing the number of posts scraped, frequency, and time period of scrapes. We used Python programming language (version 3.8.5; Python Software Foundation) using NLTK [36], Scikit-Learn [37], and TensorFlow [38] libraries. The computations were performed on an Nvidia Tesla T4 GPU (Nvidia Corporation) within Google Colab.

Ethical Considerations

Ethical considerations about social media research suggest that publicly available data (eg, posts on X) can be used for research studies without requiring additional consent or ethics approval [39]. In this study, we did not apply for the ethics approval. In addition, because we did not engage or interfere with the users whose publicly posted content was collected and analyzed, we did not require informed consent. Nonetheless, to ensure users’ anonymity and protect their privacy, any related identifying personal information (eg, user IDs and usernames) has been removed. There was no compensation provided to the users whose public X posts were used.


Parameter Sensitivity Analysis

In this study, several algorithms were used with multiple parameters that were tuned to obtain optimal performance. AdaBoost performed best with a base estimator of DT, N-estimator set to 100, and a max depth set to 5. DT achieved the best performance with the criterion parameter set to “entropy,” a minimum sample leaf set to 100 and a minimum sample split set to 5. KNN performed best with 5 neighbors and with the “Euclidean” distance metric. LR achieved the best performance with a parameter c of 0.5, “ovr” selected for multiclass and “liblinear” selected for solver parameters. RF performed best with “Gini” as the criterion parameter, N-estimator set to 300, and max depth set to 32. SVM performed best with a radial basis function kernel and C set to 5. Finally, XGBoost performed best when max depth and learning rate were set to 6 and 0.015, respectively.

Performance Analysis

Study 1

The performance of various algorithms in identifying posts relevant to dementia and COVID-19 was evaluated with the First Wave data set. SVM achieved the highest accuracy (96.17%), precision (93.08%), sensitivity (98.04%), specificity (94.86%), F1-score (95.49%), and AUC (99.41%) for the training set, followed by the RF algorithm obtaining the second-best performance across all metrics. Table 3 summarizes the performance metrics of different algorithms for the training set.

For the test set, the ALBERT algorithm achieved the best accuracy (82.92%), precision (74.55%), sensitivity (84.83%), specificity (82.24%), F1-score (78.81%), and AUC (83.53%) among all the algorithms tested. The BERT algorithm achieved the second-best performance for these metrics. Performance metrics of different algorithms for the test set are summarized in Table 4, and the receiver operating characteristic curves of competing algorithms for the training and test sets are shown in Figure 2A and 2B, respectively.

Based on the performance of various algorithms on the test set (Table 4), the ALBERT algorithm demonstrated promising performance in class-wise measurements and accurately identifying posts relevant to dementia and COVID-19. Therefore, we concluded that the ALBERT algorithm shows reliable performance for this task. The class-wise results of competing algorithms using confusion matrices are shown in Figure 3.

Table 3. The performance of different algorithms for the training set using the First Wave data set.
AlgorithmMean accuracy % (SD)Mean precision % (SD)Mean sensitivity % (SD)Mean specificity % (SD)Mean F1-score % (SD)Mean AUCa % (SD)
Logistic regression80.51 (0.18)68.39 (0.54)83.88 (0.32)78.65 (0.19)75.35 (0.31)88.96 (0.1)
Naïve Bayes78.91 (0.29)62.57 (0.62)85.07 (0.48)76.02 (0.27)72.1 (0.41)87.86 (0.13)
Multinomial Naïve Bayes79.3 (0.56)71.82 (4.46)79 (2.67)79.74 (1.82)75.06 (1.74)82.23 (1.51)
K-nearest neighbor78.41 (0.32)69.78 (2.61)78.41 (1.79)78.53 (1.08)73.78 (0.67)87.42 (0.23)
Support vector machine96.17 (0.13)b93.08 (0.23)98.04 (0.22)94.86 (0.15)95.49 (0.15)99.41 (0.05)
Decision tree77.39 (0.62)68.87 (1.48)76.84 (1.18)77.76 (0.72)72.62 (0.87)86.19 (0.79)
Random forest88.66 (0.34)80.83 (1.03)92.17 (0.42)86.49 (0.54)86.12 (0.52)96.87 (0.09)
AdaBoost75.08 (0.85)76.2 (2.34)69.6 (1.68)80.21 (0.98)72.7 (0.74)83.86 (0.55)
XGBoost73.99 (0.57)59.9 (5.67)75.65 (2.64)73.41 (1.94)66.59 (2.57)82.8 (0.24)
BERTc81.52 (0.1)75.42 (1.04)85.13 (1.01)81.78 (0.2)80.23 (1.32)82.98 (0.4)
ALBERTd81.78 (0.45)74.98 (0.98)84.08 (0.82)82.19 (0.61)79.27 (0.5)83.13 (0.45)

aAUC: area under the curve.

bThe best values for the performance metrics are in italics.

cBERT: Bidirectional Encoder Representations from Transformers.

dALBERT: A Lite BERT.

Table 4. The performance of different algorithms for test set using the First Wave data set.
AlgorithmMean accuracy % (SD)Mean precision % (SD)Mean sensitivity % (SD)Mean specificity % (SD)Mean F1-score % (SD)Mean AUCa % (SD)
Logistic regression75.13 (0.32)61.97 (2.81)76.55 (2.85)74.39 (2.38)68.43 (1.96)82.21 (0.99)
Naïve Bayes74.26 (0.48)56.19 (3.39)78.8 (3.91)72.29 (2.65)65.5 (2.65)82.13 (1.16)
Multinomial Naïve Bayes72.1 (2.67)63.6 (4.26)69.92 (4.52)73.75 (1.92)66.43 (2.78)74.37 (1.87)
K-nearest neighbor64.31 (1.79)53.54 (3.98)60.49 (4.74)66.94 (1.97)56.6 (2.5)73.87 (2.72)
Support vector machine74.75 (0.22)62.04 (2.2)75.65 (2.99)74.26 (2.1)68.13 (1.93)82.09 (1.09)
Decision tree66.49 (1.18)56.03 (2.98)62.97 (3.21)68.73 (2.4)59.25 (2.52)69.64 (2.18)
Random forest73.37 (0.42)56.98 (2.4)75.91 (3.79)72.15 (2.06)65.05 (2.46)80.48 (1.82)
AdaBoost68.13 (1.68)67.98 (4.67)62.44 (3.65)73.54 (3.04)64.94 (2.6)73.67 (1.65)
XGBoost70.32 (2.64)54.6 (5.89)70.75 (3.52)70.29 (2.88)61.4 (3.82)76.82 (1.93)
BERTb81.03 (1.01)73.87 (3.43)80.21 (3.1)77.61 (3.3)77.87 (2.34)80.89 (2.3)
ALBERTc82.92 (0.82)d74.55 (8.91)84.83 (5.86)82.24 (5.06)78.81 (4.93)83.53 (3.14)

aAUC: area under the curve.

bBERT: Bidirectional Encoder Representations from Transformers.

cALBERT: A Lite BERT.

dThe best values for the performance metrics are in italics.

Figure 2. Receiver operating characteristic curves for competing machine learning algorithms for training (A) and test sets (B). The corresponding area under the curve values are given for each algorithm. ROC: receiver operating characteristic; LR: logistics regression; AUC: area under the curve; NB: Naïve Bayes; MNB: Multinomial Naïve Bayes; kNN: k-nearest neighbor; SVM: support vector machine; DT: decision tree; RF: random forest; BERT: Bidirectional Encoder Representations from Transformers; ALBERT: A Lite BERT.
Figure 3. Confusion matrices produced by the competing algorithms for the test set using the First Wave data set. LR: logistics regression; NB: Naïve Bayes; MNB: Multinomial Naïve Bayes; kNN: k-nearest neighbor; SVM: support vector machine; DT: decision tree; RF: random forest; BERT: Bidirectional Encoder Representations from Transformers; ALBERT: A Lite BERT.
Study 2

To extend our study to other conditions and evaluate the reliability of the best-performing algorithm in identifying posts relevant to dementia and COVID-19, we applied it to the longitudinal data set. Because this data set shares similarities with the one used in study 1, the results of study 2 can indicate whether the same methods can be used for similar data sets involving mental health disorders, pandemics, and the like. The ALBERT algorithm was used to test the longitudinal data set, and the results, as shown in Table 5, confirmed its effectiveness in correctly identifying posts relevant to dementia and COVID-19 at the second stage of the pandemic.

Although the ALBERT algorithm demonstrated promising performance in study 2, it was important to assess the generalizability and robustness of the methodology introduced by using a similar data set and search terms. Testing an algorithm on a data set with similar words as the training data can result in overfitting and may not generalize well to other data sets. Furthermore, a customized method for a given data set is not overly useful. Accordingly, study 3 was conducted to evaluate the performance of the ALBERT algorithm on a new and unrelated yet adjacent data set.

Table 5. The performance of the ALBERTa algorithm using the longitudinal data set.
AlgorithmMean accuracy %Mean precision %Mean sensitivity %Mean specificity %Mean F1-score %Mean AUCb %
ALBERT81.2386.0279.4583.4182.2481.46

aALBERT: A Lite Bidirectional Encoder Representations from Transformers.

bAUC: area under the curve.

Study 3

Study 3 aimed to assess the generalizability of the ML algorithm in identifying posts relevant to different mental disorders and pandemics by examining the X discourse on dementia during Alzheimer’s Awareness Month in Canada. Given the variability of health issues and the possibility of different pandemics, it is highly beneficial to have a model that performs well across different contexts. To evaluate the performance of the ALBERT algorithm in this context, we retrained the algorithm using 10% (129/1289 labeled posts) of the Alzheimer’s Awareness Month data set (addressing different mental health disorders and containing different search terms from those in the First Wave and longitudinal data sets) and tested it on the remaining 90% (1160/1289 labeled posts). As shown in Table 6, using the Alzheimer’s Awareness Month data set ALBERT algorithm achieved a reliable and acceptable performance of 80% or higher in all the metrics considered. The results challenged any potential bias of the ALBERT algorithm toward the First Wave and longitudinal data sets and confirmed that it can be applied to identify and classify posts relevant to different mental health disorders during pandemics.

Table 6. The performance of the ALBERTa algorithm for training and test sets using the Alzheimer’s Awareness Month data set.

Mean accuracy %Mean precision %Mean sensitivity %Mean specificity %Mean F1-score %Mean AUCb %
Training set86.3989.2184.9887.2482.7786.24
Test set80.6283.4079.9880.6281.6580.30

aALBERT: A Lite Bidirectional Encoder Representations from Transformers.

bAUC: area under the curve.


Principal Findings

The COVID-19 pandemic has highlighted the needs of vulnerable populations, including patients with dementia and their caregivers. Social media platforms, particularly X, provide a valuable source of data for health care researchers, governments, policy makers, and practitioners to understand the impact of the pandemic on this population. However, the sheer volume and high percentage of irrelevant social media data make it difficult to manually mine, analyze, and identify relevant posts for this purpose.

In this study, we aimed to identify posts relevant to dementia and COVID-19 using a combination of NLP and ML algorithms. Our findings demonstrated that transfer learning algorithms such as ALBERT outperform traditional ML algorithms in identifying relevant posts. The results of study 1 showed the superiority of the ALBERT algorithm and demonstrated that it achieved the best performance for the task of identifying relevant posts. In addition, in study 2, this algorithm demonstrated its capability in identifying posts that share similar content with the data set used in study 1. This indicated that transfer learning algorithms can effectively identify posts relevant to similar disorders for which the algorithm was trained. Furthermore, study 3 revealed a high level of generalizability of these algorithms. This suggests that transfer learning algorithms can be trained for a specific disorder or pandemic and can be applied to different (especially adjacent) disorders or pandemics with comparable performance.

Our study highlights the significant applicability and value of automated approaches in identifying posts relevant to COVID-19 and dementia, and the methodology is applicable to other studies involving the analysis of social media posts. A further application of this study is potential real-time monitoring of public health sentiment during pandemics or other public health crises. Tools can be developed to continuously analyze social media data to track the sentiments, concerns, and information needs of patients with dementia, their caregivers, and the general public. This real-time monitoring can provide invaluable insights to public health authorities, researchers, and policy makers, enabling them to decide about responses, allocate resources efficiently, and provide targeted support to vulnerable populations such as patients with dementia.

Limitations

Limitations of this study mainly involve the inherent biases in using X data for research. First, X users represent a particular portion of the population, often more tech-savvy, and this bias can result in the underrepresentation of older or less internet-connected demographics. In addition, X data may not capture the perspectives of individuals who do not use this platform for discussing health-related topics, potentially leading to an inaccurate perspective on public sentiment. Furthermore, linguistic and cultural biases can exist in the language and topics discussed on X, and these biases should also be considered when interpreting the findings. Moreover, this study focused on posts filtered by the keywords “Coronavirus” and “dementia.” Although the methods and analysis were tested on different data sets, they are particularly relevant to data related to COVID-19 and dementia. To increase generalizability, future research could explore the use of various other search keywords. Another limitation is the unexplored possible effect of post length on algorithm performance. Because the study did not consider post length, future research could include sensitivity analysis to determine its impact on algorithm performance. Finally, although posts reflect people’s thoughts at a specific moment, textual features alone may not fully reflect overall sentiments. Various syntactic and semantic post features could be explored to enhance algorithm performance. These limitations do not undermine this study’s contributions.

Conclusions

The COVID-19 pandemic has created a pressing need for policy makers to make timely decisions to address the needs of vulnerable populations, such as patients with dementia and their caregivers. Social media platforms, particularly X, offer a wealth of data that can provide invaluable insights into COVID-19 and its impact on patients with dementia and other vulnerable populations. However, the sheer volume and high fraction of irrelevant data make it difficult to extract and analyze relevant content. An automated tool that can accurately identify posts relevant to a given topic would be of great value to health care researchers, social scientists, policy makers, and practitioners in analyzing such posts.

In this study, we explored the use of an automated approach to identify posts relevant to dementia and COVID-19 using various ML algorithms. Our study shows that transfer learning algorithms outperform traditional ML algorithms, with the ALBERT algorithm achieving the best performance among the algorithms tested. The reliability of the results was confirmed using independent data sets, highlighting the ability of these algorithms to be part of an automated methodology for identifying posts relevant to dementia and COVID-19. Such a methodology can also be applied to help facilitate other studies that involve analysis of social media posts, and it can aid in effective and timely decision-making during times of acute need.

Future research could include the analysis of multiple features of posts to further increase identification performance and the reliability of the algorithms. In addition, as pandemic or other communicable disease-related data continue to become available on social media, the development of a range of analytics and ML-based solutions based on processing such data can lead to enhanced support for patients and their caregivers. Other future research directions include use of specific medical-based models to improve performance, application of the approach to nonmedical data to test its generality, and incorporation of cross-language recognition to gain insight from other cultures.

Acknowledgments

The authors gratefully acknowledge the financial support from the Natural Sciences and Engineering Research Council of Canada under Discovery Grant (RGPN 2020-04467), Mitacs, and Refresh Inc.

Data Availability

The data sets generated or analyzed during this study are available from the corresponding author on reasonable request.

Authors' Contributions

All authors were involved in the conceptualization of the idea and writing and reviewing the paper. MA designed and implemented the experiment and collected, curated, analyzed, and visualized the data. AAJ analyzed and visualized the data. RJS looked after the administration, supervised the study, and provided funding and resources.

Conflicts of Interest

None declared.

Multimedia Appendix 1

The word cloud of the most frequently used word in each data set.

DOCX File , 277 KB

  1. Ding Z, Leung PY, Lee TL, Chan AS. Effectiveness of lifestyle medicine on cognitive functions in mild cognitive impairments and dementia: a systematic review on randomized controlled trials. Ageing Res Rev. 2023;86:101886. [CrossRef] [Medline]
  2. Masterson-Algar P, Allen MC, Hyde M, Keating N, Windle G. Exploring the impact of COVID-19 on the care and quality of life of people with dementia and their carers: a scoping review. Dementia. 2021;21(2):648-676. [FREE Full text] [CrossRef]
  3. Lindeza P, Rodrigues M, Costa J, Guerreiro M, Rosa MM. Impact of dementia on informal care: a systematic review of family caregivers' perceptions. BMJ Support Palliat Care. 2024;14:e38-e49. [FREE Full text] [CrossRef] [Medline]
  4. O'Connor MK, Nicholson R, Epstein C, Donley T, Salant R, Nguyen AH, et al. Telehealth support for dementia caregivers during the COVID-19 pandemic: lessons learned from the NYU family support program. Am J Geriatr Psychiatry. 2023;31(1):14-21. [FREE Full text] [CrossRef] [Medline]
  5. ONS report shows 52% increase in excess deaths of people dying of dementia—Alzheimer's Society comments. Alzheimer's Society. 2020. URL: https:/​/www.​alzheimers.org.uk/​news/​2020-06-05/​ons-report-shows-52-increase-excess-deaths-people-dying-dementia-alzheimers-society [accessed 2024-05-02]
  6. Wang Q, Davis PB, Gurney ME, Xu R. COVID-19 and dementia: analyses of risk, disparity, and outcomes from electronic health records in the US. Alzheimers Dement. 2021;17(8):1297-1306. [FREE Full text] [CrossRef] [Medline]
  7. Chayangkoon N, Srivihok A. Text classification model for methamphetamine-related tweets in Southeast Asia using dual data preprocessing techniques. Int J Electr Comput Eng. 2021;11(4):3617. [FREE Full text] [CrossRef]
  8. Khattak AM, Batool R, Satti FA, Hussain J, Khan WA, Khan AM, et al. Tweets classification and sentiment analysis for personalized tweets recommendation. Complexity. 2020;2020:1-11. [FREE Full text] [CrossRef]
  9. Daouadi KE, Rebaï RZ, Amous I. Optimizing semantic deep forest for tweet topic classification. Inf Syst. 2021;101:101801. [CrossRef]
  10. Sarsam SM, Al-Samarraie H, Alzahrani AI, Alnumay W, Smith AP. A lexicon-based approach to detecting suicide-related messages on Twitter. Biomed Signal Process Control. 2021;65:102355. [CrossRef]
  11. Jamali AA, Berger C, Spiteri RJ. Momentary depressive feeling detection using X (formerly Twitter) data: contextual language approach. JMIR AI. 2023;2:e49531. [FREE Full text] [CrossRef]
  12. Roy A, Ojha M. Twitter sentiment analysis using deep learning models. 2020. Presented at: 2020 IEEE 17th India Council International Conference (INDICON); December 10-13, 2020; New Delhi, India. [CrossRef]
  13. Chiroma F, Liu H, Cocea M. Suicide related text classification with prism algorithm. 2018. Presented at: 2018 International Conference on Machine Learning and Cybernetics (ICMLC); July 15-18, 2018; Chengdu, China. [CrossRef]
  14. Gulati K, Kumar SS, Boddu RSK, Sarvakar K, Sharma DK, Nomani MZM. Comparative analysis of machine learning-based classification models using sentiment classification of tweets related to COVID-19 pandemic. Mater Today Proc. 2022;51:38-41. [CrossRef]
  15. Satu MS, Khan MI, Mahmud M, Uddin S, Summers MA, Quinn JMW, et al. TClustVID: a novel machine learning classification model to investigate topics and sentiment in COVID-19 tweets. Knowl Based Syst. 2021;226:107126. [FREE Full text] [CrossRef] [Medline]
  16. Bacsu JD, O'Connell ME, Cammer A, Azizi M, Grewal K, Poole L, et al. Using Twitter to understand the COVID-19 experiences of people with dementia: infodemiology study. J Med Internet Res. 2021;23(2):e26254. [FREE Full text] [CrossRef] [Medline]
  17. Singh AK, Shashi M. Vectorization of text documents for identifying unifiable news articles. Int J Adv Comput Sci Appl. 2019;10(7):305-310. [FREE Full text] [CrossRef]
  18. Zhang Y, Zhou Y, Yao J. Feature extraction with TF-IDF and game-theoretic shadowed sets. In: Wilbik A, Bouchon-Meunier B, Carvalho JP, Reformat MZ, Lesot MJ, Yager RR, et al, editors. Information Processing and Management of Uncertainty in Knowledge-Based Systems. Cham, Switzerland. Springer International Publishing; 2020:722-733.
  19. Ramos J. Using TF-IDF to determine word relevance in document queries. Proc First Instr Conf Mach Learn. 2003;242(1):29-48. [FREE Full text]
  20. Riffenburgh RH, Gillen DL. Logistic regression for binary outcomes. In: Statistics in Medicine, 4th Edition. London, United Kingdom. Academic Press; 2020:437-457.
  21. Huang Y, Li L. Naive Bayes classification algorithm based on small sample set. 2011. Presented at: 2011 IEEE International Conference on Cloud Computing and Intelligence Systems; September 15-17, 2011; Beijing, China. [CrossRef]
  22. Jiang L, Wang S, Li C, Zhang L. Structure extended multinomial naive Bayes. Inf Sci. 2016;329:346-356. [CrossRef]
  23. Kibriya A, Frank E, Pfahringer B, Holmes G. Multinomial naive Bayes for text categorization revisited. In: Webb GI, Webb G, Yu X, editors. AI 2004: Advances in Artificial Intelligence. Berlin, Heidelberg. Springer Berlin Heidelberg; 2005:488-499.
  24. Taunk K, De S, Verma S, Swetapadma A. A brief review of nearest neighbor algorithm for learning and classification. 2019. Presented at: 2019 International Conference on Intelligent Computing and Control Systems (ICCS); May 15-17, 2019; Madurai, India. [CrossRef]
  25. Alsagri HS, Ykhlef M. Machine learning-based approach for depression detection in Twitter using content and activity features. IEICE Trans Inf Syst. 2020;E103.D(8):1825-1832. [FREE Full text] [CrossRef]
  26. Müller AC, Guido S. Introduction to Machine Learning with Python: A Guide for Data Scientists. Sebastopol, California. O'Reilly Media, Inc; 2016.
  27. Che D, Liu Q, Rasheed K, Tao X. Decision tree and ensemble learning algorithms with their applications in bioinformatics. Adv Exp Med Biol. 2011;696:191-199. [CrossRef] [Medline]
  28. Breiman L. Bagging predictors. Mach Learn. 1996;24(2):123-140. [FREE Full text] [CrossRef]
  29. Kowsari K, Meimandi KJ, Heidarysafa M, Mendu S, Barnes L, Brown D. Text classification algorithms: a survey. Information. 2019;10(4):150. [FREE Full text] [CrossRef]
  30. Chen T, Guestrin C. XGBoost: a scalable tree boosting system. 2016. Presented at: KDD '16: The 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; August 13-17, 2016; San Francisco, California, USA. [CrossRef]
  31. Zeng W, Gautam A, Huson DH. On the application of advanced machine learning methods to analyze enhanced, multimodal data from persons infected with COVID-19. Computation. 2021;9(1):4. [FREE Full text] [CrossRef]
  32. Zhuang F, Qi Z, Duan K, Xi D, Zhu Y, Zhu H, et al. A comprehensive survey on transfer learning. Proc IEEE. 2021;109(1):43-76. [CrossRef]
  33. Deepa M. Bidirectional Encoder Representations from Transformers (BERT) language model for sentiment analysis task. Turk J Comput Math Educ. 2021;12(7):1708-1721. [FREE Full text]
  34. González-Carvajal S, Garrido-Merchán EC. Comparing BERT against traditional machine learning text classification. Journal of Computational and Cognitive Engineering. 2020;2(4):352-356. [FREE Full text]
  35. Lan Z, Chen M, Goodman S, Gimpel K, Sharma P, Soricut R. Albert: a lite bert for self-supervised learning of language representations. INTERNATIONAL CONFERENCE ON LEARNING REPRESENTATIONS (ICLR). 2020:1-17. [FREE Full text]
  36. Perkins J. Python Text Processing With NLTK 2.0 Cookbook. England. PACKT publishing; 2010.
  37. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, et al. Scikit-learn: machine learning in Python. J Mach Learn Res. 2011;12:2825-2830. [FREE Full text]
  38. Silaparasetty N. The tensorflow machine learning library. In: Machine Learning Concepts with Python and the Jupyter Notebook Environment: Using Tensorflow 2.0. New York, NY. Springer; 2020:149-171.
  39. Williams ML, Burnap P, Sloan L. Towards an ethical framework for publishing Twitter data in social research: taking into account users' views, online context and algorithmic estimation. Sociology. 2017;51(6):1149-1168. [FREE Full text] [CrossRef] [Medline]


ALBERT: A lite BERT
AUC: area under the curve
BERT: Bidirectional Encoder Representations from Transformers
CV: cross-validation
DT: decision tree
kNN: k-nearest neighbor
LR: logistics regression
ML: machine learning
MNB: multinomial naïve Bayes
NB: naïve Bayes
NLP: natural language processing
RF: random forest
SVM: support vector machine
TF-IDF: term frequency–inverse document frequency
XGBoost: Extreme Gradient Boosting


Edited by A Mavragani; submitted 01.06.23; peer-reviewed by LH Yao, M Chatzimina; comments to author 30.11.23; revised version received 11.12.23; accepted 03.04.24; published 04.06.24.

Copyright

©Mehrnoosh Azizi, Ali Akbar Jamali, Raymond J Spiteri. Originally published in JMIR Formative Research (https://formative.jmir.org), 04.06.2024.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on https://formative.jmir.org, as well as this copyright and license information must be included.