<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JFR</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR Form Res</journal-id>
      <journal-title>JMIR Formative Research</journal-title>
      <issn pub-type="epub">2561-326X</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v9i1e64536</article-id>
      <article-id pub-id-type="pmid">40810448</article-id>
      <article-id pub-id-type="doi">10.2196/64536</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Early Diagnosis of Knee Osteoarthritis With a Natural Language Processing–Driven Approach Based on Clinician Notes: Development and Validation Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Mavragani</surname>
            <given-names>Amaryllis</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Kimia</surname>
            <given-names>Amir</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Azar</surname>
            <given-names>Danielle</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author">
          <name name-style="western">
            <surname>Thanyakunsajja</surname>
            <given-names>Narathip</given-names>
          </name>
          <degrees>MS</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0007-1086-8830</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Jitkajornwanich</surname>
            <given-names>Kulsawasd</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <address>
            <institution/>
            <institution>Texas Tech University</institution>
            <addr-line>2500 Broadway</addr-line>
            <addr-line>Lubbock, TX, 79409</addr-line>
            <country>United States</country>
            <phone>1 6824149164</phone>
            <email>kulsawasd.jitkajornwanich@ttu.edu</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-6926-7577</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Xu</surname>
            <given-names>Shan</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2251-8682</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Shin</surname>
            <given-names>Donghee</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-5439-4493</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author">
          <name name-style="western">
            <surname>Charoenporn</surname>
            <given-names>Pattama</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2213-7169</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>King Mongkut’s Institute of Technology Ladkrabang</institution>
        <addr-line>Bangkok</addr-line>
        <country>Thailand</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Texas Tech University</institution>
        <addr-line>Lubbock, TX</addr-line>
        <country>United States</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Kulsawasd Jitkajornwanich <email>kulsawasd.jitkajornwanich@ttu.edu</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <year>2025</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>14</day>
        <month>8</month>
        <year>2025</year>
      </pub-date>
      <volume>9</volume>
      <elocation-id>e64536</elocation-id>
      <history>
        <date date-type="received">
          <day>19</day>
          <month>7</month>
          <year>2024</year>
        </date>
        <date date-type="rev-request">
          <day>1</day>
          <month>10</month>
          <year>2024</year>
        </date>
        <date date-type="rev-recd">
          <day>20</day>
          <month>1</month>
          <year>2025</year>
        </date>
        <date date-type="accepted">
          <day>5</day>
          <month>5</month>
          <year>2025</year>
        </date>
      </history>
      <copyright-statement>©Narathip Thanyakunsajja, Kulsawasd Jitkajornwanich, Shan Xu, Donghee Shin, Pattama Charoenporn. Originally published in JMIR Formative Research (https://formative.jmir.org), 14.08.2025.</copyright-statement>
      <copyright-year>2025</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on https://formative.jmir.org, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://formative.jmir.org/2025/1/e64536" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Knee osteoarthritis (OA) is a common form of knee arthritis that can cause significant disability and affect a patient’s quality of life. Although this disease is chronic and irreversible, the patient’s condition can be improved and the progression of the disease can be prevented if the disease is diagnosed early and the patient receives appropriate treatment immediately. Therefore, the prediction of knee OA is considered one of the essential steps to effectively diagnose and prevent further severe OA conditions. Knee OA is commonly diagnosed by medical experts or physicians, and the diagnosis of OA is mainly based on patients’ laboratory results and medical images, including x-ray and magnetic resonance images. However, diagnosis through such data is often time-consuming. Moreover, the diagnosis results can vary among physicians depending on their expertise. Previous studies mostly focused on using approaches, such as those involving artificial intelligence, to automatically detect knee OA through such data. However, these studies did not incorporate clinicians’ or doctors’ notes (text data) into the analysis, although these data involving reported symptoms and behaviors are already available and easier to collect and access than laboratory data and image data.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>We propose a novel natural language processing–driven approach based on clinicians’ or doctors’ notes of patient-reported symptoms (text data only) for diagnosing knee OA.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>The textual information from clinicians’ or doctors’ notes was first preprocessed using text analysis algorithms with respect to natural language processing. We then incorporated deep learning models, including convolutional neural networks, bidirectional long short-term memory (BiLSTM), and gated recurrent units. Lastly, a disease-specific standard questionnaire called WOMAC (Western Ontario and McMaster Universities Arthritis Index) was taken into account to improve the overall performance of the models.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>Our experiment included 5849 records (OA: 3455; non-OA: 2394). Before applying our WOMAC-based processing approach, the best-performing model was BiLSTM (area under the curve, 0.85; accuracy, 0.87; precision, 0.85; sensitivity, 0.95; specificity, 0.76; <italic>F</italic><sub>1</sub>-score, 0.90), and there was an improvement in the results with BiLSTM after applying our approach (area under the curve, 0.91; accuracy, 0.91; precision, 0.91; sensitivity, 0.94; specificity, 0.87; <italic>F</italic><sub>1</sub>-score, 0.93).</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>Our proposed method for predicting the occurrence of knee OA showed better performance than other conventional methods that use image data and statistical laboratory data. The findings indicate the feasibility of using text data (symptom descriptions reported by patients and recorded by doctors) to predict knee OA. Medical notes of symptom reports can be considered a valuable data source for predicting whether a particular knee is likely to experience OA progression.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>computational communication</kwd>
        <kwd>health informatics</kwd>
        <kwd>natural language processing</kwd>
        <kwd>artificial intelligence</kwd>
        <kwd>text mining</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <sec>
        <title>Background</title>
        <p>Knee osteoarthritis (OA) is a long-term condition, and its cause is still unknown. The most common form of arthritis involves cartilage degradation or bone abnormalities [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref3">3</xref>]. Joint pain, stiffness, and limitation of knee movement are some of the primary symptoms of knee OA. Knee OA is a slow but aggressive disease, and as the damage to the cartilage and bones cannot be restored, a prevention protocol could be beneficial [<xref ref-type="bibr" rid="ref4">4</xref>]. Knee OA affects 10%-15% of adults worldwide. In 2015, the World Health Organization (WHO) estimated that 9.6% of men and 18% of women aged 60 years or older had symptomatic knee OA. Knee OA symptoms affect mobility in 80% of cases, while 25% of people with symptomatic knee OA struggle to perform their daily activities. Knee OA can not only affect individuals both physically and psychologically but also create a burden on society and the economy [<xref ref-type="bibr" rid="ref5">5</xref>]. In 2008, knee OA was ranked as one of the most expensive diseases to treat. The cost of hospital treatment for arthroplasty surgery has reached approximately US $40 billion worldwide [<xref ref-type="bibr" rid="ref6">6</xref>]. Having said that, knee OA can be diagnosed at an early stage of the disease. Early detection and diagnosis are possible and crucial to overcome the issues associated with this highly disabling disease, as early-stage treatment can prevent further breakdown of the bone and cartilage [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref7">7</xref>].</p>
        <p>Owing to the presence of big data and the advancement of artificial intelligence (AI), many researchers are using these to identify key factors influencing health and disease-related issues with the help of machine learning models. Furthermore, deep learning is one of the new and rising trends and opportunities in the field of machine learning for many studies in diverse domains. This is particularly true in the medical domain, where deep learning and natural language processing (NLP) could be used for medical text classification tasks [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>]. This has led to the emergence of new research areas in medical text classification and summarization. Machine learning approaches for text classification can be roughly divided into 2 types: traditional machine learning approaches and deep learning–based approaches. Because traditional approaches have a limited ability to process large text data, which requires complex computations in the corpus, the performance of models is greatly affected. On the other hand, deep learning models have been shown to be highly effective in text classification [<xref ref-type="bibr" rid="ref10">10</xref>]. Moreover, many deep learning models, such as recurrent neural networks (RNNs), long short-term memory (LSTM), convolutional neural networks (CNNs), and gated recurrent units (GRUs), are widely used in the field of text classification [<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref12">12</xref>]. To use deep learning in NLP, we need a form of word representation to transform words into vectors, known as word embeddings. One of the most widely used methods for this is Word2Vec [<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref14">14</xref>], which we adopted in our work.</p>
        <p>In this study, we have proposed an NLP model to detect the occurrence of knee OA by incorporating deep learning models (GRUs, bidirectional long short-term memory [BiLSTM], and CNNs) based on the textual data of knee OA symptoms reported in doctors’ notes. Moreover, we have performed a literature review. The methodology of the proposed sentiment analysis has been explained in the Methods section, and the results have been provided in the Results section. Finally, discussions, conclusions, and future work have been communicated in the Discussion section.</p>
      </sec>
      <sec>
        <title>Literature Review</title>
        <p>In the medical AI–related literature, deep learning has been used to help detect and diagnose diseases. Often, each study is differentiated by the type of data used, such as images and statistical data, including laboratory results. We have discussed several studies in this section related to the use of AI, machine learning, and deep learning for detecting and predicting knee OA based on both medical images and statistical data, including laboratory results.</p>
        <p>In the study by Brahim et al [<xref ref-type="bibr" rid="ref1">1</xref>], x-ray images of the knees of patients, together with machine learning algorithms, were used to develop a computer-aided diagnosis (CAD) system for detecting early knee OA. The machine learning algorithms used were naïve Bayes and random forest models, which were adopted for classification tasks. The approach used 1025 knee x-ray images from the public Knee Osteoarthritis Initiative database [<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref16">16</xref>]. Yong et al [<xref ref-type="bibr" rid="ref17">17</xref>] proposed a method to detect Kellgren/Lawrence (KL) grading [<xref ref-type="bibr" rid="ref18">18</xref>], using an ordinal regression module (ORM) in neural networks. The module can predict the inclusion of images in multiple classes that follow a natural severity order, ranging from KL grades 0 to 4. They also used data from the Knee Osteoarthritis Initiative [<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref16">16</xref>]. A total of 4130 pairs of knee images were labeled with KL grades and collected from the baseline cohort. These studies yielded good results, but they only considered x-ray images (data were obtained from x-ray machines). Halilaj et al [<xref ref-type="bibr" rid="ref19">19</xref>] proposed a mixed effects model to differentiate clusters of knee OA progression. They used data from 1234 patients in the Knee Osteoarthritis Initiative, who were classified as being at high risk of developing knee OA based on age, BMI, and medical and occupational histories, to identify knee OA progression. They also developed models to predict the probability of belonging to a cluster using Least Absolute Shrinkage and Selection Operator (LASSO) regression. Lim et al [<xref ref-type="bibr" rid="ref6">6</xref>] proposed a feed-forward neural network trained with standard back propagation as a deep learning model to detect the occurrence of knee OA using the statistics of patients, including medical and health-related behavioral information. They used data from 5749 subjects in the Korea National Health and Nutrition Examination Survey [<xref ref-type="bibr" rid="ref20">20</xref>], which was performed by the Korea Centers for Disease Control and Prevention (KCDC). Wang et al [<xref ref-type="bibr" rid="ref21">21</xref>] developed a model to assess the severity of radiographic knee OA and identify key risk factors for early intervention and treatment. To predict the KL grade, they used a LSTM model with an attention mechanism based on data from 518 patients in the Knee Osteoarthritis Initiative [<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref16">16</xref>]. From these previous studies, we noticed that even though the statistical methods were faster than the image methods, the high accuracy of the image methods yielded better results compared to the statistical methods. Recently, Tiulpin et al [<xref ref-type="bibr" rid="ref22">22</xref>] developed a machine learning approach to predict the progression of structural knee OA, using raw radiographic image data, physical examination results, patient medical history, anthropometric information, and KL grade as features. They included data from the Knee Osteoarthritis Initiative (2711 subjects) [<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref16">16</xref>] in the training set and data from the Multicenter Osteoarthritis Study (2129 subjects) [<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref23">23</xref>] in the test set. The experiments combined both the image method and statistical method and yielded good performance. However, this approach considered x-ray images, which need to be retrieved from x-ray machines and take a long time to process.</p>
        <p>Recently, the field of NLP has shown significant technological advances, especially in text classification involving assigning a label or class to a given text. Text classification has many practical applications, such as spam detection in email servers, news categorization, sentiment analysis, hate speech detection in social media, etc. It has also been applied to medical fields. For example, Liu et al [<xref ref-type="bibr" rid="ref8">8</xref>] applied text classification to label words retrieved from social media big data as medical terms and categorized them accordingly.</p>
        <p>Traditional text classification approaches are dictionary-based and involve basic machine learning methods [<xref ref-type="bibr" rid="ref24">24</xref>]. Differentiating texts related to patients is challenging. This is because the texts used in the medical field are still unstructured data and can be complicated to process, especially when processed by basic machine learning approaches [<xref ref-type="bibr" rid="ref25">25</xref>]. On the other hand, deep learning methods have shown strong performance in image classification as well as speech recognition. These methods have been widely used in the fields of NLP and text classification and have yielded good results [<xref ref-type="bibr" rid="ref26">26</xref>]. CNNs and RNNs are among the most widely adopted deep learning models for text classification.</p>
        <p>Jang et al [<xref ref-type="bibr" rid="ref24">24</xref>] proposed a method to classify movie user reviews into positive, negative, and neutral categories using a sentiment analysis model that combines BiLSTM and CNN. The Internet Movie Database (IMDb) movie review dataset was used for this experiment. The model used a CNN to reduce the number of input features. The features extracted from the CNN were input into a BiLSTM model for the final classification. Qing et al [<xref ref-type="bibr" rid="ref26">26</xref>] proposed a neural network approach to classify medical texts into one of four dataset categories: (1) clinical records from Traditional Chinese Medicine in the Classified Medical Records of Distinguished Physicians Continued Two (TCM), (2) an open inpatient medical records dataset from the China Conference on Knowledge Graph and Semantic Computing (CCKS), (3) a corpus of biomedical publication abstracts labeled for cancer hallmarks, and (4) a dataset of Activating Invasion and Metastasis (AIM). They employed a convolutional layer to extract features from sentences and used a bidirectional gated recurrent unit (Bi-GRU) for classification.</p>
        <p>Many previous studies have focused on image and statistical data for knee OA prediction with machine learning and deep learning techniques. However, a limitation of image data is that they must be generated and retrieved through x-ray machines or magnetic resonance imaging (MRI) scanners. Moreover, statistical data have limitations associated with the integrity of the factors and variables related to patients.</p>
      </sec>
      <sec>
        <title>Objectives</title>
        <p>NLP or text classification has potential and could be beneficial in knee OA prediction. For the early detection of knee OA, we propose a novel approach based on an NLP or text classification method that uses only textual data, instead of x-ray or MRI image data and statistical data, and incorporates deep learning models, including GRU, BiLSTM, and CNN models. We used textual data in our research, which included clinicians’ or doctors’ notes on patient symptoms related to knee OA.</p>
      </sec>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Ethical Considerations</title>
        <p>Data permission was approved and granted by the study hospital and the associated institutions according to the institutional review board (IRB2023-736; develop language models for OA diagnosis). Regarding our dataset, no identifiable data of patients were collected in this study. The CONSORT (Consolidated Standards of Reporting Trials) (<xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>), CONSORT-EHEALTH (Consolidated Standards of Reporting Trials of Electronic and Mobile Health Applications and Online Telehealth) (<xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>), and STARD (Standards for Reporting of Diagnostic Accuracy Studies) (<xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>) checklists [<xref ref-type="bibr" rid="ref27">27</xref>-<xref ref-type="bibr" rid="ref29">29</xref>] were followed, and training was provided in human ethics for conducting research in humans.</p>
      </sec>
      <sec>
        <title>Resources</title>
        <p>All models were trained using a MacBook Pro 2018 laptop having an Intel Core i5 quad-core CPU (2.3 GHz), an Intel Iris Plus 655 GPU, and a Macintosh disk storage (256 GB).</p>
      </sec>
      <sec>
        <title>Study Design and Data Sources</title>
        <p>The approach comprised 6 stages. First, deidentified data on patients’ reported symptoms from clinicians’ or doctors’ notes were obtained from a hospital in Thailand (N=6737; knee OA: 3945, non-OA: 2792). No personal attributes or biospecimen data, such as laboratory results, were used. Second, data selection, filtering, cleaning, and preparation processes were performed for the data of patients who met the criteria of the experiments (N=5849; knee OA: 3455, non-OA: 2394). Third, text analysis and mining processes were performed on the datasets. Fourth, to enhance the efficiency of the model, we incorporated our text analysis and mining models with the WOMAC (Western Ontario and McMaster Universities Osteoarthritis Index) and specific concepts in our AI model. Fifth, to conform to reporting guidelines, we randomized our data into 3 groups: training, validation, and test sets. Finally, we applied and conducted comparison experiments with deep learning models (ie, CNNs, BiLSTM, and GRUs) to predict knee OA diagnosis.</p>
        <p>Texts in the medical field mainly contain 3 types of words. The first category is standardized medical terminology, which refers to specific concepts in the medical field. The second category is user-based health terminology, such as “can’t sleep” and “stomach pain.” The third category includes words that are considered noise, such as “because” and “so.” These words typically lack relevant health information and can cause interference [<xref ref-type="bibr" rid="ref8">8</xref>]. Our dataset primarily combined the first 2 categories. However, the data in this work consisted of words describing symptoms, which can be varied (such as pain [“<inline-graphic xlink:href="formative_v9i1e64536_fig8.png" xlink:type="simple" mimetype="image"/>” in Thai] and ache [“<inline-graphic xlink:href="formative_v9i1e64536_fig9.png" xlink:type="simple" mimetype="image"/>” in Thai], where both words mean “pain”). Furthermore, the data may have some incorrect words in sentences. Therefore, one of the core steps in our methodology was to use a WOMAC-based processing approach to transform some words in sentences that are similar in the WOMAC corpus to medical terminology by applying the <italic>SequenceMatcher</italic> library [<xref ref-type="bibr" rid="ref30">30</xref>]. We defined a cutoff score of similarity of 0.85 to replace old words with WOMAC words.</p>
        <p>To conform to appropriate reporting guidelines, CONSORT 2010 [<xref ref-type="bibr" rid="ref27">27</xref>] and CONSORT-EHEALTH [<xref ref-type="bibr" rid="ref28">28</xref>] guidelines were applied in randomizing our data into 3 groups: training, validation, and test sets. Moreover, STARD 2015 [<xref ref-type="bibr" rid="ref29">29</xref>] guidelines were used to ensure the quality and clarity of reporting of our diagnostic accuracy studies. <xref rid="figure1" ref-type="fig">Figure 1</xref> illustrates the process in detail.</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Schematic representation of our knee osteoarthritis prediction model, consisting of (A) textual raw symptom reports, (B) data cleaning, (C) text processing and analysis, (D) randomizing data (with respect to CONSORT [Consolidated Standards of Reporting Trials], CONSORT-EHEALTH [Consolidated Standards of Reporting Trials of Electronic and Mobile Health Applications and Online Telehealth], and STARD [Standards for Reporting of Diagnostic Accuracy Studies] checklists), (E) training models, and (F) classification. WOMAC: Western Ontario and McMaster Universities Osteoarthritis Index.</p>
          </caption>
          <graphic xlink:href="formative_v9i1e64536_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Data Preparation and Experiment Plan</title>
        <sec>
          <title>Data Collection</title>
          <p>The text reports of knee OA were provided by a hospital in Thailand and were collected by orthopedists between 2019 and 2021. Specifically, we focused only on textual knee OA reports on symptoms in the Thai language, and we considered only data from patients who first visited the hospital. To conduct fair experiments, we only considered patients who were new and had not been treated by orthopedists before, as we focused on patients who had not received treatment by orthopedists. We did not consider patients who had been treated previously, as their symptoms may not be clearly mentioned in their reports, which can degrade the accuracy of our model and predictions. The medical terminology (from the WOMAC corpus) was downloaded from the Royal College of Physiatrists of Thailand [<xref ref-type="bibr" rid="ref31">31</xref>].</p>
        </sec>
        <sec>
          <title>Data Cleaning</title>
          <p>The data cleaning process involved selection, cleaning, filtering, and preparation. After obtaining medical text reports from the hospital, we investigated the data and excluded patient data that were incomplete or duplicate in order to keep only complete patient data. Patients aged between 45 and 79 years were considered for 2 reasons. First, many previous studies focused on data from patients aged between 45 and 79 years [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref22">22</xref>]. Second, in the Knee Osteoarthritis Initiative datasets [<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref16">16</xref>], which are considered some of the standard resources to study the progression of OA and develop innovative approaches to treat this chronic disease, the focus is on patients aged 45-79 years.</p>
        </sec>
        <sec>
          <title>Text Analysis and Processing</title>
          <p>Data preprocessing was conducted. Punctuations were removed, and the texts were filtered to include only Thai alphabets and numbers. Moreover, word segmentation, tokenization, and correction of misspelled words were performed to ensure data quality. Next, we incorporated word embedding with the WOMAC and concepts. Each word was initially transformed into its embedding vector, with positional encoding and attention applied to assign weight to the words in each sentence. This yielded the final sentence vector.</p>
        </sec>
        <sec>
          <title>Experiment Plan</title>
          <p>We planned to use and experiment with 3 classification models, including CNN, BiLSTM, and GRU models, along with NLP components, which have been discussed in detail in the next section.</p>
        </sec>
      </sec>
      <sec>
        <title>NLP and Deep Learning Techniques</title>
        <p>In this section, we describe the NLP components used in our research, namely, word segmentation and Word2Vec, as well as deep learning models.</p>
        <sec>
          <title>Word Segmentation</title>
          <p>Since Thai sentences can be verbose and lead to ambiguity, word segmentation was considered before conducting further analysis of the text. The 3 most prevalent word segmentation approaches for Thai sentences are presented below.</p>
        </sec>
        <sec>
          <title>Word_Tokenize</title>
          <p>Word_Tokenize [<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref33">33</xref>] is one of the functions in the PyThaiNLP library. Word_Tokenize uses the maximal matching algorithm, which involves dictionary-based maximum matching and uses the Thai character cluster. Results demonstrated that it is the fastest word tokenizer on the BEST 2010 [<xref ref-type="bibr" rid="ref34">34</xref>] benchmark, with 71.18% accuracy [<xref ref-type="bibr" rid="ref35">35</xref>,<xref ref-type="bibr" rid="ref36">36</xref>].</p>
        </sec>
        <sec>
          <title>DeepCut</title>
          <p>DeepCut [<xref ref-type="bibr" rid="ref37">37</xref>] is a technique built on a CNN that uses characters and character types as features. The goal of the model is to predict whether a character is the beginning of a word before word segmentation can be processed. The BEST 2010 corpus [<xref ref-type="bibr" rid="ref34">34</xref>] was used for training when building the model. Results showed that the model had 97.8% precision, 98.5% recall, and 98.1% <italic>F</italic><sub>1</sub>-score [<xref ref-type="bibr" rid="ref32">32</xref>].</p>
        </sec>
        <sec>
          <title>AttaCut</title>
          <p>Inspired by DeepCut, AttaCut [<xref ref-type="bibr" rid="ref38">38</xref>] was designed to speed up the segmentation process. It achieves this goal by disabling neurons in DeepCut’s layers, and at the end, its speed performance increased by 6 times. Results on the in-domain dataset revealed 89%-91% <italic>F</italic><sub>1</sub>-scores, and those on the out-domain dataset revealed 63%-81% <italic>F</italic><sub>1</sub>-scores [<xref ref-type="bibr" rid="ref32">32</xref>].</p>
        </sec>
        <sec>
          <title>Word2Vec</title>
          <p>The well-known word embedding or word vector is an approach for representing documents and words. It is defined as a numeric vector input that allows words with similar meanings to have the same representation. Word2Vec is one of the models used to transform words and represent them as distributed vectors. It is commonly used as an initial step in predictive models for tasks related to semantics and information retrieval [<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref39">39</xref>,<xref ref-type="bibr" rid="ref40">40</xref>]. A neural network is then used to generate word embeddings as weights (by considering relationships between words and their surrounding context). Word2Vec generates word vectors using 2 main algorithms: continuous bag of words (CBOW) and skip-gram [<xref ref-type="bibr" rid="ref41">41</xref>,<xref ref-type="bibr" rid="ref42">42</xref>]. The CBOW and skip-gram architectures are shown in <xref rid="figure2" ref-type="fig">Figures 2</xref> and <xref rid="figure3" ref-type="fig">3</xref>, respectively.</p>
          <fig id="figure2" position="float">
            <label>Figure 2</label>
            <caption>
              <p>Continuous bag of words model architecture: one of the algorithms inside Word2Vec to produce the word vector of the current word by considering surrounding words.</p>
            </caption>
            <graphic xlink:href="formative_v9i1e64536_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <fig id="figure3" position="float">
            <label>Figure 3</label>
            <caption>
              <p>Skip-gram model architecture: one of the algorithms inside Word2Vec to produce word vectors for the surrounding words by considering the current word.</p>
            </caption>
            <graphic xlink:href="formative_v9i1e64536_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <p>Skip-gram is slower than CBOW in training and has lower performance on frequent words, but it is better suited for rare words or phrases. Since our data included medical terminology, which can contain rare words and phrases, we focused on Word2Vec by using skip-gram.</p>
          <p>The skip-gram model generates word-context pairs (<italic>w</italic>, <italic>c</italic>) from text data as input for a neural network.</p>
          <disp-formula>(<italic>w</italic>, <italic>c</italic>) = (<italic>w</italic><sub>1</sub>, <italic>c</italic><sub>1</sub>), (<italic>w</italic><sub>2</sub>, <italic>c</italic><sub>2</sub>), … , (<italic>w<sub>n</sub></italic>, <italic>c<sub>n</sub></italic>) <bold>(1)</bold></disp-formula>
          <disp-formula><italic>w<sub>i</sub></italic> ∈ <italic>V<sub>w</sub></italic>, <italic>c<sub>i</sub></italic> ∈ <italic>V<sub>c</sub></italic></disp-formula>
          <p>where <italic>V<sub>w</sub></italic> is the input vector (center word) and <italic>V<sub>c</sub></italic> is the output vector (context word) that represent the word and context vocabularies with sizes <italic>S<sub>w</sub></italic> and <italic>S<sub>c</sub></italic>, respectively. The goal of the skip-gram model is to learn the distribution of each word (based on context) by maximizing the probability of context words, as shown below.</p>
          <disp-formula>
            <graphic xlink:href="formative_v9i1e64536_fig10.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </disp-formula>
          <p>where <italic>E</italic> and <italic>F</italic> refer to the parameter matrices of (<italic>S<sub>w</sub></italic>×<italic>d</italic>) and (<italic>S<sub>c</sub></italic>×<italic>d</italic>), respectively, and <italic>d</italic> refers to the dimensionality of the embedding vector space.</p>
        </sec>
        <sec>
          <title>Deep Learning Models</title>
          <p>As mentioned in the literature review section, deep learning methods have demonstrated strong performance in image classification and speech recognition. In recent years, they have also been widely adopted in NLP and have achieved good performance as well [<xref ref-type="bibr" rid="ref26">26</xref>]. In NLP tasks, the deep learning models widely used include GRU, LSTM, and CNN [<xref ref-type="bibr" rid="ref11">11</xref>]. Therefore, we focused on these deep learning models for the detection of knee OA.</p>
        </sec>
        <sec>
          <title>GRU Model</title>
          <p>GRU is a type of RNN specifically designed to address the vanishing gradient problem when capturing long-term dependencies in RNNs. It is also very efficient in processing sequential or time series data in NLP tasks. The structure of GRU consists of an update gate and a reset gate [<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref43">43</xref>]. GRU uses gating mechanisms to control which information should be passed to the output or remain in the network. GRU model architecture is illustrated in <xref rid="figure4" ref-type="fig">Figure 4</xref>.</p>
          <fig id="figure4" position="float">
            <label>Figure 4</label>
            <caption>
              <p>Gated recurrent unit model architecture: one of the recurrent neural network types. Its structure consists of input state (x<sub>t</sub>), hidden state (h<sub>t</sub>), output state (o<sub>t</sub>), and update gate (z<sub>t</sub>).</p>
            </caption>
            <graphic xlink:href="formative_v9i1e64536_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <p>The GRU structure is illustrated in <xref rid="figure4" ref-type="fig">Figure 4</xref>. In the equations listed below, the input <italic>x<sub>t</sub></italic> (the sequence vector of the current word at time <italic>t</italic>) is the input for the hidden state <italic>h<sub>t</sub></italic>.</p>
          <disp-formula><italic>r<sub>t</sub></italic> = <italic>σ</italic>(<italic>W<sub>r</sub>x<sub>r</sub></italic> + <italic>U<sub>r</sub>h<sub>t</sub></italic><sub>-1</sub> + <italic>b<sub>r</sub></italic>) <bold>(3)</bold></disp-formula>
          <disp-formula><italic>z<sub>t</sub></italic> = <italic>σ</italic>(<italic>W<sub>z</sub>x<sub>t</sub></italic> + <italic>U<sub>z</sub>h<sub>t</sub></italic><sub>-1</sub> + <italic>b<sub>z</sub></italic>) <bold>(4)</bold></disp-formula>
          <disp-formula>
            <graphic xlink:href="formative_v9i1e64536_fig11.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </disp-formula>
          <disp-formula>
            <graphic xlink:href="formative_v9i1e64536_fig12.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </disp-formula>
          <p>where the update gate and the reset gate are represented by <italic>z<sub>t</sub></italic> and <italic>r<sub>t</sub></italic>, respectively.The previous state is represented by <italic>h<sub>t</sub></italic><sub>-1</sub>, while the candidate state is represented by <inline-graphic xlink:href="formative_v9i1e64536_fig13.png" xlink:type="simple" mimetype="image"/> at time <italic>t</italic>. The sigmoid and hyperbolic tangent functions are represented by σ and <italic>tanh</italic>, respectively. <italic>b<sub>z</sub></italic>, <italic>b<sub>h</sub></italic>, and <italic>b<sub>r</sub></italic> are all biases, whereas the operator <inline-graphic xlink:href="formative_v9i1e64536_fig16.png" xlink:type="simple" mimetype="image"/> is element-wise multiplication.</p>
        </sec>
        <sec>
          <title>LSTM Model</title>
          <p>LSTM is considered among the most widely adopted RNN types for sequential data-related tasks. It is designed to solve the problem of the traditional RNN model in capturing long-term dependencies, similar to GRU. The structure is shown in <xref rid="figure5" ref-type="fig">Figure 5</xref>.</p>
          <fig id="figure5" position="float">
            <label>Figure 5</label>
            <caption>
              <p>Long short-term memory model architecture: one of the recurrent neural network types. Its structure consists of input state (x<sub>t</sub>), hidden state (h<sub>t</sub>), output state (o<sub>t</sub>), forget gate (z<sub>t</sub>), and memory cell (c<sub>t</sub>).</p>
            </caption>
            <graphic xlink:href="formative_v9i1e64536_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <p>Often, RNNs face the vanishing gradient problem when trying to capture long-term dependencies. Consequently, LSTM is designed to solve this issue by using some internal gates [<xref ref-type="bibr" rid="ref44">44</xref>]. The LSTM architecture consists of 3 gates (input gate, output gate, and forget gate) along with a memory cell [<xref ref-type="bibr" rid="ref45">45</xref>]. The goal of the memory cell is to store information over long sequences by capturing previous words (<italic>c<sub>t</sub></italic><sub>-1</sub>, <italic>h<sub>t</sub></italic><sub>-1</sub>), with some being dropped (forgotten) by the forget gate, which regulates the flow of information in (<italic>x</italic>) and out (<italic>h</italic>) of the cell by the 3 gates. The input gate, forget gate, output gate, and cell output can be illustrated as follows:</p>
          <disp-formula><italic>f<sub>t</sub></italic> = <italic>σ</italic>(<italic>W</italic><sup>(</sup><italic><sup>f</sup></italic><sup>)</sup><italic>x<sub>t</sub></italic> + <italic>U</italic><sup>(</sup><italic><sup>f</sup></italic><sup>)</sup><italic>h<sub>t</sub></italic><sub>-1</sub> + <italic>b</italic><sup>(</sup><italic><sup>f</sup></italic><sup>)</sup>) <bold>(7)</bold></disp-formula>
          <disp-formula><italic>i<sub>t</sub></italic> = <italic>σ</italic>(<italic>W</italic><sup>(</sup><italic><sup>i</sup></italic><sup>)</sup><italic>x<sub>t</sub></italic> + <italic>U</italic><sup>(</sup><italic><sup>i</sup></italic><sup>)</sup><italic>h<sub>t</sub></italic><sub>-1</sub> + <italic>b</italic><sup>(</sup><italic><sup>i</sup></italic><sup>)</sup>) <bold>(8)</bold></disp-formula>
          <disp-formula><italic>o<sub>t</sub></italic> = <italic>σ</italic>(<italic>W</italic><sup>(</sup><italic><sup>o</sup></italic><sup>)</sup><italic>x<sub>t</sub></italic> + <italic>U</italic><sup>(</sup><italic><sup>o</sup></italic><sup>)</sup><italic>h<sub>t</sub></italic><sub>-1</sub> + <italic>b</italic><sup>(</sup><italic><sup>o</sup></italic><sup>)</sup>) <bold>(9)</bold></disp-formula>
          <disp-formula><italic>c<sub>t</sub></italic> = <italic>f<sub>t</sub></italic> <inline-graphic xlink:href="formative_v9i1e64536_fig17.png" xlink:type="simple" mimetype="image"/> <italic>c<sub>t</sub></italic><sub>-1</sub> + <italic>i<sub>t</sub></italic> <inline-graphic xlink:href="formative_v9i1e64536_fig17.png" xlink:type="simple" mimetype="image"/> <italic>tanh</italic> (<italic>W</italic><sup>(</sup><italic><sup>c</sup></italic><sup>)</sup><italic>x<sub>t</sub></italic> + <italic>U</italic><sup>(</sup><italic><sup>c</sup></italic><sup>)</sup><italic>h<sub>t</sub></italic><sub>-1</sub> + <italic>b</italic><sup>(</sup><italic><sup>c</sup></italic><sup>)</sup>) <bold>(10)</bold></disp-formula>
          <disp-formula><italic>h<sub>t</sub></italic> = <italic>o<sub>t</sub></italic> <inline-graphic xlink:href="formative_v9i1e64536_fig17.png" xlink:type="simple" mimetype="image"/> <italic>tanh</italic> (<italic>c<sub>t</sub></italic>) <bold>(11)</bold></disp-formula>
          <p>where the input is represented by <italic>x<sub>t</sub></italic> ∈ <italic>R<sup>d</sup></italic> at time step <italic>t</italic>, while <italic>d</italic> is the feature dimension for each word. The element-wise sigmoid function and the element-wise dot product are represented by σ and <inline-graphic xlink:href="formative_v9i1e64536_fig17.png" xlink:type="simple" mimetype="image"/>, respectively. The memory cell represented by <italic>c<sub>t</sub></italic> aims to solve the problem of vanishing gradients, which facilitates the learning of long-term dependencies. The forget gate, represented by <italic>f<sub>t</sub></italic>, is used to decide what information to store in the memory cell, whereas the input gate and output gate, represented by <italic>i<sub>t</sub></italic> and <italic>o<sub>t</sub></italic>, respectively, aim to control the input and output of the memory cell, respectively. Finally, the output of LSTM is represented by <italic>h<sub>t</sub></italic>.</p>
          <p>However, LSTM still has some limitations, since it can only read in 1 direction (forward direction). In this case, the results may show low classification performance because some contextual information from future words may be ignored by considering just 1 forward direction. Thus, Alharbi et al [<xref ref-type="bibr" rid="ref14">14</xref>] proposed a solution to address the limitations of LSTM, which is called BiLSTM. BiLSTM can capture long-term dependencies from both the past and the future of a sequence (2 directions), which is important for understanding the full context of the sequence. As an extension of LSTM, the BiLSTM network processes both the forward and backward directions of a sequence to capture dependencies [<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref46">46</xref>,<xref ref-type="bibr" rid="ref47">47</xref>]. The BiLSTM model architecture is illustrated in <xref rid="figure6" ref-type="fig">Figure 6</xref>.</p>
          <fig id="figure6" position="float">
            <label>Figure 6</label>
            <caption>
              <p>Bidirectional long short-term memory model architecture: one of the recurrent neural network types. Its structure consists of 2 long short-term memory parts: one processes input sequences in the forward direction, and the other processes input sequences in the backward direction. LSTM: long short-term memory.</p>
            </caption>
            <graphic xlink:href="formative_v9i1e64536_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>CNN Model</title>
          <p>CNNs are generally developed for image classification. Currently, CNNs have shown strong performance in various NLP tasks, particularly in text classification. They are often considered appropriate for short text classification tasks [<xref ref-type="bibr" rid="ref47">47</xref>-<xref ref-type="bibr" rid="ref49">49</xref>]. A CNN model typically consists of several layers, which can be separated into 3 groups. First, the convolutional layer transforms the text data to obtain a feature map to detect the presence of a specific feature. Second, the pooling layer follows the convolutional layer, and it is designed to reduce the input data while preserving important information further. Third, the final stage of the CNN is the fully connected layer, which is responsible for using the features extracted in the previous layers for classification. The CNN architecture is illustrated in <xref rid="figure7" ref-type="fig">Figure 7</xref>.</p>
          <p>Embedding layer: Let <italic>w<sub>i</sub></italic> ∈ <italic>R<sup>m</sup></italic> be a word vector with m-dimensional representation associated with a word in a sentence. A sentence of length <italic>n</italic> is represented as follows:</p>
          <disp-formula><italic>w</italic><sub>1:</sub><italic><sub>n</sub></italic> = <italic>w</italic><sub>1</sub><inline-graphic xlink:href="formative_v9i1e64536_fig15.png" xlink:type="simple" mimetype="image"/><italic>w</italic><sub>2</sub><inline-graphic xlink:href="formative_v9i1e64536_fig15.png" xlink:type="simple" mimetype="image"/> ...<inline-graphic xlink:href="formative_v9i1e64536_fig15.png" xlink:type="simple" mimetype="image"/><italic>w</italic><sub>n</sub> <bold>(12)</bold></disp-formula>
          <p>where <inline-graphic xlink:href="formative_v9i1e64536_fig15.png" xlink:type="simple" mimetype="image"/> denotes the concatenation operator.</p>
          <p>Convolutional layer: The filter is represented by <italic>x</italic> ∈ <italic>R<sup>h</sup></italic><sup>×</sup><italic><sup>m</sup></italic>, where <italic>h</italic> is the number of words in a window and <italic>m</italic> is the dimension of a word vector. A feature map c = [<italic>c</italic><sub>1</sub>, <italic>c</italic><sub>2</sub>, …, <italic>c<sub>n</sub></italic><sub>-</sub><italic><sub>h</sub></italic><sub>+1</sub>] can be generated by:</p>
          <disp-formula><italic>c<sub>i</sub></italic> = <italic>f</italic>(<italic>x</italic>∙<italic>w<sub>i</sub></italic><sub>:</sub><italic><sub>i</sub></italic><sub>+</sub><italic><sub>h</sub></italic><sub>-1</sub> + <italic>b</italic>) <bold>(13)</bold></disp-formula>
          <p>where <italic>b</italic> ∈ <italic>R</italic> denotes a bias term and <italic>f</italic> is a nonlinear function.</p>
          <p>Pooling layer: The pooling operation was applied for the respective filter to select the most important feature from each feature map. We used the max pooling operation to take the maximum value in the feature map.</p>
          <disp-formula><inline-graphic xlink:href="formative_v9i1e64536_fig18.png" xlink:type="simple" mimetype="image"/> = <italic>max</italic> {<italic>c</italic>} <bold>(14)</bold></disp-formula>
          <p>Fully connected layer: The selected features from the pooling layer are flattened into a single vector. It is passed to a fully connected layer to assign a probability for a specific label.</p>
          <disp-formula>
            <graphic xlink:href="formative_v9i1e64536_fig14.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </disp-formula>
          <fig id="figure7" position="float">
            <label>Figure 7</label>
            <caption>
              <p>Convolutional neural network model architecture. Its structure consists of an embedding layer (input vectors), a 1D convolutional layer to extract local patterns, a pooling layer to select the most important features from the convolutional layer, a fully connected layer to prepare for the classification, and an output layer to calculate the output from the fully connected layer to the final binary classification result.</p>
            </caption>
            <graphic xlink:href="formative_v9i1e64536_fig7.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
      </sec>
      <sec>
        <title>Algorithmic Details</title>
        <p>According to the deep learning models mentioned in the previous subsections, the key procedures can be described as follows: (1) import a dataset of reported symptoms of diseases (that are in Excel files) to our program; (2) perform data cleaning in sentences from Excel files; (3) create a vocabulary set with indexes from sentences to transform the sentences into vectors using a tokenizer; (4) create a word embedding from unique vocabulary for pretraining by using Word2Vec (skip-gram approach); (5) identify the longest vector with the maximum number of words in a sentence and adjust all the vectors to the same length as the longest (since the word count of words in a sentence can vary, the size of each vector can be different); (6) pad such vectors with fewer words than the maximum one with zeroes; (7) separate training data, validation data, and test data; (8) build the proposed models with BiLSTM, GRU, and CNN; (9) train the models according to the range given by the hyperparameters; and (10) evaluate the models.</p>
        <p>We present the pseudocode of our methods, in which algorithm 1 was designed for BiLSTM and GRU (<xref ref-type="boxed-text" rid="box1">Textbox 1</xref>) and algorithm 2 was designed for CNN (<xref ref-type="boxed-text" rid="box2">Textbox 2</xref>). For CNN, we used a 1D convolution layer, and the feature map was selected by max pooling in convolutional features with kernel sizes of 4, 6, and 8. We used dropout and rectified linear unit (ReLU) as activation functions. Lastly, we used the Softmax activation function in a fully connected layer.</p>
        <boxed-text id="box1" position="float">
          <title>Pseudocodes of bidirectional long short-term memory and gated recurrent unit for knee osteoarthritis predictions.</title>
          <p>Algorithm 1: Pseudocodes of bidirectional long short-term memory (BiLSTM) and gated recurrent unit (GRU)</p>
          <p>Input: A collection of medical symptom reports for diseases in Excel files (Let <italic>D</italic> be one report and <italic>S(i)</italic> be an <italic>i<sup>th</sup></italic> sentence: <italic>D</italic> = {<italic>S</italic>(1), <italic>S</italic>(2), <italic>S</italic>(3), … , <italic>S</italic>(<italic>i</italic>)}, <italic>i</italic> ∈ <italic>N</italic>)</p>
          <p>Output: Prediction results/performance of BiLSTM and GRU</p>
          <p>1: Import the reports of knee osteoarthritis data from Excel files into our program</p>
          <p>2: while i in (1, n) do;</p>
          <p>Select characters (that are Thai alphabets only) in the sentences</p>
          <p>Tokenize by Word_Tokenize, DeepCut, AttaCut</p>
          <p>Remove Thai stop words in sentences</p>
          <p>Edit missing words with WOMAC (Western Ontario and McMaster Universities Osteoarthritis Index)-based processing approach (cutoff = 0.85)</p>
          <p>end while</p>
          <p>3: Create a set of vocabulary along with indexes by tokenizer from TensorFlow</p>
          <p>4. Generate word embedding by Word2Vec (skip-gram approach)</p>
          <p>vector_size = 128, epochs = 1000</p>
          <p>5: Identify the max-length vector and adjust all vectors accordingly</p>
          <p>6: Pad all smaller vectors with zeroes</p>
          <p>7: Separate training, validation, and test datasets [training dataset: 80%, validation dataset: 10%, and test dataset: 10%]</p>
          <p>8: Build the BiLSTM or GRU model</p>
          <p>Set vocab_size, drop_out, BatchNomalization, and activation function parameters</p>
          <p>9: while i &#60;= epochs; // e = epoch numbers</p>
          <p>execute the attention module;</p>
          <p>import input vector;</p>
          <p>Dropout; // (0.5)</p>
          <p>BatchNormalization;</p>
          <p>Update/optimize parameters with Adam optimization to minimize the loss function in the model</p>
          <p>Verify the BiLSTM or GRU model using</p>
          <p>validation data;</p>
          <p>end while</p>
          <p>10: Evaluate model by the test data;</p>
          <p>Return the performance of BiLSTM and GRU</p>
        </boxed-text>
        <boxed-text id="box2" position="float">
          <title>Pseudocodes of convolutional neural network for knee osteoarthritis predictions.</title>
          <p>Algorithm 2: Pseudocodes of convolutional neural network (CNN)</p>
          <p>Input: A collection of medical symptom reports for diseases in Excel files (Let <italic>D</italic> be one report and <italic>S(i)</italic> be an <italic>i<sup>th</sup></italic> sentence: <italic>D</italic> = {<italic>S</italic>(1), <italic>S</italic>(2), <italic>S</italic>(3), … , <italic>S</italic>(<italic>i</italic>)}, <italic>i</italic> ∈ <italic>N</italic>)</p>
          <p>Output: The result of the prediction from CNN</p>
          <p>1: Import the report of knee osteoarthritis data</p>
          <p>2: while i in (1, n) do;</p>
          <p>Select characters (that are Thai alphabets only) in the sentences</p>
          <p>Tokenize by Word_Tokenize, DeepCut, AttaCut</p>
          <p>Remove Thai stop words in sentences</p>
          <p>Edit missing words with the WOMAC (Western Ontario and McMaster Universities Osteoarthritis Index)-based processing approach (cutoff = 0.85)</p>
          <p>end while</p>
          <p>3: Create a set of vocabulary along with indexes by tokenizer from TensorFlow</p>
          <p>4. Generate word embedding by Word2Vec (skip-gram approach)</p>
          <p>vector_size = 128, epochs = 1000</p>
          <p>5: Identify the max-length vector and adjust all vectors accordingly</p>
          <p>6: Pad all smaller vectors with zeroes</p>
          <p>7: Separate training, validation, and test datasets [training dataset: 80%, validation dataset: 10%, and test dataset: 10%]</p>
          <p>8: Build the CNN model</p>
          <p>Set vocab_size, drop_out, BatchNormalization, and activation function parameters</p>
          <p>9: while i &#60;= epochs; // e = epoch numbers</p>
          <p>execute the attention module;</p>
          <p>import input vector;</p>
          <p>convolutional layer; // kernel_size = 4, 6, 8</p>
          <p>Dropout; // (0.5)</p>
          <p>Max pooling;</p>
          <p>Fully connected layer;</p>
          <p>Update/optimize parameters with Adam optimization to minimize the loss function in the model</p>
          <p>Verify the CNN model using validation data;</p>
          <p>end while</p>
          <p>10: Evaluate model by using test data;</p>
          <p>Return the performance of the CNN</p>
        </boxed-text>
      </sec>
      <sec>
        <title>Evaluation Metrics</title>
        <p>We evaluated our models by using 5 metrics: accuracy, sensitivity (or recall), specificity, precision, and <italic>F</italic><sub>1</sub>-score. These parameters are defined as follows:</p>
        <disp-formula>    Accuracy = (TP + TN) / (TP + FN + FP + TN) <bold>(16)</bold></disp-formula>
        <disp-formula> Sensitivity/recall = TP / (TP + FN) <bold>(17)</bold></disp-formula>
        <disp-formula>  Specificity = TN / (TN + FP) <bold>(18)</bold></disp-formula>
        <disp-formula>  Precision = TP / (TP + FP) <bold>(19)</bold></disp-formula>
        <disp-formula><italic>F</italic><sub>1</sub>-score = (2 × [precision × recall]) / (precision + recall) <bold>(20)</bold></disp-formula>
        <p>where true positive (TP) is the total number of patients classified and tagged in the dataset as OA, true negative (TN) is the total number of patients classified and tagged in the dataset as non-OA, false negative (FN) is the total number of patients classified as non-OA but tagged in the dataset as OA, and false positive (FP) is the total number of patients classified as OA but tagged in the dataset as non-OA. Accuracy is the ratio of true results of both OA and non-OA among the total number of patients examined. Sensitivity is the proportion of patients classified as OA by the model among all OA-tagged data. Specificity is the proportion of patients classified as non-OA by the model among all non-OA–tagged data. Precision is the percentage of patients classified as OA by the model, and <italic>F</italic><sub>1</sub>-score is the average of recall and precision.</p>
        <p>In summary, our contributions are as follows. First, our approach proposes to use clinicians’ or doctors’ notes (instead of x-ray or MRI images and statistical laboratory results), which contain meaningful messages from patients who have the same symptoms, to help in diagnosis. Second, we used the power of text classification through NLP in diagnosis, which can be used in complement with other diagnostic methods (such as using medical images or biospecimen laboratory results) to enhance accuracy and achieve a timely diagnosis. Third, we introduced the high potential of textual analysis based on doctors’ notes of symptoms to the field of knee OA as an alternative, which is new, and combined it with the WOMAC (dictionary/key terms for knee OA). Fourth, although our approach is performed on Thai datasets, the methodology can be adapted or repeated for diagnosing knee OA in other languages. Fifth, we combined and compared the text classification methods with deep learning models, including GRU, BiLSTM, and CNN, and at the end, we were able to identify the best combination in diagnosing knee OA in our approach. Sixth, to further tune our approach for better performance, we fine-tuned our method with our technique in adjusting misleading or wrong words by using real experts and physicians, as well as the Thai-verified version of the WOMAC.</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Records</title>
        <p>To ensure data quality, we removed 888 (incomplete) records from 6737 knee OA records, leaving 5849 records containing complete patient reports. The dataset was separated into training, validation, and test sets comprising 4788, 598, and 463 records, respectively. We used the GRU, BiLSTM, and CNN models for knee OA prediction.</p>
      </sec>
      <sec>
        <title>Hyperparameters of GRU, BiLSTM, and CNN</title>
        <p>The hyperparameters were adjusted for each model used in the experiments, including the quantity of neurons in each layer, the choice of the activation function, the optimization method, and the regularization techniques. In selecting hyperparameters for our experiments, we referred to relevant research studies in the literature that demonstrated strong performance [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref44">44</xref>,<xref ref-type="bibr" rid="ref48">48</xref>,<xref ref-type="bibr" rid="ref50">50</xref>]. The hyperparameters of each model are shown in <xref ref-type="table" rid="table1">Table 1</xref>.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Hyperparameters of gated recurrent unit, bidirectional long short-term memory, and convolutional neural network.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="211"/>
            <col width="285"/>
            <col width="279"/>
            <col width="225"/>
            <thead>
              <tr valign="top">
                <td>Hyperparameter<sup>a</sup></td>
                <td>GRU<sup>b</sup></td>
                <td>BiLSTM<sup>c</sup></td>
                <td>CNN<sup>d</sup></td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Filter</td>
                <td>—<sup>e</sup></td>
                <td>—</td>
                <td>32</td>
              </tr>
              <tr valign="top">
                <td>Kernel</td>
                <td>—</td>
                <td>—</td>
                <td>4, 6, 8</td>
              </tr>
              <tr valign="top">
                <td>Activation function (last layer)</td>
                <td>ReLU<sup>f</sup> (Softmax)</td>
                <td>ReLU (Softmax)</td>
                <td>ReLU (Sigmoid)</td>
              </tr>
              <tr valign="top">
                <td>Dropout</td>
                <td>0.5</td>
                <td>0.5</td>
                <td>0.5</td>
              </tr>
              <tr valign="top">
                <td>Epoch</td>
                <td>50</td>
                <td>50</td>
                <td>50</td>
              </tr>
              <tr valign="top">
                <td>Batch size</td>
                <td>128</td>
                <td>128</td>
                <td>128</td>
              </tr>
              <tr valign="top">
                <td>Loss function</td>
                <td>categorical_crossentropy</td>
                <td>categorical_crossentropy</td>
                <td>binary_crossentropy</td>
              </tr>
              <tr valign="top">
                <td>Optimizer/learning rate</td>
                <td>Adam/0.0001</td>
                <td>Adam/0.0001</td>
                <td>Adam/0.0001</td>
              </tr>
              <tr valign="top">
                <td>Regularization</td>
                <td>Batch normalization</td>
                <td>Batch normalization</td>
                <td> Batch normalization</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>BiLSTM: bidirectional long short-term memory.</p>
            </fn>
            <fn id="table1fn2">
              <p><sup>b</sup>GRU: gated recurrent unit.</p>
            </fn>
            <fn id="table1fn3">
              <p><sup>c</sup>CNN: convolutional neural network.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <p>For GRU and BiLSTM, the weights were passed to the output layer with a Softmax function to diagnose knee OA. For CNN, the weights were passed to the output layer with a sigmoid function to diagnose knee OA. For all models, the hidden layer was trained and regularized using ReLU and batch normalization, respectively. To reduce the overfitting problem, we applied dropout [<xref ref-type="bibr" rid="ref14">14</xref>], with a probability of 0.5 after each model while training with the Adam optimizer [<xref ref-type="bibr" rid="ref10">10</xref>], using a learning rate of 0.0001 to optimize the model parameters. The experiments have two parts: (1) experiments that use the baseline approach and (2) experiments that use our WOMAC-based processing approach. In <xref ref-type="table" rid="table2">Tables 2</xref> and <xref ref-type="table" rid="table3">3</xref>, we present the experimental results of the models “before” processing the missing or edited words in sentences by using our WOMAC-based processing approach. The best-performing model was the BiLSTM model that used the tokenization of the AttaCut algorithm, with the following metrics: area under the curve (AUC), 0.85; accuracy, 0.87; precision, 0.85; sensitivity, 0.95; specificity, 0.76; and <italic>F</italic><sub>1</sub>-score, 0.90.</p>
        <p>In <xref ref-type="table" rid="table4">Tables 4</xref> and <xref ref-type="table" rid="table5">5</xref>, we present the experimental results of the models “after” processing the missing or edited words in sentences by using our WOMAC-based processing approach. The best-performing model was the BiLSTM model that used the tokenization of the AttaCut algorithm, with the following metrics: AUC, 0.91; accuracy, 0.91; precision, 0.91; sensitivity, 0.94; specificity, 0.87; and <italic>F</italic><sub>1</sub>-score, 0.93.</p>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Results of the baseline approach (n=463) in terms of true positive, false positive, false negative, true negative, and runtime for each model.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="0"/>
            <col width="220"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <thead>
              <tr valign="top">
                <td colspan="3">Model</td>
                <td>TP<sup>a</sup>, n (%)</td>
                <td>FP<sup>b</sup>, n (%)</td>
                <td>FN<sup>c</sup>, n (%)</td>
                <td>TN<sup>d</sup>, n (%)</td>
                <td>Runtime (s)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="8">
                  <bold>Word_Tokenize (n=463)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td/>
                <td colspan="2">BiLSTM<sup>e</sup></td>
                <td>248 (53.6)</td>
                <td>36 (7.8)</td>
                <td>26 (5.6)</td>
                <td>153 (33.0)</td>
                <td>389.01</td>
              </tr>
              <tr valign="top">
                <td/>
                <td colspan="2">GRU<sup>f</sup></td>
                <td>249 (53.8)</td>
                <td>40 (8.6)</td>
                <td>25 (5.4)</td>
                <td>149 (32.2)</td>
                <td>303.88</td>
              </tr>
              <tr valign="top">
                <td/>
                <td colspan="2">CNN<sup>g</sup></td>
                <td>252 (54.4)</td>
                <td>43 (9.3)</td>
                <td>22 (4.8)</td>
                <td>146 (31.5)</td>
                <td>138.72</td>
              </tr>
              <tr valign="top">
                <td colspan="8">
                  <bold>DeepCut (n=463)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td/>
                <td colspan="2">BiLSTM</td>
                <td>258 (55.7)</td>
                <td>47 (10.2)</td>
                <td>16 (3.5)</td>
                <td>142 (30.7)</td>
                <td>476.81</td>
              </tr>
              <tr valign="top">
                <td/>
                <td colspan="2">GRU</td>
                <td>249 (53.8)</td>
                <td>46 (9.9)</td>
                <td>25 (5.4)</td>
                <td>143 (30.9)</td>
                <td>411.99</td>
              </tr>
              <tr valign="top">
                <td/>
                <td colspan="2">CNN</td>
                <td>258 (55.7)</td>
                <td>49 (10.6)</td>
                <td>16 (3.5)</td>
                <td>140 (30.2)</td>
                <td>152.16</td>
              </tr>
              <tr valign="top">
                <td colspan="8">
                  <bold>AttaCut (n=463)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td/>
                <td colspan="2">BiLSTM</td>
                <td>260 (56.2)</td>
                <td>46 (9.9)</td>
                <td>14 (3.0)</td>
                <td>143 (30.9)</td>
                <td>467.10</td>
              </tr>
              <tr valign="top">
                <td/>
                <td colspan="2">GRU</td>
                <td>251 (54.2)</td>
                <td>40 (8.6)</td>
                <td>23 (5.0)</td>
                <td>149 (32.2)</td>
                <td>387.65</td>
              </tr>
              <tr valign="top">
                <td/>
                <td colspan="2">CNN</td>
                <td>248 (53.6)</td>
                <td>35 (7.6)</td>
                <td>26 (5.6)</td>
                <td>154 (33.3)</td>
                <td>146.02</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>TP: true positive.</p>
            </fn>
            <fn id="table2fn2">
              <p><sup>b</sup>FP: false positive.</p>
            </fn>
            <fn id="table2fn3">
              <p><sup>c</sup>FN: false negative.</p>
            </fn>
            <fn id="table2fn4">
              <p><sup>d</sup>TN: true negative.</p>
            </fn>
            <fn id="table2fn5">
              <p><sup>e</sup>BiLSTM: bidirectional long short-term memory.</p>
            </fn>
            <fn id="table2fn6">
              <p><sup>f</sup>GRU: gated recurrent unit.</p>
            </fn>
            <fn id="table2fn7">
              <p><sup>g</sup>CNN: convolutional neural network.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Results of the baseline approach (n=463) in terms of accuracy, precision, sensitivity, specificity, and F1-score for each model.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="220"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <thead>
              <tr valign="top">
                <td colspan="2">Model</td>
                <td>Accuracy (%)</td>
                <td>Precision (%)</td>
                <td>Sensitivity (%)</td>
                <td>Specificity (%)</td>
                <td><italic>F</italic><sub>1</sub>-score (%)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="7">
                  <bold>Word_Tokenize</bold>
                </td>
              </tr>
              <tr valign="top">
                <td/>
                <td>BiLSTM<sup>a</sup></td>
                <td>0.86</td>
                <td>0.85</td>
                <td>0.92</td>
                <td>0.77</td>
                <td>0.89</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>GRU<sup>b</sup></td>
                <td>0.86</td>
                <td>0.86</td>
                <td>0.91</td>
                <td>0.79</td>
                <td>0.89</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>CNN<sup>c</sup></td>
                <td>0.87</td>
                <td>0.87</td>
                <td>0.91</td>
                <td>0.91</td>
                <td>0.89</td>
              </tr>
              <tr valign="top">
                <td colspan="7">
                  <bold>DeepCut</bold>
                </td>
              </tr>
              <tr valign="top">
                <td/>
                <td>BiLSTM</td>
                <td>0.86</td>
                <td>0.85</td>
                <td>0.85</td>
                <td>0.94</td>
                <td>0.75</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>GRU</td>
                <td>0.85</td>
                <td>0.84</td>
                <td>0.83</td>
                <td>0.91</td>
                <td>0.84</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>CNN</td>
                <td>0.86</td>
                <td>0.85</td>
                <td>0.85</td>
                <td>0.94</td>
                <td>0.75</td>
              </tr>
              <tr valign="top">
                <td colspan="7">
                  <bold>AttaCut</bold>
                </td>
              </tr>
              <tr valign="top">
                <td/>
                <td>BiLSTM</td>
                <td>0.87</td>
                <td>0.85</td>
                <td>0.95</td>
                <td>0.76</td>
                <td>0.90</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>GRU</td>
                <td>0.86</td>
                <td>0.86</td>
                <td>0.92</td>
                <td>0.79</td>
                <td>0.89</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>CNN</td>
                <td>0.87</td>
                <td>0.88</td>
                <td>0.91</td>
                <td>0.81</td>
                <td>0.89</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table3fn1">
              <p><sup>a</sup>BiLSTM: bidirectional long short-term memory.</p>
            </fn>
            <fn id="table3fn2">
              <p><sup>b</sup>GRU: gated recurrent unit.</p>
            </fn>
            <fn id="table3fn3">
              <p><sup>c</sup>CNN: convolutional neural network.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table4">
          <label>Table 4</label>
          <caption>
            <p>Results of the Western Ontario and McMaster Universities Osteoarthritis Index–based processing approach (n=463) in terms of true positive, false positive, false negative, true negative, and runtime for each model.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="220"/>
            <col width="0"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <thead>
              <tr valign="top">
                <td colspan="3">Model</td>
                <td>TP<sup>a</sup>, n (%)</td>
                <td>FP<sup>b</sup>, n (%)</td>
                <td>FN<sup>c</sup>, n (%)</td>
                <td>TN<sup>d</sup>, n (%)</td>
                <td>Runtime (s)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="8">
                  <bold>Word_Tokenize (n=463)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td/>
                <td>BiLSTM<sup>e</sup></td>
                <td colspan="2">252 (54.4)</td>
                <td>24 (5.2)</td>
                <td>22 (4.8)</td>
                <td>165 (35.6)</td>
                <td>359.62</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>GRU<sup>f</sup></td>
                <td colspan="2">256 (55.3)</td>
                <td>36 (7.8)</td>
                <td>18 (3.9)</td>
                <td>153 (33.0)</td>
                <td>294.49</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>CNN<sup>g</sup></td>
                <td colspan="2">254 (54.9)</td>
                <td>25 (5.4)</td>
                <td>20 (4.3)</td>
                <td>164 (35.4)</td>
                <td>128.76</td>
              </tr>
              <tr valign="top">
                <td colspan="8">
                  <bold>DeepCut (n=463)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td/>
                <td>BiLSTM</td>
                <td colspan="2">251 (54.2)</td>
                <td>25 (5.4)</td>
                <td>23 (5.0)</td>
                <td>164 (35.4)</td>
                <td>429.84</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>GRU</td>
                <td colspan="2">250 (54.0)</td>
                <td>27 (5.8)</td>
                <td>24 (5.2)</td>
                <td>162 (35.0)</td>
                <td>367.58</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>CNN</td>
                <td colspan="2">258 (55.7)</td>
                <td>36 (7.8)</td>
                <td>16 (3.5)</td>
                <td>153 (33.0)</td>
                <td>129.69</td>
              </tr>
              <tr valign="top">
                <td colspan="8">
                  <bold>AttaCut (n=463)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td/>
                <td>BiLSTM</td>
                <td colspan="2">257 (55.5)</td>
                <td>24 (5.2)</td>
                <td>17 (3.7)</td>
                <td>165 (35.6)</td>
                <td>397.51</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>GRU</td>
                <td colspan="2">254 (54.9)</td>
                <td>31 (6.7)</td>
                <td>20 (4.3)</td>
                <td>158 (34.1)</td>
                <td>343.44</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>CNN</td>
                <td colspan="2">251 (54.2)</td>
                <td>22 (4.8)</td>
                <td>23 (5.0)</td>
                <td>167 (36.1)</td>
                <td>117.44</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table4fn1">
              <p><sup>a</sup>TP: true positive.</p>
            </fn>
            <fn id="table4fn2">
              <p><sup>b</sup>FP: false positive.</p>
            </fn>
            <fn id="table4fn3">
              <p><sup>c</sup>FN: false negative.</p>
            </fn>
            <fn id="table4fn4">
              <p><sup>d</sup>TN: true negative.</p>
            </fn>
            <fn id="table4fn5">
              <p><sup>e</sup>BiLSTM: bidirectional long short-term memory.</p>
            </fn>
            <fn id="table4fn6">
              <p><sup>f</sup>GRU: gated recurrent unit.</p>
            </fn>
            <fn id="table4fn7">
              <p><sup>g</sup>CNN: convolutional neural network.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table5">
          <label>Table 5</label>
          <caption>
            <p>Results of the Western Ontario and McMaster Universities Osteoarthritis Index–based processing approach (n=463) in terms of accuracy, precision, sensitivity, specificity, and F1-score for each model.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="220"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <thead>
              <tr valign="top">
                <td colspan="2">Model</td>
                <td>Accuracy (%)</td>
                <td>Precision (%)</td>
                <td>Sensitivity (%)</td>
                <td>Specificity (%)</td>
                <td><italic>F</italic><sub>1</sub>-score (%)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="7">
                  <bold>Word_Tokenize</bold>
                </td>
              </tr>
              <tr valign="top">
                <td/>
                <td>BiLSTM<sup>a</sup></td>
                <td>0.90</td>
                <td>0.91</td>
                <td>0.93</td>
                <td>0.87</td>
                <td>0.92</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>GRU<sup>b</sup></td>
                <td>0.88</td>
                <td>0.88</td>
                <td>0.93</td>
                <td>0.81</td>
                <td>0.90</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>CNN<sup>c</sup></td>
                <td>0.90</td>
                <td>0.91</td>
                <td>0.92</td>
                <td>0.87</td>
                <td>0.92</td>
              </tr>
              <tr valign="top">
                <td colspan="7">
                  <bold>DeepCut</bold>
                </td>
              </tr>
              <tr valign="top">
                <td/>
                <td>BiLSTM</td>
                <td>0.89</td>
                <td>0.88</td>
                <td>0.94</td>
                <td>0.81</td>
                <td>0.91</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>GRU</td>
                <td>0.89</td>
                <td>0.90</td>
                <td>0.91</td>
                <td>0.86</td>
                <td>0.91</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>CNN</td>
                <td>0.90</td>
                <td>0.91</td>
                <td>0.92</td>
                <td>0.87</td>
                <td>0.91</td>
              </tr>
              <tr valign="top">
                <td colspan="7">
                  <bold>AttaCut</bold>
                </td>
              </tr>
              <tr valign="top">
                <td/>
                <td>BiLSTM</td>
                <td>0.91</td>
                <td>0.91</td>
                <td>0.94</td>
                <td>0.87</td>
                <td>0.93</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>GRU</td>
                <td>0.89</td>
                <td>0.89</td>
                <td>0.93</td>
                <td>0.84</td>
                <td>0.91</td>
              </tr>
              <tr valign="top">
                <td/>
                <td>CNN</td>
                <td>0.90</td>
                <td>0.92</td>
                <td>0.92</td>
                <td>0.88</td>
                <td>0.92</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table5fn1">
              <p><sup>a</sup>BiLSTM: bidirectional long short-term memory.</p>
            </fn>
            <fn id="table5fn2">
              <p><sup>b</sup>GRU: gated recurrent unit.</p>
            </fn>
            <fn id="table5fn3">
              <p><sup>c</sup>CNN: convolutional neural network.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Overview</title>
        <p>While conventional medical data, such as images and statistical or laboratory data, have been extensively used, the use of textual data through NLP for knee OA diagnosis is still a relatively new approach. As mentioned previously, we propose an NLP model to detect the occurrence of knee OA by incorporating deep learning models, including GRU, BiLSTM, and CNN, based on textual data of knee OA symptoms reported in doctors’ notes. We discuss principal findings, comparison with prior work, strengths, limitations, and future work in the following subsections.</p>
      </sec>
      <sec>
        <title>Principal Findings</title>
        <p>A key finding in this study is that our method, using only textual data from doctors’ notes for knee OA prediction, performed better than previous studies that used image data and statistical or laboratory data. Moreover, in our experiments, the models trained very quickly due to the nature of our textual data (short sentences), which, in turn, can reduce the model’s runtime. We also found that the best tokenization algorithm and model for this experiment was BiLSTM combined with AttaCut tokenization.</p>
        <p>Although CNNs can extract relevant local features in sentences where specific combinations of words (local features) are important, it was not designed to handle sequential data like our data. Therefore, CNNs may not effectively handle variable-length sequences. GRUs have a simpler architecture and train faster than BiLSTM. However, GRUs may not perform as well as BiLSTM on very complex sequence data. This may be because BiLSTM can address the issues of complex sequences or long-distance dependencies in sentences better, as shown in our experiments, and thus, BiLSTM had the best performance. The best-performing model “before” applying our WOMAC-based processing approach was BiLSTM (AUC, 0.85; accuracy, 0.87; precision, 0.85; sensitivity, 0.95; specificity, 0.76; <italic>F</italic><sub>1</sub>-score, 0.90), and after applying our approach, there was an improvement in the results with BiLSTM (AUC, 0.91; accuracy, 0.91; precision, 0.91; sensitivity, 0.94; specificity, 0.87; <italic>F</italic><sub>1</sub>-score, 0.93). Therefore, our WOMAC-based processing approach can increase the performance of models. A comparison of the baseline model and the model using the WOMAC-based processing approach is shown in <xref ref-type="table" rid="table6">Table 6</xref>.</p>
        <table-wrap position="float" id="table6">
          <label>Table 6</label>
          <caption>
            <p>Comparison between the baseline approach and Western Ontario and McMaster Universities Osteoarthritis Index–based processing approach, and the improvement rate from baseline for each model.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="23"/>
            <col width="196"/>
            <col width="0"/>
            <col width="75"/>
            <col width="0"/>
            <col width="75"/>
            <col width="0"/>
            <col width="110"/>
            <col width="0"/>
            <col width="0"/>
            <col width="75"/>
            <col width="0"/>
            <col width="75"/>
            <col width="0"/>
            <col width="110"/>
            <col width="0"/>
            <col width="0"/>
            <col width="75"/>
            <col width="0"/>
            <col width="75"/>
            <col width="0"/>
            <col width="110"/>
            <col width="1"/>
            <thead>
              <tr valign="top">
                <td rowspan="2" colspan="3">Model</td>
                <td colspan="7">Baseline</td>
                <td colspan="7">WOMAC<sup>a</sup>-based processing approach</td>
                <td colspan="6">Improvement rate</td>
              </tr>
              <tr valign="top">
                <td colspan="2"><italic>F</italic><sub>1</sub>-score (%)</td>
                <td colspan="2">Macro avg<sup>b</sup> (%)</td>
                <td colspan="2">Weighted avg<sup>c</sup> (%)</td>
                <td colspan="3"><italic>F</italic><sub>1</sub>-score (%)</td>
                <td colspan="2">Macro avg<sup>b</sup> (%)</td>
                <td colspan="2">Weighted avg<sup>c</sup> (%)</td>
                <td colspan="3"><italic>F</italic><sub>1</sub>-score (%)</td>
                <td colspan="2">Macro avg<sup>b</sup> (%)</td>
                <td colspan="2">Weighted avg<sup>c</sup> (%)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="23">
                  <bold>Word_Tokenize</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>BiLSTM<sup>d</sup></td>
                <td colspan="2">0.89</td>
                <td colspan="2">0.88</td>
                <td colspan="2">0.89</td>
                <td colspan="3">0.92</td>
                <td colspan="2">0.90</td>
                <td colspan="2">0.90</td>
                <td colspan="3">0.03</td>
                <td colspan="2">0.02</td>
                <td colspan="2">0.01</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>GRU<sup>e</sup></td>
                <td colspan="2">0.88</td>
                <td colspan="2">0.85</td>
                <td colspan="2">0.86</td>
                <td colspan="3">0.90</td>
                <td colspan="2">0.88</td>
                <td colspan="2">0.88</td>
                <td colspan="3">0.02</td>
                <td colspan="2">0.03</td>
                <td colspan="2">0.02</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>CNN<sup>f</sup></td>
                <td colspan="2">0.89</td>
                <td colspan="2">0.86</td>
                <td colspan="2">0.87</td>
                <td colspan="3">0.92</td>
                <td colspan="2">0.90</td>
                <td colspan="2">0.90</td>
                <td colspan="3">0.03</td>
                <td colspan="2">0.04</td>
                <td colspan="2">0.03</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td colspan="23">
                  <bold>DeepCut</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>BiLSTM</td>
                <td colspan="2">0.89</td>
                <td colspan="2">0.85</td>
                <td colspan="2">0.86</td>
                <td colspan="3">0.91</td>
                <td colspan="2">0.88</td>
                <td colspan="2">0.89</td>
                <td colspan="3">0.02</td>
                <td colspan="2">0.03</td>
                <td colspan="2">0.03</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>GRU</td>
                <td colspan="2">0.88</td>
                <td colspan="2">0.84</td>
                <td colspan="2">0.85</td>
                <td colspan="3">0.91</td>
                <td colspan="2">0.89</td>
                <td colspan="2">0.89</td>
                <td colspan="3">0.03</td>
                <td colspan="2">0.04</td>
                <td colspan="2">0.04</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>CNN</td>
                <td colspan="2">0.89</td>
                <td colspan="2">0.85</td>
                <td colspan="2">0.86</td>
                <td colspan="3">0.91</td>
                <td colspan="2">0.89</td>
                <td colspan="2">0.90</td>
                <td colspan="3">0.02</td>
                <td colspan="2">0.04</td>
                <td colspan="2">0.04</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td colspan="23">
                  <bold>AttaCut</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>BiLSTM</td>
                <td colspan="2">0.90</td>
                <td colspan="2">0.86</td>
                <td colspan="2">0.87</td>
                <td colspan="3">0.93</td>
                <td colspan="2">0.91</td>
                <td colspan="2">0.91</td>
                <td colspan="3">0.03</td>
                <td colspan="2">0.05</td>
                <td colspan="2">0.04</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>GRU</td>
                <td colspan="2">0.89</td>
                <td colspan="2">0.86</td>
                <td colspan="2">0.86</td>
                <td colspan="3">0.91</td>
                <td colspan="2">0.88</td>
                <td colspan="2">0.89</td>
                <td colspan="3">0.02</td>
                <td colspan="2">0.02</td>
                <td colspan="2">0.03</td>
                <td>
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>CNN</td>
                <td colspan="2">0.89</td>
                <td colspan="2">0.86</td>
                <td colspan="2">0.87</td>
                <td colspan="3">0.92</td>
                <td colspan="2">0.88</td>
                <td colspan="2">0.89</td>
                <td colspan="3">0.03</td>
                <td colspan="2">0.02</td>
                <td colspan="2">0.02</td>
                <td>
                  <break/>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table6fn1">
              <p><sup>a</sup>AUC: area under the curve.</p>
            </fn>
            <fn id="table6fn2">
              <p><sup>b</sup>MAE: mean absolute error.</p>
            </fn>
            <fn id="table6fn3">
              <p><sup>c</sup>Not applicable.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Comparison With Prior Work</title>
        <p>In comparisons of our study with prior studies, we considered the primary data source used (image data or statistical data). Sources of data are presented in <xref ref-type="supplementary-material" rid="app4">Multimedia Appendix 4</xref>. A detailed comparison of the model in our study with the models in previous studies is shown in <xref ref-type="table" rid="table7">Table 7</xref>. When compared to prior models that used image data [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref22">22</xref>], our model showed better accuracy [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref17">17</xref>] and a higher AUC [<xref ref-type="bibr" rid="ref22">22</xref>]. When compared to prior models that used statistical data [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref19">19</xref>], our model showed higher accuracy, sensitivity, specificity, and AUC [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref6">6</xref>]. However, the AUC of our model was lower than that of the model in the study by Halilaj et al [<xref ref-type="bibr" rid="ref19">19</xref>].</p>
        <table-wrap position="float" id="table7">
          <label>Table 7</label>
          <caption>
            <p>Comparison of our model with models in previous studies.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="153"/>
            <col width="101"/>
            <col width="128"/>
            <col width="77"/>
            <col width="77"/>
            <col width="112"/>
            <col width="128"/>
            <col width="128"/>
            <col width="96"/>
            <thead>
              <tr valign="top">
                <td>Study (data source)</td>
                <td>Number of patients</td>
                <td>Accuracy (%)</td>
                <td>AUC<sup>a</sup> (%)</td>
                <td>MAE<sup>b</sup> (%)</td>
                <td>Precision (%)</td>
                <td>Sensitivity (%)</td>
                <td>Specificity (%)</td>
                <td><italic>F</italic><sub>1</sub>-score (%)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Brahim et al [<xref ref-type="bibr" rid="ref1">1</xref>] (images)</td>
                <td>1024</td>
                <td>0.83</td>
                <td>—<sup>c</sup></td>
                <td>—</td>
                <td>—</td>
                <td>0.87</td>
                <td>0.81</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>Shen et al [<xref ref-type="bibr" rid="ref2">2</xref>] (stats)</td>
                <td>195</td>
                <td>—</td>
                <td>—</td>
                <td>0.11</td>
                <td>—</td>
                <td>0.71</td>
                <td>0.67</td>
                <td>0.69</td>
              </tr>
              <tr valign="top">
                <td>Halilaj et al [<xref ref-type="bibr" rid="ref19">19</xref>] (stats)</td>
                <td>3924</td>
                <td>—</td>
                <td>0.95</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>Yong et al [<xref ref-type="bibr" rid="ref17">17</xref>] (images)</td>
                <td>4796</td>
                <td>0.88</td>
                <td>—</td>
                <td>0.33</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>Lim et al [<xref ref-type="bibr" rid="ref6">6</xref>] (stats)</td>
                <td>5749</td>
                <td>0.72</td>
                <td>0.77</td>
                <td>—</td>
                <td>—</td>
                <td>0.67</td>
                <td>0.73</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>Tiulpin et al [<xref ref-type="bibr" rid="ref22">22</xref>] (images)</td>
                <td>4928</td>
                <td>—</td>
                <td>0.79</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>Our study</td>
                <td>5849</td>
                <td>0.91</td>
                <td>0.91</td>
                <td>0.09</td>
                <td>0.91</td>
                <td>0.94</td>
                <td>0.87</td>
                <td>0.93</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table7fn1">
              <p><sup>a</sup>AUC: area under the curve.</p>
            </fn>
            <fn id="table7fn2">
              <p><sup>b</sup>MAE: mean absolute error.</p>
            </fn>
            <fn id="table7fn3">
              <p><sup>c</sup>Not applicable.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Strengths</title>
        <p>The overall performance of our model surpasses that of models in previous studies. Our proposed method introduces a novel approach to knee OA prediction research. Our research has 3 strengths. First, our model uses textual data, a valuable information source for AI learning. While textual data is critical, investigation and verification steps are also important. Second, since the nature of our dataset is textual and typically consists of short sentences, it does not require much processing time; hence, our model can be trained very quickly. Third, our patient dataset is larger than the datasets in previous studies (<xref ref-type="table" rid="table7">Table 7</xref>).</p>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>A limitation of our method is that it currently works only with the Thai language. As a result, it requires local application programming interfaces (APIs) that can handle the Thai language for processing. However, these local APIs may not have a programmer community as large as that for English or global APIs, and therefore, they may not be as productive. Our work focused on using textual data only in our model, and thus, we did not incorporate other types of data, such as image data and statistical or laboratory data, into our model, which could have improved diagnostic performance.</p>
      </sec>
      <sec>
        <title>Future Work</title>
        <p>In the future, we aim to improve our proposed method by incorporating image data and statistical data to achieve even better knee OA prediction. Additionally, we want to extend our work to make the model more precise by predicting the stage of knee OA (there are 5 stages based on WOMAC scores, ranging from 0 to 4) [<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref18">18</xref>]. Furthermore, our proposed method can be applied to other medical conditions or diseases, such as depression.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>We proposed a method to predict the occurrence of knee OA by using clinicians’ or doctors’ notes. This method has better prediction performance than other conventional methods that use images or statistical laboratory data. We confirmed the possibility of using symptom texts reported by patients (recorded by doctors) to predict knee OA conditions. Moreover, we demonstrated that the text of symptom reports can be a valuable data source to predict or classify a disease (whether a particular knee will have OA progression or not). Furthermore, our method involving text data (mostly consisting of short sentences) can save a considerable amount of time compared with other methods that use image data and statistical laboratory data (patients must visit the hospital and undergo imaging or tests to generate such data).</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>CONSORT (Consolidated Standards of Reporting Trials) checklist.</p>
        <media xlink:href="formative_v9i1e64536_app1.pdf" xlink:title="PDF File  (Adobe PDF File), 66 KB"/>
      </supplementary-material>
      <supplementary-material id="app2">
        <label>Multimedia Appendix 2</label>
        <p>CONSORT-EHEALTH (Consolidated Standards of Reporting Trials of Electronic and Mobile Health Applications and Online Telehealth) checklist.</p>
        <media xlink:href="formative_v9i1e64536_app2.pdf" xlink:title="PDF File  (Adobe PDF File), 1119 KB"/>
      </supplementary-material>
      <supplementary-material id="app3">
        <label>Multimedia Appendix 3</label>
        <p>STARD (Standards for Reporting of Diagnostic Accuracy Studies) checklist.</p>
        <media xlink:href="formative_v9i1e64536_app3.pdf" xlink:title="PDF File  (Adobe PDF File), 195 KB"/>
      </supplementary-material>
      <supplementary-material id="app4">
        <label>Multimedia Appendix 4</label>
        <p>Sources of data used according to the literature.</p>
        <media xlink:href="formative_v9i1e64536_app4.png" xlink:title="PNG File , 57 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AI</term>
          <def>
            <p>artificial intelligence</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">API</term>
          <def>
            <p>application programming interface</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">AUC</term>
          <def>
            <p>area under the curve</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">BiLSTM</term>
          <def>
            <p>bidirectional long short-term memory</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">CBOW</term>
          <def>
            <p>continuous bag of words</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">CNN</term>
          <def>
            <p>convolutional neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">GRU</term>
          <def>
            <p>gated recurrent unit</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">KL</term>
          <def>
            <p>Kellgren/Lawrence</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">LSTM</term>
          <def>
            <p>long short-term memory</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">MRI</term>
          <def>
            <p>magnetic resonance imaging</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">NLP</term>
          <def>
            <p>natural language processing</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">OA</term>
          <def>
            <p>osteoarthritis</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb13">ReLU</term>
          <def>
            <p>rectified linear unit</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb14">RNN</term>
          <def>
            <p>recurrent neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb15">WOMAC</term>
          <def>
            <p>Western Ontario and McMaster Universities Arthritis Index</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <notes>
      <sec>
        <title>Data Availability</title>
        <p>The datasets used in this study are not publicly available, as they require institutional approval on a case-by-case basis from the relevant authorities. However, the contact information of the relevant authorities can be requested by contacting the corresponding author.</p>
      </sec>
    </notes>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Brahim</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Jennane</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Riad</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Janvier</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Khedher</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Toumi</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Lespessailles</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>A decision support tool for early detection of knee OsteoArthritis using X-ray imaging and machine learning: Data from the OsteoArthritis Initiative</article-title>
          <source>Comput Med Imaging Graph</source>
          <year>2019</year>
          <month>04</month>
          <volume>73</volume>
          <fpage>11</fpage>
          <lpage>18</lpage>
          <pub-id pub-id-type="doi">10.1016/j.compmedimag.2019.01.007</pub-id>
          <pub-id pub-id-type="medline">30784984</pub-id>
          <pub-id pub-id-type="pii">S0895-6111(19)30003-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Yue</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A clinical model to predict the progression of knee osteoarthritis: data from Dryad</article-title>
          <source>J Orthop Surg Res</source>
          <year>2023</year>
          <month>08</month>
          <day>28</day>
          <volume>18</volume>
          <issue>1</issue>
          <fpage>628</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://josr-online.biomedcentral.com/articles/10.1186/s13018-023-04118-4"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s13018-023-04118-4</pub-id>
          <pub-id pub-id-type="medline">37635226</pub-id>
          <pub-id pub-id-type="pii">10.1186/s13018-023-04118-4</pub-id>
          <pub-id pub-id-type="pmcid">PMC10464113</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Farrow</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Ashcroft</surname>
              <given-names>GP</given-names>
            </name>
            <name name-style="western">
              <surname>Zhong</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Anderson</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Using artificial intelligence to revolutionise the patient care pathway in hip and knee arthroplasty (ARCHERY): Protocol for the development of a clinical prediction model</article-title>
          <source>JMIR Res Protoc</source>
          <year>2022</year>
          <month>05</month>
          <day>11</day>
          <volume>11</volume>
          <issue>5</issue>
          <fpage>e37092</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.researchprotocols.org/2022/5/e37092/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/37092</pub-id>
          <pub-id pub-id-type="medline">35544289</pub-id>
          <pub-id pub-id-type="pii">v11i5e37092</pub-id>
          <pub-id pub-id-type="pmcid">PMC9133991</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>CPC</given-names>
            </name>
            <name name-style="western">
              <surname>Chung</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>MS</given-names>
            </name>
            <name name-style="western">
              <surname>Cheng</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Liao</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Predicting the risk of total hip replacement by using a deep learning algorithm on plain pelvic radiographs: Diagnostic study</article-title>
          <source>JMIR Form Res</source>
          <year>2023</year>
          <month>10</month>
          <day>20</day>
          <volume>7</volume>
          <fpage>e42788</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://formative.jmir.org/2023//e42788/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/42788</pub-id>
          <pub-id pub-id-type="medline">37862084</pub-id>
          <pub-id pub-id-type="pii">v7i1e42788</pub-id>
          <pub-id pub-id-type="pmcid">PMC10625092</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Luo</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Effectiveness of internet-based telehealth programs in patients with hip or knee osteoarthritis: Systematic review and meta-analysis</article-title>
          <source>J Med Internet Res</source>
          <year>2024</year>
          <month>09</month>
          <day>30</day>
          <volume>26</volume>
          <fpage>e55576</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2024//e55576/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/55576</pub-id>
          <pub-id pub-id-type="medline">39348685</pub-id>
          <pub-id pub-id-type="pii">v26i1e55576</pub-id>
          <pub-id pub-id-type="pmcid">PMC11474128</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lim</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Cheon</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A deep neural network-based method for early detection of osteoarthritis using statistical data</article-title>
          <source>Int J Environ Res Public Health</source>
          <year>2019</year>
          <month>04</month>
          <day>10</day>
          <volume>16</volume>
          <issue>7</issue>
          <fpage>1281</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=ijerph16071281"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/ijerph16071281</pub-id>
          <pub-id pub-id-type="medline">30974803</pub-id>
          <pub-id pub-id-type="pii">ijerph16071281</pub-id>
          <pub-id pub-id-type="pmcid">PMC6480580</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nowinka</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Alagha</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Mahmoud</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Jones</surname>
              <given-names>GG</given-names>
            </name>
          </person-group>
          <article-title>Predicting depression in patients with knee osteoarthritis using machine learning: model development and validation study</article-title>
          <source>JMIR Form Res</source>
          <year>2022</year>
          <month>09</month>
          <day>13</day>
          <volume>6</volume>
          <issue>9</issue>
          <fpage>e36130</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://formative.jmir.org/2022/9/e36130/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/36130</pub-id>
          <pub-id pub-id-type="medline">36099008</pub-id>
          <pub-id pub-id-type="pii">v6i9e36130</pub-id>
          <pub-id pub-id-type="pmcid">PMC9518113</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Medical social media text classification integrating consumer health terminology</article-title>
          <source>IEEE Access</source>
          <year>2019</year>
          <volume>7</volume>
          <fpage>78185</fpage>
          <lpage>78193</lpage>
          <pub-id pub-id-type="doi">10.1109/access.2019.2921938</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Helwan</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Azar</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Ozsahin</surname>
              <given-names>DU</given-names>
            </name>
          </person-group>
          <article-title>Medical Reports Summarization Using Text-To-Text Transformer</article-title>
          <year>2023</year>
          <conf-name>2023 Advances in Science and Engineering Technology International Conferences (ASET)</conf-name>
          <conf-date>February 20-23, 2023</conf-date>
          <conf-loc>Dubai, United Arab Emirates</conf-loc>
          <pub-id pub-id-type="doi">10.1109/aset56582.2023.10180671</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wazrah</surname>
              <given-names>AA</given-names>
            </name>
            <name name-style="western">
              <surname>Alhumoud</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Sentiment analysis using stacked gated recurrent unit for Arabic tweets</article-title>
          <source>IEEE Access</source>
          <year>2021</year>
          <volume>9</volume>
          <fpage>137176</fpage>
          <lpage>137187</lpage>
          <pub-id pub-id-type="doi">10.1109/access.2021.3114313</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Santur</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Sentiment Analysis Based on Gated Recurrent Unit</article-title>
          <year>2019</year>
          <conf-name>2019 International Artificial Intelligence and Data Processing Symposium (IDAP)</conf-name>
          <conf-date>September 21-22, 2019</conf-date>
          <conf-loc>Malatya, Turkey</conf-loc>
          <pub-id pub-id-type="doi">10.1109/IDAP.2019.8875985</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ashalatha</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>T</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Geetha</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>T</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mittal</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>K</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>Predicting severity of osteoarthritis using recurrent convolutional neural networks (Rcnn) and medical imaging data</article-title>
          <source>South Eastern European Journal of Public Health</source>
          <year>2024</year>
          <month>09</month>
          <day>02</day>
          <fpage>47</fpage>
          <lpage>61</lpage>
          <pub-id pub-id-type="doi">10.70135/seejph.vi.894</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Al-Saqqa</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Awajan</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>The Use of Word2vec Model in Sentiment Analysis: A Survey</article-title>
          <source>AIRC '19: Proceedings of the 2019 International Conference on Artificial Intelligence, Robotics and Control</source>
          <year>2019</year>
          <conf-name>2019 International Conference on Artificial Intelligence, Robotics and Control</conf-name>
          <conf-date>December 14-16, 2019</conf-date>
          <conf-loc>Cairo, Egypt</conf-loc>
          <pub-id pub-id-type="doi">10.1145/3388218.3388229</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Alharbi</surname>
              <given-names>O</given-names>
            </name>
          </person-group>
          <article-title>A deep learning approach combining CNN and Bi-LSTM with SVM classifier for Arabic sentiment analysis</article-title>
          <source>International Journal of Advanced Computer Science and Applications</source>
          <year>2021</year>
          <volume>12</volume>
          <issue>6</issue>
          <fpage>0120618</fpage>
          <pub-id pub-id-type="doi">10.14569/ijacsa.2021.0120618</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hampton</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Osteoarthritis initiative</article-title>
          <source>JAMA</source>
          <year>2004</year>
          <month>04</month>
          <day>28</day>
          <volume>291</volume>
          <issue>16</issue>
          <fpage>1951</fpage>
          <pub-id pub-id-type="doi">10.1001/jama.291.16.1951-a</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xin Teoh</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Othmani</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Li Goh</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Usman</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lai</surname>
              <given-names>KW</given-names>
            </name>
            <name name-style="western">
              <surname>Lai</surname>
              <given-names>KW</given-names>
            </name>
          </person-group>
          <article-title>Deciphering knee osteoarthritis diagnostic features with explainable artificial intelligence: A systematic review</article-title>
          <source>IEEE Access</source>
          <year>2024</year>
          <volume>12</volume>
          <fpage>109080</fpage>
          <lpage>109108</lpage>
          <pub-id pub-id-type="doi">10.1109/access.2024.3439096</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yong</surname>
              <given-names>CW</given-names>
            </name>
            <name name-style="western">
              <surname>Teo</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Murphy</surname>
              <given-names>BP</given-names>
            </name>
            <name name-style="western">
              <surname>Hum</surname>
              <given-names>YC</given-names>
            </name>
            <name name-style="western">
              <surname>Tee</surname>
              <given-names>YK</given-names>
            </name>
            <name name-style="western">
              <surname>Xia</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Lai</surname>
              <given-names>KW</given-names>
            </name>
          </person-group>
          <article-title>Knee osteoarthritis severity classification with ordinal regression module</article-title>
          <source>Multimed Tools Appl</source>
          <year>2021</year>
          <month>01</month>
          <day>26</day>
          <volume>81</volume>
          <issue>29</issue>
          <fpage>41497</fpage>
          <lpage>41509</lpage>
          <pub-id pub-id-type="doi">10.1007/s11042-021-10557-0</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Helwan</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Azar</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Abdellatef</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>An update on the knee osteoarthritis severity grading using wide residual learning</article-title>
          <source>XST</source>
          <year>2022</year>
          <month>09</month>
          <day>30</day>
          <volume>30</volume>
          <issue>5</issue>
          <fpage>1009</fpage>
          <lpage>1021</lpage>
          <pub-id pub-id-type="doi">10.3233/xst-221190</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Halilaj</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Le</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Hicks</surname>
              <given-names>JL</given-names>
            </name>
            <name name-style="western">
              <surname>Hastie</surname>
              <given-names>TJ</given-names>
            </name>
            <name name-style="western">
              <surname>Delp</surname>
              <given-names>SL</given-names>
            </name>
          </person-group>
          <article-title>Modeling and predicting osteoarthritis progression: data from the osteoarthritis initiative</article-title>
          <source>Osteoarthritis Cartilage</source>
          <year>2018</year>
          <month>12</month>
          <volume>26</volume>
          <issue>12</issue>
          <fpage>1643</fpage>
          <lpage>1650</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1063-4584(18)31415-8"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.joca.2018.08.003</pub-id>
          <pub-id pub-id-type="medline">30130590</pub-id>
          <pub-id pub-id-type="pii">S1063-4584(18)31415-8</pub-id>
          <pub-id pub-id-type="pmcid">PMC6469859</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kweon</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Jang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Choi</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Chun</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Khang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Oh</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Data resource profile: the Korea National Health and Nutrition Examination Survey (KNHANES)</article-title>
          <source>Int J Epidemiol</source>
          <year>2014</year>
          <month>02</month>
          <volume>43</volume>
          <issue>1</issue>
          <fpage>69</fpage>
          <lpage>77</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/24585853"/>
          </comment>
          <pub-id pub-id-type="doi">10.1093/ije/dyt228</pub-id>
          <pub-id pub-id-type="medline">24585853</pub-id>
          <pub-id pub-id-type="pii">dyt228</pub-id>
          <pub-id pub-id-type="pmcid">PMC3937975</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>You</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Chyr</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lan</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Noble</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>X</given-names>
            </name>
          </person-group>
          <article-title>Causal discovery in radiographic markers of knee osteoarthritis and prediction for knee osteoarthritis severity with attention-long short-term memory</article-title>
          <source>Front Public Health</source>
          <year>2020</year>
          <month>12</month>
          <day>18</day>
          <volume>8</volume>
          <fpage>604654</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/33409263"/>
          </comment>
          <pub-id pub-id-type="doi">10.3389/fpubh.2020.604654</pub-id>
          <pub-id pub-id-type="medline">33409263</pub-id>
          <pub-id pub-id-type="pmcid">PMC7779681</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tiulpin</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Klein</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Bierma-Zeinstra</surname>
              <given-names>SMA</given-names>
            </name>
            <name name-style="western">
              <surname>Thevenot</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Rahtu</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Meurs</surname>
              <given-names>JV</given-names>
            </name>
            <name name-style="western">
              <surname>Oei</surname>
              <given-names>EHG</given-names>
            </name>
            <name name-style="western">
              <surname>Saarakkala</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Multimodal machine learning-based knee osteoarthritis progression prediction from plain radiographs and clinical data</article-title>
          <source>Sci Rep</source>
          <year>2019</year>
          <month>12</month>
          <day>27</day>
          <volume>9</volume>
          <issue>1</issue>
          <fpage>20038</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-019-56527-3"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-019-56527-3</pub-id>
          <pub-id pub-id-type="medline">31882803</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-019-56527-3</pub-id>
          <pub-id pub-id-type="pmcid">PMC6934728</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Segal</surname>
              <given-names>NA</given-names>
            </name>
            <name name-style="western">
              <surname>Nevitt</surname>
              <given-names>MC</given-names>
            </name>
            <name name-style="western">
              <surname>Gross</surname>
              <given-names>KD</given-names>
            </name>
            <name name-style="western">
              <surname>Hietpas</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Glass</surname>
              <given-names>NA</given-names>
            </name>
            <name name-style="western">
              <surname>Lewis</surname>
              <given-names>CE</given-names>
            </name>
            <name name-style="western">
              <surname>Torner</surname>
              <given-names>JC</given-names>
            </name>
          </person-group>
          <article-title>The Multicenter Osteoarthritis Study: opportunities for rehabilitation research</article-title>
          <source>PM R</source>
          <year>2013</year>
          <month>08</month>
          <volume>5</volume>
          <issue>8</issue>
          <fpage>647</fpage>
          <lpage>54</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/23953013"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.pmrj.2013.04.014</pub-id>
          <pub-id pub-id-type="medline">23953013</pub-id>
          <pub-id pub-id-type="pii">S1934-1482(13)00197-4</pub-id>
          <pub-id pub-id-type="pmcid">PMC3867287</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jang</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Harerimana</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Kang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>JW</given-names>
            </name>
          </person-group>
          <article-title>Bi-LSTM model to increase accuracy in text classification: Combining Word2vec CNN and attention mechanism</article-title>
          <source>Applied Sciences</source>
          <year>2020</year>
          <month>08</month>
          <day>24</day>
          <volume>10</volume>
          <issue>17</issue>
          <fpage>5841</fpage>
          <pub-id pub-id-type="doi">10.3390/app10175841</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yahia</surname>
              <given-names>HS</given-names>
            </name>
            <name name-style="western">
              <surname>Abdulazeez</surname>
              <given-names>AM</given-names>
            </name>
          </person-group>
          <article-title>Medical text classification based on convolutional neural network: a review</article-title>
          <source>International Journal of Science and Business</source>
          <year>2021</year>
          <volume>5</volume>
          <issue>3</issue>
          <fpage>27</fpage>
          <lpage>41</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://ideas.repec.org/a/aif/journl/v5y2021i3p27-41.html"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Qing</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Linhong</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Xuehai</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>A novel neural network-based method for medical text classification</article-title>
          <source>Future Internet</source>
          <year>2019</year>
          <month>12</month>
          <day>10</day>
          <volume>11</volume>
          <issue>12</issue>
          <fpage>255</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.3390/fi11120255"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/fi11120255</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Schulz</surname>
              <given-names>KF</given-names>
            </name>
            <name name-style="western">
              <surname>Altman</surname>
              <given-names>DG</given-names>
            </name>
            <name name-style="western">
              <surname>Moher</surname>
              <given-names>D</given-names>
            </name>
            <collab>CONSORT Group</collab>
          </person-group>
          <article-title>CONSORT 2010 statement: updated guidelines for reporting parallel group randomized trials</article-title>
          <source>Ann Intern Med</source>
          <year>2010</year>
          <month>06</month>
          <day>01</day>
          <volume>152</volume>
          <issue>11</issue>
          <fpage>726</fpage>
          <lpage>32</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.acpjournals.org/doi/abs/10.7326/0003-4819-152-11-201006010-00232?url_ver=Z39.88-2003&#38;rfr_id=ori:rid:crossref.org&#38;rfr_dat=cr_pub  0pubmed"/>
          </comment>
          <pub-id pub-id-type="doi">10.7326/0003-4819-152-11-201006010-00232</pub-id>
          <pub-id pub-id-type="medline">20335313</pub-id>
          <pub-id pub-id-type="pii">0003-4819-152-11-201006010-00232</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Eysenbach</surname>
              <given-names>G</given-names>
            </name>
            <collab>CONSORT-EHEALTH Group</collab>
          </person-group>
          <article-title>CONSORT-EHEALTH: improving and standardizing evaluation reports of Web-based and mobile health interventions</article-title>
          <source>J Med Internet Res</source>
          <year>2011</year>
          <month>12</month>
          <day>31</day>
          <volume>13</volume>
          <issue>4</issue>
          <fpage>e126</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2011/4/e126/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/jmir.1923</pub-id>
          <pub-id pub-id-type="medline">22209829</pub-id>
          <pub-id pub-id-type="pii">v13i4e126</pub-id>
          <pub-id pub-id-type="pmcid">PMC3278112</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bossuyt</surname>
              <given-names>PM</given-names>
            </name>
            <name name-style="western">
              <surname>Reitsma</surname>
              <given-names>JB</given-names>
            </name>
            <name name-style="western">
              <surname>Bruns</surname>
              <given-names>DE</given-names>
            </name>
            <name name-style="western">
              <surname>Gatsonis</surname>
              <given-names>CA</given-names>
            </name>
            <name name-style="western">
              <surname>Glasziou</surname>
              <given-names>PP</given-names>
            </name>
            <name name-style="western">
              <surname>Irwig</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Lijmer</surname>
              <given-names>JG</given-names>
            </name>
            <name name-style="western">
              <surname>Moher</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Rennie</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>de Vet</surname>
              <given-names>HCW</given-names>
            </name>
            <name name-style="western">
              <surname>Kressel</surname>
              <given-names>HY</given-names>
            </name>
            <name name-style="western">
              <surname>Rifai</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Golub</surname>
              <given-names>RM</given-names>
            </name>
            <name name-style="western">
              <surname>Altman</surname>
              <given-names>DG</given-names>
            </name>
            <name name-style="western">
              <surname>Hooft</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Korevaar</surname>
              <given-names>DA</given-names>
            </name>
            <name name-style="western">
              <surname>Cohen</surname>
              <given-names>JF</given-names>
            </name>
            <collab>STARD Group</collab>
          </person-group>
          <article-title>STARD 2015: an updated list of essential items for reporting diagnostic accuracy studies</article-title>
          <source>BMJ</source>
          <year>2015</year>
          <month>10</month>
          <day>28</day>
          <volume>351</volume>
          <fpage>h5527</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.bmj.com/lookup/pmidlookup?view=long&#38;pmid=26511519"/>
          </comment>
          <pub-id pub-id-type="doi">10.1136/bmj.h5527</pub-id>
          <pub-id pub-id-type="medline">26511519</pub-id>
          <pub-id pub-id-type="pmcid">PMC4623764</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wołk</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Marasek</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <person-group person-group-type="editor">
            <name name-style="western">
              <surname>Rocha</surname>
              <given-names>Á</given-names>
            </name>
            <name name-style="western">
              <surname>Correia</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Tan</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Stroetmann</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>A Sentence Meaning Based Alignment Method for Parallel Text Corpora Preparation</article-title>
          <source>New Perspectives in Information Systems and Technologies, Volume 1. Advances in Intelligent Systems and Computing, vol 275</source>
          <year>2014</year>
          <publisher-loc>Cham</publisher-loc>
          <publisher-name>Springer</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jinks</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Jordan</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Croft</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Measuring the population impact of knee pain and disability with the Western Ontario and McMaster Universities Osteoarthritis Index (WOMAC)</article-title>
          <source>Pain</source>
          <year>2002</year>
          <month>11</month>
          <volume>100</volume>
          <issue>1-2</issue>
          <fpage>55</fpage>
          <lpage>64</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/S0304-3959(02)00239-7"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/s0304-3959(02)00239-7</pub-id>
          <pub-id pub-id-type="medline">12435459</pub-id>
          <pub-id pub-id-type="pii">S0304395902002397</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Arreerard</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Mander</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Piao</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Survey on Thai NLP Language Resources and Tools</article-title>
          <source>Proceedings of the Thirteenth Language Resources and Evaluation Conference</source>
          <year>2022</year>
          <conf-name>Thirteenth Language Resources and Evaluation Conference</conf-name>
          <conf-date>June 2022</conf-date>
          <conf-loc>Marseille, France</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chormai</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Prasertsom</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Cheevaprawatdomrong</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Rutherford</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Syllable-based Neural Thai Word Segmentation</article-title>
          <source>Proceedings of the 28th International Conference on Computational Linguistics</source>
          <year>2020</year>
          <conf-name>28th International Conference on Computational Linguistics</conf-name>
          <conf-date>December 2020</conf-date>
          <conf-loc>Barcelona, Spain (Online)</conf-loc>
          <pub-id pub-id-type="doi">10.18653/v1/2020.coling-main.407</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Boriboon</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kriengket</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Chootrakool</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Phaholphinyo</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Purodakananda</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Thanakulwarapas</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>BEST Corpus Development and Analysis</article-title>
          <year>2009</year>
          <conf-name>2009 International Conference on Asian Language Processing</conf-name>
          <conf-date>December 07-09, 2009</conf-date>
          <conf-loc>Singapore</conf-loc>
          <pub-id pub-id-type="doi">10.1109/IALP.2009.76</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Phatthiyaphaibun</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Chaovavanich</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Polpanumas</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Suriyawongkul</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lowphansirikul</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Chormai</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Limkonchotiwat</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Suntorntip</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Udomcharoenchaikit</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>PyThaiNLP: Thai Natural Language Processing in Python</article-title>
          <source>arXiv</source>
          <access-date>2025-06-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/2312.04649">https://arxiv.org/abs/2312.04649</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Theeramunkong</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Sornlertlamvanich</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Tanhermhong</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Chinnan</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>Character cluster based Thai information retrieval</article-title>
          <source>IRAL '00: Proceedings of the Fifth International Workshop on Information Retrieval with Asian Languages</source>
          <year>2000</year>
          <conf-name>Fifth International Workshop on Information Retrieval with Asian Languages</conf-name>
          <conf-date>September 30, 2000-October 1, 2000</conf-date>
          <conf-loc>Hong Kong China</conf-loc>
          <pub-id pub-id-type="doi">10.1145/355214.355225</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Seeha</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Bilan</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Sanchez</surname>
              <given-names>LM</given-names>
            </name>
            <name name-style="western">
              <surname>Huber</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Matuschek</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Schütze</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>ThaiLMCut: Unsupervised Pretraining for Thai Word Segmentation</article-title>
          <source>Proceedings of the Twelfth Language Resources and Evaluation Conference</source>
          <year>2020</year>
          <conf-name>Twelfth Language Resources and Evaluation Conference</conf-name>
          <conf-date>May 2020</conf-date>
          <conf-loc>Marseille, France</conf-loc>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chormai</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Prasertsom</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Rutherford</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>AttaCut: A Fast and Accurate Neural Thai Word Segmenter</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <access-date>2025-06-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1911.07056">https://arxiv.org/abs/1911.07056</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shao</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Taylor</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Marshall</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Morioka</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Zeng-Treitler</surname>
              <given-names>Q</given-names>
            </name>
          </person-group>
          <article-title>Clinical Text Classification with Word Embedding Features vs. Bag-of-Words Features</article-title>
          <year>2018</year>
          <conf-name>2018 IEEE International Conference on Big Data (Big Data)</conf-name>
          <conf-date>December 10-13, 2018</conf-date>
          <conf-loc>Seattle, WA, USA</conf-loc>
          <pub-id pub-id-type="doi">10.1109/bigdata.2018.8622345</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jatnika</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Bijaksana</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Suryani</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Word2Vec model analysis for semantic similarities in English words</article-title>
          <source>Procedia Computer Science</source>
          <year>2019</year>
          <volume>157</volume>
          <fpage>160</fpage>
          <lpage>167</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/j.procs.2019.08.153"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.procs.2019.08.153</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hung</surname>
              <given-names>PT</given-names>
            </name>
            <name name-style="western">
              <surname>Yamanishi</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Word2vec skip-gram dimensionality selection via sequential normalized maximum likelihood</article-title>
          <source>Entropy (Basel)</source>
          <year>2021</year>
          <month>07</month>
          <day>31</day>
          <volume>23</volume>
          <issue>8</issue>
          <fpage>997</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=e23080997"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/e23080997</pub-id>
          <pub-id pub-id-type="medline">34441136</pub-id>
          <pub-id pub-id-type="pii">e23080997</pub-id>
          <pub-id pub-id-type="pmcid">PMC8391962</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref42">
        <label>42</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Irsoy</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Benton</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Stratos</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Corrected CBOW Performs as well as Skip-gram</article-title>
          <source>arXiv</source>
          <year>2020</year>
          <access-date>2025-06-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/2012.15332">https://arxiv.org/abs/2012.15332</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref43">
        <label>43</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zulqarnain</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ghazali</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Mohmad Hassim</surname>
              <given-names>YM</given-names>
            </name>
            <name name-style="western">
              <surname>Rehan</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Text classification based on gated recurrent unit combines with support vector machine</article-title>
          <source>IJECE</source>
          <year>2020</year>
          <month>08</month>
          <day>01</day>
          <volume>10</volume>
          <issue>4</issue>
          <fpage>3734</fpage>
          <pub-id pub-id-type="doi">10.11591/ijece.v10i4.pp3734-3742</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref44">
        <label>44</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Minaee</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Azimi</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Abdolrashidi</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Deep-Sentiment: Sentiment Analysis Using Ensemble of CNN and Bi-LSTM Models</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <access-date>2025-06-12</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1904.04206">https://arxiv.org/abs/1904.04206</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref45">
        <label>45</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Otter</surname>
              <given-names>DW</given-names>
            </name>
            <name name-style="western">
              <surname>Medina</surname>
              <given-names>JR</given-names>
            </name>
            <name name-style="western">
              <surname>Kalita</surname>
              <given-names>JK</given-names>
            </name>
          </person-group>
          <article-title>A survey of the usages of deep learning for natural language processing</article-title>
          <source>IEEE Trans Neural Netw Learning Syst</source>
          <year>2021</year>
          <month>2</month>
          <volume>32</volume>
          <issue>2</issue>
          <fpage>604</fpage>
          <lpage>624</lpage>
          <pub-id pub-id-type="doi">10.1109/tnnls.2020.2979670</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref46">
        <label>46</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Meng</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Qiu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>X</given-names>
            </name>
          </person-group>
          <article-title>Sentiment analysis of comment texts based on BiLSTM</article-title>
          <source>IEEE Access</source>
          <year>2019</year>
          <volume>7</volume>
          <fpage>51522</fpage>
          <lpage>51532</lpage>
          <pub-id pub-id-type="doi">10.1109/access.2019.2909919</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref47">
        <label>47</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yao</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Mao</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Luo</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Clinical text classification with rule-based features and knowledge-guided convolutional neural networks</article-title>
          <source>BMC Med Inform Decis Mak</source>
          <year>2019</year>
          <month>04</month>
          <day>04</day>
          <volume>19</volume>
          <issue>Suppl 3</issue>
          <fpage>71</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmcmedinformdecismak.biomedcentral.com/articles/10.1186/s12911-019-0781-4"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12911-019-0781-4</pub-id>
          <pub-id pub-id-type="medline">30943960</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12911-019-0781-4</pub-id>
          <pub-id pub-id-type="pmcid">PMC6448186</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref48">
        <label>48</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Guo</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Ma</surname>
              <given-names>X</given-names>
            </name>
          </person-group>
          <article-title>Improving text classification with weighted word embeddings via a multi-channel TextCNN model</article-title>
          <source>Neurocomputing</source>
          <year>2019</year>
          <month>10</month>
          <volume>363</volume>
          <fpage>366</fpage>
          <lpage>374</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/j.neucom.2019.07.052"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.neucom.2019.07.052</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref49">
        <label>49</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Jeong</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Sentiment classification using convolutional neural networks</article-title>
          <source>Applied Sciences</source>
          <year>2019</year>
          <month>06</month>
          <day>07</day>
          <volume>9</volume>
          <issue>11</issue>
          <fpage>2347</fpage>
          <pub-id pub-id-type="doi">10.3390/app9112347</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref50">
        <label>50</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Salur</surname>
              <given-names>MU</given-names>
            </name>
            <name name-style="western">
              <surname>Aydin</surname>
              <given-names>I</given-names>
            </name>
          </person-group>
          <article-title>A novel hybrid deep learning model for sentiment classification</article-title>
          <source>IEEE Access</source>
          <year>2020</year>
          <volume>8</volume>
          <fpage>58080</fpage>
          <lpage>58093</lpage>
          <pub-id pub-id-type="doi">10.1109/access.2020.2982538</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
