Published on in Vol 13 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/70940, first published .
Extracting Symptoms of Complex Conditions From Online Discourse (Subreddit to Symptomatology): Lexicon-Based Approach

Extracting Symptoms of Complex Conditions From Online Discourse (Subreddit to Symptomatology): Lexicon-Based Approach

Extracting Symptoms of Complex Conditions From Online Discourse (Subreddit to Symptomatology): Lexicon-Based Approach

Original Paper

1Department of Computer Science and Engineering, Bangladesh University of Engineering and Technology, Dhaka, Bangladesh

2Department of Computer Science, Dartmouth College, Hanover, NH, United States

3Department of Biomedical Data Science, Geisel School of Medicine, Hanover, NH, United States

4Department of Computer Science and Engineering, Bangladesh University of Professionals, Dhaka, Bangladesh

5Department of Obstetrics & Gynecology, Bangladesh College of Physicians and Surgeons, Dhaka, Bangladesh

6Faculty of Information Technology, Monash University, Clayton, Australia

Corresponding Author:

Mohammed Eunus Ali, PhD

Faculty of Information Technology

Monash University

20 Exhibition Walk (Woodside Building)

Clayton, VIC 3168

Australia

Phone: 61 3 9905 1741

Email: eunus.ali@monash.edu


Background: Millions of people affected with complex medical conditions with diverse symptoms often turn to online discourse to share their experiences. While some studies have explored natural language processing methods and medical information extraction tools, these typically focus on generic symptoms in clinical notes and struggle to identify patient-reported, disease-specific, subtle symptoms from online health discourse.

Objective: We aimed to extract patient-reported, disease-specific symptoms shared on social media reflecting the lived experiences of thousands of affected individuals and explore the characteristics, prevalence, and occurrence patterns of the symptoms.

Methods: We propose a lexicon-based symptom extraction (LSE) method to identify a diverse list of disease-specific, patient-reported symptoms. We initially used a large language model to accelerate the extraction of symptom-related key phrases that formed the lexicon. We evaluated the effectiveness of lexicon extraction against human annotation using a Jaccard index score. We then leveraged BERT-Base, BioBERT, and Phrase-BERT–based embeddings to learn representations of these symptom-related key phrases and cluster similar symptoms using k-means and hierarchical density-based spatial clustering of applications with noise (HDBSCAN). Among the different options explored in our experiments, BioBERT-based k-means clustering was found to be the most effective. Finally, we applied symptom normalization to eliminate duplicate and redundant entries in the comprehensive symptom list.

Results: In a real-world polycystic ovary syndrome (PCOS) subreddit dataset, we found that LSE significantly outperformed state-of-the-art baselines, achieving at least 41% and 20% higher F1-scores (mean 86.10) than automatic medical extraction tools and large language models, respectively. Notably, the comprehensive list of 64 PCOS symptoms generated via LSE ensured extensive coverage of symptoms reported in 7 reputable eHealth forums. Analyzing PCOS symptomatology revealed 28 potentially emerging symptoms and 8 self-reported comorbidities co-occurring with PCOS.

Conclusions: The comprehensive patient-reported, disease-specific symptom list can help patients and health practitioners resolve uncertainties surrounding the disease, eliminating the variability of PCOS symptoms prevailing in the community. Analyzing PCOS symptomatology across varied dimensions provides valuable insights for public health research.

JMIR Med Inform 2025;13:e70940

doi:10.2196/70940

Keywords



Background

The rise of health 2.0 has significantly increased the consumption of health-related information on social media in recent years. Worldwide, 34.7% of adult internet users frequently search for health-related information online, with approximately 58.5% of adults in the United States specifically engaging in such searches as of 2023 [1,2]. Analyzing online discourse on health-related topics offers valuable insights into the lived experiences and perspectives of affected individuals [3-5]. The critical findings of these analyses are particularly significant for understanding complex medical conditions, such as different types of cancer, cognitive and mental health conditions, autoimmune disorders, and metabolic and endocrinological conditions. These medical conditions often manifest with diverse symptoms that vary significantly among patients [6-8]. Furthermore, there is often no consensus among established medical sources regarding the full range of possible symptoms for such conditions [9,10]. This variability poses challenges for patients, health care providers, and researchers seeking a comprehensive understanding of these symptoms and their characteristics. Thus, the study and analysis of symptoms associated with a disease, referred to as symptomatology, is crucial for understanding complex diseases [11]. We argue that a robust approach to extracting disease-specific symptoms from individuals’ lived experiences shared on social media can address uncertainties surrounding complex medical conditions. This approach would also allow for the identification of symptoms that patients associate with a condition (eg, brain fog with post–COVID-19 condition), prevalence of symptom co-occurrence (eg, unexplained weight loss and pain with cancer), and how symptoms vary based on patients’ comorbidities and preexisting conditions.

Recent advancements in natural language processing (NLP) have enabled public health researchers to explore medical information extraction tools [12-15] and develop novel techniques [16-18] for analyzing online health discourse, including patient-reported data from social media. However, off-the-shelf medical information extraction tools primarily excel in extracting well-known medical entities (eg, diseases, drugs, and basic symptoms). Their limitations become evident with granular, patient-centric details (eg, unusual symptom descriptions, implicit context, or subtle adverse events), which often require specialized rules or retrained models for accurate detection [19-22]. Modern transformer-based models used with medical named entity recognition have shown promising results in symptom extraction [23,24]. However, these models typically require large amounts of domain-specific data and are mostly trained on clinical notes (electronic health records), which limits their effectiveness when dealing with colloquial health discourse generated by patients [7,23-25]. A family of bidirectional encoder representations from transformers (BERT)–based models fine-tuned on colloquial data markedly improves symptom recognition in informal texts; nevertheless, their error analysis indicates that the best-trained models still capture only approximately 25% to 50% of lengthy, figurative symptom phrases and continue to misclassify many subtle or previously unseen expressions (eg, “mild difficulty taking deep breaths” or “chest again felt a little tight”) [26]. A recent NLP pipeline [27] used to extract symptoms from patient narratives on social media proposed a dual-corpus symptom lexicon containing keywords (eg, fever and cough) and related modifiers (eg, mild or severe) followed by applying a rule-based postprocessing step to merge the 2. However, this fixed rule-based approach may struggle to report symptoms whenever people describe them implicitly, with spelling noise, or in new phrasing that never entered the seed lists (eg, “can’t breathe properly” or “my lungs felt on fire”).

While the existing lexicon-based approaches have successfully used predefined resources and deep learning frameworks for broad public health surveillance, their reliance on explicit wording limits their ability to capture the full range of informal expressions. On the other hand, large language models (LLMs) have significantly accelerated medical NLP research, especially in the field of information extraction [16,17,21,28]. While some studies have demonstrated the potential of LLMs for health information extraction from social media [29,30], they face challenges when extracting specific contexts such as health advice and disease-specific treatment [12,16].

Objectives

To overcome the aforementioned limitations of symptom extraction from social media discourse, we propose a novel NLP pipeline that systematically combines the capabilities of LLMs and other NLP methods with feedback from domain experts, resulting in a scalable and practical solution to the problem of symptom extraction from patient-reported social media discourse.

While the proposed pipeline can be applied to explore a plethora of medical conditions for which there is an abundance of patient-reported online discourse, we focused on a common chronic condition with significant uncertainty, namely, polycystic ovary syndrome (PCOS), as a use case in this paper. According to the World Health Organization [31], an estimated 116 million women (3.4% of reproductive-age women) worldwide are affected by PCOS, and its prevalence is on the rise. Despite its widespread impact, the full spectrum of PCOS remains insufficiently explored [32]. Patients report a diverse range of symptoms that vary in severity depending on previous medical history and their effect on daily life. The combination of varied symptoms, insufficient treatment guidelines, and instances of medical gaslighting often leads to psychological distress, prompting many individuals to seek support and share experiences on social media platforms [33], including Reddit. Existing research leveraging social media data to study PCOS is still underdeveloped [4,34-36]. This creates a unique opportunity to systematically identify self-reported symptoms and related symptomatology based on the lived experiences of thousands of individuals and, thereby, inform the clinical understanding of this complex medical condition. We make 3 primary contributions. First, we developed a novel lexicon-based symptom extraction (LSE) method to identify diverse disease-specific symptoms from individuals’ experiences shared on social media. This method can effectively detect even subtle symptoms mentioned in user discourse. We applied our LSE method to a novel PCOS subreddit dataset, achieving statistically significant improvements over state-of-the-art medical NLP baselines.

Second, we conducted an in-depth analysis of peer discussion threads on the extracted PCOS symptoms to reveal a set of lesser-known yet prevalent symptoms that individuals associate with PCOS and associated co-occurring symptoms and comorbidities. These findings can contribute to clinical research and public health research in a community-informed way.

Third, we curated a novel Reddit dataset containing 59,962 and 641,441 relevant posts and comments, respectively. This dataset captures self-reported PCOS symptoms and their context from 68,010 affected individuals. In addition, we curated a human-annotated, labeled dataset of 1000 posts of patient-reported symptoms related to PCOS based on guidance from domain experts and patient input. We release these resources in a public domain. These unique, rich resources can enable medical NLP and computational health researchers to further investigate the symptomatology of complex medical conditions identified and analyzed online.


Data Collection and Preprocessing

We selected Reddit as a source of data because of its anonymous nature. It encourages users to share personal experiences on stigmatized topics [4,37]. The data collection process began with the identification of the r/PCOS subreddit by considering the group posts’ availability and popularity among 153,000 members engaging in discussions on PCOS. We systematically collected data from this subreddit using the Python Reddit API Wrapper package (Python Software Foundation) and the Pushshift application programming interface (API) [38,39]. We obtained all the posts and comments from the r/PCOS subreddit from August 4, 2010, to December 31, 2022. This timeline was chosen to ensure a comprehensive capture of recent as well as previous discussions. For each post, we retained the date, flair name, unique post ID, post title, post content, score, upvotes, and comments due to their utility and relevance.

During preprocessing, we discarded entries with deleted posts or those in which the post content had <5 words as the brevity poses challenges to deriving meaningful insights from posts. We also removed posts containing irrelevant text (eg, polls, survey links only, or URLs). The URL-heavy posts were classified using a rule-based heuristic combining personal pronoun frequency and URL-to-text ratio. We provide a summary of the steps to remove irrelevant text from social media data in Table S1 in Multimedia Appendix 1. This filtering process enabled us to focus on patient-authored, experience-rich discussions. Finally, we obtained a primary dataset comprising 59,962 posts and 641,441 comments by 68,010 active unique users. Identifying task-relevant data is crucial for data-driven approaches [12] as it significantly impacts the efficiency of downstream tasks. We leveraged Reddit’s post “flair” option to filter relevant data. On Reddit, a post flair is a customizable text or icon that appears next to a post title to provide additional context or categorization, similar to hashtags available on other platforms. We reviewed 30 randomly selected posts from several flairs of the PCOS subreddit. We identified that, among all the flairs, posts in the “General/Advice” flair were more likely to share knowledge and seek advice on symptoms. Further investigation on the most frequent unigrams of these flairs supported our qualitative analysis, where the presence of symptom-related unigrams such as “period,” “symptom,” “doctor,” “birth,” “control,” and “diagnosed” was predominantly observed in the “General/Advice” flair. The distribution of frequent unigrams in the “General/Advice” flair is shown in Figure S1 in Multimedia Appendix 1. On the basis of these findings, we chose the “General/Advice” flair for extracting a wide range of symptoms. In addition, the “General/Advice” flair is more popular in the community as it contains the highest number of posts, as shown in Table S2 in Multimedia Appendix 1. This selection ensured a broad range of symptoms and valuable insights. Finally, we developed a corpus of 10,500 posts for the symptom identification task, randomly selected from the posts with the “General/Advice” flair.

Our Method: LSE

We propose an LSE method to analyze self-reported symptoms of a complex medical condition using social media discourse. Our proposed method consists of three main steps: (1) symptom-related lexicon extraction, (2) symptom identification, and (3) symptom normalization, as illustrated in Figure 1.

Figure 1. Steps of the lexicon-based symptom extraction method. JI: Jaccard index; LLM: large language model.
Symptom-Related Lexicon Extraction

We leveraged symptom-related key phrases for lexicon extraction as it helped distill the most relevant terms and phrases associated with physical manifestations from user narratives. We focused on extracting the key symptom-related phrases instead of the symptoms as the latter results in extracting (1) phrases that are not relevant and (2) repetitive mentions of the same symptoms, rendering the subsequent analysis challenging. Compared to the tedious and time-intensive task of manual extraction, LLMs emerge as a more economical choice, offering consistency in their responses, especially when the temperature setting is near 0 [40,41]. We used the OpenAI API gpt-3.5-turbo-16k [42] to annotate 10,500 user posts. To ensure precision in identifying the physical manifestation of the condition, we emphasized physical symptoms while excluding psychological aspects in our base prompt using a zero-shot prompting strategy (Figure S2 in Multimedia Appendix 1). The API responses were stored in a JSON line format containing key phrases and corresponding posts. This process yielded a lexicon of 30,370 symptom-related key phrases that we used in the subsequent steps.

We resorted to human annotation to validate the extraction conducted by ChatGPT (OpenAI). We employed 2 graduate students in the science, technology, engineering, and mathematics field particularly active in women’s health–related social media forums. Before annotation, they underwent multiple training sessions led by a domain expert specializing in computational health. These sessions were designed to refine their understanding on key phrase extraction and promote annotation consistency. On a sample set of 30 posts, the interannotator agreement reached 92% in a relaxed match. Finally, we randomly selected 500 entries from 10,500 user posts. We recognize that many social media and health NLP studies use larger annotated sets, often containing 2000 to 4000 messages, and few drop below 1000 annotations [12,43-45]. However, those datasets are typically used to validate end-to-end model performance, whereas our 500-post sample supported a specific component of our pipeline. We deliberately divided tasks between the LLM and human experts to balance coverage and precision while remaining practical as (1) the cost in terms of time and resources was significant (eg, annotating just 500 entries required 4 training sessions × 3 hours = 12 hours, 35 hours for literature study and documentation, and 208.3 hours approximately for annotation per post; 500 posts × 25 minutes), making larger samples prohibitively expensive; (2) we observed a significant drop in new symptom-related key phrase discoveries after the first 200 posts, mirroring the classic “diminishing returns” shown in Table S3 in Multimedia Appendix 1, also evident in social media symptom mining and related tasks [46]; and (3) during manual inspection, we observed minimal disagreement as the annotation task was primarily syntactic rather than interpretive. Taking into consideration the above 3 points, we assigned 250 unique posts to each annotator to minimize bias and enhance the quality of our annotated set.

We choose the Jaccard index (JI) to measure the similarity between key phrases extracted by ChatGPT and the human annotators [12]. This required normalizing the extracted key phrases (eg, changing “pressure in pelvic area” to “pelvic pressure”) to maintain homogeneity in both lists. We computed the JI similarity for each post in the sample using exact key phrase match. Table 1 shows 3 posts from our sample set with extracted key phrases and their corresponding JI scores. The average JI score was 74%, indicating moderate lexical similarity between the key phrases extracted by ChatGPT and the human annotators considering the complexity of the task [47]. Our manual investigation of all missed or incorrectly extracted key phrases by the LLM revealed 5 distinct error categories, as outlined in Table S4 in Multimedia Appendix 1. For instance, in row 1 of Table 1, the LLM missed “absence of period” as it was implicitly mentioned in the post as “haven’t had a period since December.” Similarly, in row 3, “insulin resistance” is an ambiguous expression to be accepted as a symptom. However, a deeper analysis revealed 3 findings supporting the acceptability of the lexicon.

First, post-level JI analysis; our manual investigation showed that 12.6% (63/500) of the posts had a JI score of 0, which substantially affected the average JI. We found that most of these JIs with a value of 0 resulted from very short posts (eg, “Does anyone know if weightloss in PCOS helps or throws off hormones?”) and were annotated as no symptom by humans, whereas the LLM incorrectly extracted the key phrases weightloss and hormones. The median JI score across all posts was 1.0 (IQR 0.5-1.0), indicating strong agreement between the LLM and human annotations in most cases.

Second, lexicon-level robustness; we observed that the LLM missed a key phrase in a post, resulting in a mismatch, but extracted the same key phrase in another post. Similar cases were found for human annotation, resulting in incorrect extraction for the LLM. We identified such cases for 61% (104/172) of the total missed or incorrectly extracted key phrases, which affected the average JI. As the final lexicon aggregates key phrases across all posts, these disagreements at the post level did not propagate to the lexicon level. When comparing the LLM-derived lexicon with the human-annotated lexicon, we observed a JI of 0.89, demonstrating strong alignment between the 2. This high agreement confirmed that discrepancies in per-post extraction had minimal impact on the final lexicon’s accuracy and coverage.

Third, downstream impact on symptom list; the frequency of unique incorrectly extracted key phrases in the lexicon was very low (maximum of 3). Consequently, their influence on the cluster centroids was negligible. Because our symptom list was derived by selecting the most frequent phrases within each cluster, the effect of noisy or rare phrases was minimized. Such key phrases either formed microclusters (eg, ultrasound) that were discarded during the normalization step or were absorbed into broader, validated symptom cluster categories.

Table 1. Sample excerpts of human and ChatGPT annotations of posts. The Jaccard index (JI) was calculated for the similarity between the 2 annotations based on the extracted key phrases.
PostHuman annotationChatGPT annotationJI score
“I haven’t had a period since December, however in the past week I keep getting really bad cramping with sharp pains (almost feels like my ovaries), but no period...has anyone else experienced this? Is this normal?”“Absence of period,” “cramping,” and “sharp pains”“Cramping” and “sharp pains”0.67
“I’m 22, F and over the years, I have gained 50 pounds. Recently, my acne is getting worse and I feel pressure in my pelvic area. I am trying to lose weight and thereby started working out but the scale is still not budging.”“Weight gain,” “acne,” and “pelvic pressure”“Weight gain,” “acne,” and “pelvic pressure”1.0
“Due to PCOS and insulin resistance, Does anyone else get symptoms like they’re getting their period (sore breasts, crampy feeling) and just overall malaise when you’re ‘due’ for your period you don’t actually get?”“Insulin resistance,” “sore breasts,” and “crampy feeling”“Sore breasts,” “crampy feeling,” and “malaise”0.5

In summary, while there were post-level discrepancies, the high lexicon-level JI (0.89) and the robustness of frequency-based clustering ensured that the final symptom list remained accurate and representative of human annotation without prohibitive annotation burden.

Symptom Identification
Overview

In this step, we identified a distinct set of disease symptoms using the symptom-related lexicon. As the key phrases in the lexicon were captured exactly as mentioned in users’ posts, many were semantically similar with different phrasing. For example, key phrases such as cystic ovary, ovarian cysts, and multiple cysts in ovaries have the same semantic meaning with different surface forms. Thus, we used clustering to categorize the semantically similar key phrases and identify such categories as unique symptoms. To ensure accurate clustering, it is crucial to generate meaningful representations of symptom-related key phrases that effectively capture their semantic meaning. To achieve this, we used embeddings, which convert words into numeric representations in lower dimensions that preserve their contexts. The process of embedding and clustering the key phrases is described in the following sections.

Choosing the Right Embedding

We explored and evaluated widely used transformer models such as BERT-Base [48], Phrase-BERT [49], and BioBERT [50] to generate embeddings of symptom-related key phrases. We observed the performance of these embeddings by applying various clustering methods. For k-means clustering, we analyzed the silhouette score for each BERT-based embedding. The silhouette score considers both within-cluster and intercluster variances to determine cluster quality [51]. We plotted silhouette scores of different embedding approaches, shown in Figure 2A. The curves for BERT-Base and Phrase-BERT show a general increase but noticeable fluctuation, representing inconsistency in the results. In contrast, BioBERT shows a consistent and uniform ascent compared to the other 2 in the silhouette curve. Furthermore, applying L2 normalization during embedding improved the uniqueness of the clusters [48]. The consistent superiority of BioBERT across all clustering approaches steered our decision to adopt BioBERT as our embedding method.

Figure 2. Characteristic analysis of embedding methods for selecting optimal clustering methods—(A) silhouette score analysis of different embeddings on k-means and (B) rate of change in inertia (within-cluster sum of the squares) decreasing as the number of clusters increases. BERT: bidirectional encoder representations from transformers.
Selecting the Optimal Clustering Method

We exploited k-means and hierarchical density-based spatial clustering of applications with noise (HDBSCAN) with BioBERT embeddings for clustering symptom-related key phrases. Comparing k-means and HDBSCAN is fairly tricky as their hyperparameters (number of clusters vs density thresholds) are fundamentally different. Although it is widely known that HDBSCAN performs better for noisy social media data as it filters out “low-density” points as noise, the choice of clustering method depends on the downstream task and the performance on specific data. In Table S5 in Multimedia Appendix 1, we report the performance of HDBSCAN under a grid of min_cluster_size and min_samples settings with the observations of cluster numbers, density-based clustering validation (DBCV) score, and noise. The DBCV score measures how dense and well separated core clusters are in a density-based clustering method [52]. Thus, HDBSCAN with lower minimum_cluster_size increases in DBCV scores indicates compactness of the clusters but may generate spurious microclusters. Manual observation shows that HDBSCAN with lower minimum_cluster_size provides better intracluster coherence but reduce intercluster separation and generate more semantically similar clusters, such as increased testosterone and elevated level of testosterone. In contrast, an increasing minimum_cluster_size results in lower DBCV scores. As a result, the number of clusters decreases, which may contribute to losing meaningful clusters by being too strict on cluster size. For such instances, elements of the cysts on ovaries cluster represents less coherence and semantic similarity in HDBSCAN compared to k-means shown in (Figure 3). In our use case, HDBSCAN labeled 31% to 56% of the points as noise across settings. This affected the downstream task as our goal was to maximize symptom coverage. On the other hand, k-means has no noise filtering, which ensures maximum coverage of symptoms. In addition, we implemented k-means using the cosine distance to improve the cluster uniqueness with fewer duplicates compared to using the Euclidean distance [53].

Figure 3. Comparison of elements within the cluster "cysts on ovaries" between 2 clustering methods: (A) k-means, containing semantically similar key phrases that are evenly distributed, and (B) hierarchical density-based spatial clustering of applications with noise, containing key phrases conflicting with the cluster name, highlighted in red, and showing more uneven distribution.
Sensitivity and Stability Analysis in Choosing the k Value

We varied the number of clusters (k) to determine the optimal number based on the silhouette score (in Figure 2A) and rate of change in inertia (in Figure 2B). In our PCOS use case, we identified that, after k=100, the change in inertia became flat and the rate of increment in silhouette score was also negligible. Therefore, we set k to range from 10 to 100, and at k=96, we observed an optimal silhouette score. In addition, we explored a –10 to +10 window for k=96 to verify the sensitivity of the resulting clusters. Across k=86 to 106, the silhouette score remained within approximately –0.01 to +0.01 of its value at k=96, and the inertia curve showed minimal change, confirming that the choice of k sat in the middle of a stable region. In addition, we report the bootstrap stability results for 3 representative choices between k=86 and 96. All 3 settings yielded high median Adjusted Rand Index (ARI) values (≥0.75) and narrow 95% CIs, demonstrating that cluster assignments are largely consistent when up to 20% of the data are held out. We provide a summary of the sensitivity and stability analysis of the choice of k in k-means in Table S6 in Multimedia Appendix 1.

Symptom Normalization
Overview

Each of the obtained clusters represented a symptom as the elements of the clusters were semantically similar. We considered the most frequent element of the cluster as the key symptom. The justification for selecting the most frequent element as the key symptom was 2-fold. First, its occurrences were significantly higher than those of other elements within the cluster. For instance, occurrences of cysts on ovaries compared to other elements within the cluster in Figure 3A were negligible. Second, all elements within a cluster consistently pointed toward the same physical condition as the key symptom. In the k-means cluster in Figure 3A, all the cluster elements referred to ovarian cysts. However, there were repetitions of symptoms in the obtained list. Therefore, we conducted symptom normalization by applying semantic mapping and validation by experts, as outlined in the following sections.

Applying Semantic Mapping

We applied semantic mapping of clusters, a manual process that involves merging symptoms with similar meanings and concepts using Unified Medical Language System Metathesaurus synonyms to avoid repetition. Through progressive discussions with our medical specialist, we formalized 4 semantic mapping rules suitable for this task. For instance, irregular bleeding, irregular menstrual cycles, and irregular cycles were mapped to irregular periods using rule 1, “synonym merging.” Applying these rules, we consolidated 11 such mappings from the initial 96 PCOS symptoms to 68 unique symptoms. The complete rule set is provided in Table S7 in Multimedia Appendix 1 to ensure reproducibility.

Validation by an Expert

We further refined the symptom list by validating the relevance of the lived experiences of patients to the disease. One of our coauthors, a gynecologist with >10 years of experience in treating PCOS, conducted this validation step. The expert’s role was to discard symptoms that were not related to the disease and resolve ambiguities between similar symptoms. For instance, the medical specialist discarded 4 irrelevant symptoms—ultrasound, birth control, low-carb diet, and neck—by manually investigating the cluster elements, resulting in a comprehensive list of 64 unique PCOS symptoms.

Ethical Considerations

We acknowledge that the research on symptom extraction from social media has multifaceted ethical concerns and requires careful attention to privacy, consent, and the responsible use of data and technology. The research was approved by Bangladesh University of Engineering and Technology's Institutional Review Board. Although Reddit is an anonymous platform, we ensured that no personally identifiable information was inadvertently revealed during data collection, analysis, or publication. All Reddit posts and comments cited as examples in this paper were paraphrased to protect user privacy. We also recognize that the demographic composition of Reddit users might not be fully representative of the broader population with PCOS, potentially limiting generalizability. Therefore, findings derived from Reddit discourse should be generalized cautiously to other populations or settings. In addition, we recognize that any biases in the dataset and the proposed model are unintentional. As the outcomes of this work are based on social media content, we emphasize the superiority of medical evidence–based facts in cases of conflict, particularly concerning direct clinical practice and medical diagnosis. We strongly encourage the use of the model for research and exploratory purposes, serving as a foundation for further developments in computational health research.


Test Data Curation and Ground Truth Annotation

We initially separated 1000 sample posts of various lengths from our dataset for results evaluation. Considering the influence of post length on symptom identification, we categorized the posts into three groups: (1) short posts of <100 words, (2) medium-length posts of between 100 and 300 words, and (3) long posts of >300 words. Post length distribution in the dataset revealed that 45% (450/1000) of the posts belonged to the medium-length category (Table S8 in Multimedia Appendix 1). To ensure a balance across all categories, we selected 500 posts from the medium-length category and 250 posts from each of the short and long post categories for curating test data.

We recruited a medical graduate with domain knowledge and a patient with PCOS lived experience to perform annotation. Before annotation, both annotators participated in progressive training sessions and interactive discussions with medical specialists to refine their understanding and improve annotation consistency. During this phase, the medical specialist primarily identified 3 potential disagreements between the annotators. These observed disagreements were instances in which (1) the medical graduate often used clinical terminology (eg, “amenorrhea”) whereas the patient used informal language (eg, “no period for months”); (2) the patient included self-reported experiences (eg, “hypothyroidism”) that lacked clinical validation whereas the clinician strictly adhered to established PCOS diagnostic criteria; and (3) the medical graduate inferred implicit symptoms (eg, laboratory test results) from domain knowledge, which was lacking in the patient annotation. Given our goal to prioritize the capture of patient-experienced symptoms, medical specialists suggested resolving these discrepancies by (1) avoiding overly technical clinical terms unless they were explicitly mentioned, (2) retaining self-reported experiences unless they were clinically invalid, and (3) including implicit symptoms as long as they were related to PCOS. Following these guidelines, both annotators extracted symptoms independently. We adopted the combination of both annotators’ symptom lists per post to maximize coverage and ensure the consistency and reliability of the ground truth dataset.

In our ground truth annotation, annotators were not restricted to predefined symptom lists. They identified any symptoms that they perceived as PCOS related, resulting in free-text labels. Moreover, each post included multiple symptoms, making the traditional Cohen κ not suitable. Instead, we evaluated interannotator agreement using relaxed-match F1-score due to the generative nature of our annotation [16,54,55]. We used BioBERT-based semantic similarity with the medical graduate’s annotation as a reference. We achieved an F1-score of 0.82, with a recall of 0.84 at a similarity threshold of 0.6, indicating substantial agreement considering the realistic complexity inherent in patient-derived, generative-nature symptom annotation.

Symptom Comparison: LSE With eHealth Forums

LSE resulted in a comprehensive list of 64 unique PCOS symptoms (Table 2) from 10,500 user posts. We also determined the unique users reporting each symptom in our primary dataset, with counts ranging from 31 to 16,001. This confirmed that LSE captures widely experienced symptoms, whereas less common symptoms are not ignored. In addition, this list provides a wide range of coverage as we cross-checked it with 7 evidence-based consumer health websites: the World Health Organization [31]; the Mayo Clinic [56]; WebMD [57]; the International Classification of Diseases, 10th Revision [58]; Johns Hopkins Medicine [59]; the National Health Service [60]; and the Centers for Disease Control and Prevention [61]. All symptoms mentioned in these sources were included in our list except for 4: sleep apnea, high blood pressure, low progesterone, and dandruff. Upon examining the clusters, we found that these 4 symptoms were represented within their respective key clusters—insomnia, high cholesterol, low estrogen, and high DHEA—due to their close medical relevance. Furthermore, all the forums recognized the 10 most frequently reported symptoms in our dataset (Table S9 in Multimedia Appendix 1). On the other hand, we identified 28 symptoms not acknowledged by any of the eHealth forums that we mention as potentially emerging symptoms in the Discussion section. Thus, our comprehensive symptom list reflects the diversity of self-reported PCOS lived experience symptoms and aligns closely with authoritative medical sources, helping standardize the variability in the symptom descriptions.

Table 2. Comprehensive list of 64 polycystic ovary syndrome symptoms extracted via lexicon-based symptom extraction. The counts represent the number of unique users reporting each symptom across the primary dataset. The symptoms in italics represent potentially emerging symptoms.
SymptomUnique users reporting the symptom, n
Hirsutism16,001
Acne14,256
Cysts on ovaries13,781
Pain13,495
Weight gain13,150
Mood swings11,523
Insulin resistance9859
Irregular period9702
Infertility7558
Cramps7043
Diabetes6690
Spotting6469
Hypothyroidism5366
No period4846
Anovulation4755
Nausea4600
Cravings4566
Fatigue4342
Bloating4048
Headaches4028
Heavy period3846
Hair loss3517
High testosterone2794
Skin issues2695
Hormone imbalance2648
Inflammation2457
Endometriosis2303
High DHEAa2231
Loss of appetite2013
Painful periods1906
Insomnia1513
Dizziness1471
Night sweats1457
Stomach pain1374
Blood clots1187
Anemia1162
Brain fog1158
Vomiting1134
High androgen1121
Oily skin912
Back pain843
Digestive issues826
Constipation813
Pelvic pain742
Belly fat731
Acanthosis nigricans616
High cholesterol613
High insulin575
Low libido575
Abdominal pain564
Ovary pain496
Dry skin480
High estrogen480
Stretch marks439
Low estrogen436
Brown discharge426
Skin tags399
Breast tenderness376
High prolactin331
Pain during sex261
Frequent urination160
Redness108
Weight loss45
High LHb level31

aDHEA: dehydroepiandrosterone.

bLH: luteinizing hormone.

Baseline Comparisons

Overview

We compared the performance of LSE with leading baselines, including 2 popular off-the-shelf models—Amazon Comprehend Medical (ACM) [62] and Google Healthcare Natural Language API (GHNLP) [63]—in their default configuration. In addition, we evaluated both open-source and closed-source LLMs, namely, Llama 3.1 (8B) and GPT-4o in a zero-shot, near-0 temperature (0.25) setting. We identified symptoms in each post of the test dataset using all the baselines for comparison with the ground truth. For LSE, we used GPT-4o to identify symptoms guided by the comprehensive symptom list obtained through our method. Therefore, it was not strictly zero-shot but, rather, a form of lexicon-guided prompting. The model’s task was to match symptoms in the text against the LSE-provided symptom lexicon. To ensure reproducibility, we included the LLM prompt templates used for the baseline comparisons (GPT-4o and Llama 3.1), shown in Figure S3 in Multimedia Appendix 1, and LSE evaluation using GPT-4o, as detailed in Figure S4 in Multimedia Appendix 1. We evaluated competitive models using standard performance metrics—precision, recall, and F1-score. To measure the similarity between symptoms, we used BioBERT-based semantic similarity as it performs optimally with medical data. Given the nuanced nature of symptom expressions, we set the different semantic similarity thresholds at 0.4, 0.6, and 0.8 for match acceptance, and the corresponding performance results are reported in Table 3. Experimental results show that LSE outperformed all baselines significantly across all metrics at every threshold level. As expected, increasing the similarity threshold imposed stricter matching criteria, resulting in fewer accepted symptom matches. Consequently, the performance of all baseline models dropped significantly with higher thresholds. In contrast, the performance of LSE decreased only marginally, highlighting the stability and robustness of our framework. For all subsequent analyses in this study, we adopted a semantic similarity threshold of 0.6 as the default setting. Several additional useful insights from the results are discussed in the following sections.

Table 3. Comparing lexicon-based symptom extraction (LSE) with off-the-shelf medical information extraction tools and large language models in terms of precision, recall, and F1-score across different post lengths.
ModelPost length


≤100 words>100 to 300 words>300 wordsFor all length, mean F1-score

PrecisionRecallF1-scorePrecisionRecallF1-scorePrecisionRecallF1-score
Similarity threshold of 0.4

ACMa55.2779.8561.21b39.2889.5051.3031.7791.2145.3652.30

GHNLPc59.9663.4158.69b49.3965.7951.7645.2968.3051.2553.37

Llama 3.1 (8B)53.2673.0057.3862.4979.1165.91b63.0774.0064.6663.47

GPT-4o64.1560.2560.7779.6174.7274.8184.2171.6275.33b71.41

LSE86.0985.0284.6087.8588.5486.8992.0990.7690.33b87.17d
Similarity threshold of 0.6

ACM48.6368.6153.45b33.4776.2943.8528.8581.1440.3045.37

GHNLP47.8249.5446.64b35.5246.7837.3135.2453.0040.0240.32

Llama 3.1 (8B)48.2063.4251.6056.4771.4859.64b55.8664.9657.2757.04

GPT-4o59.0555.6356.0272.1168.8668.5777.9766.8670.14b65.80

LSE84.8983.8283.4086.6487.4285.7891.2089.5789.42b86.10d
Similarity threshold of 0.8

ACM37.6651.1041.13b20.6846.0726.8718.5852.8126.6330.38

GHNLP37.8540.0137.62b24.4533.1526.1825.8339.2329.5429.88

Llama 3.1 (8B)41.1252.2343.60b40.8151.5943.1340.3945.4140.9642.71

GPT-4o49.7647.8843.6055.1752.1252.2160.4951.0754.18b51.60

LSE80.0079.3378.8183.0483.2982.0286.7885.3385.12b81.99d

aACM: Amazon Comprehend Medical.

bFor these values, the highest value obtained by a model among the 3-length range.

cGHNLP: Google Healthcare Natural Language Application Programming Interface.

dFor these values, the LSE obtained the highest F1-scores among all models in all the threshold-settings.

Off-the-Shelf Models Tend to Overpredict

The GHNLP and ACM models largely extracted generic symptoms that were not disease specific, such as pregnancy, cold, sick, illness, and discomfort. In addition, they extracted mental conditions such as anxiety, overwhelmed, and feeling bad as PCOS symptoms. As a result, they identified more symptoms (4.84 and 9.05 per post by GHNLP and ACM on average, respectively) compared to the ground truth (mean 2.76, SD 1.90 per post). Moreover, they struggled to comprehend the context of the post content, often extracting medical jargon such as side effect, lab report, and medication as symptoms. Substantially, GHNLP and ACM extracted 64% (2874/4463) and 70% (5691/8161) incorrect symptoms, respectively, of the total symptoms identified (Figure 4). Longer posts aggravated these issues as the presence of more jargon further reduced the precision and F1-scores of these models (Table 3). However, both models achieved better recall than precision due to overprediction.

Figure 4. Performance comparison of different models in identifying symptom occurrences from 1000 sample posts. Each bar represents the number of correctly identified, not identified, and incorrectly identified symptoms stacked on top of each other. ACM: Amazon Comprehend Medical; GHNLP: Google Healthcare Natural Language Application Programming Interface; LSE: lexicon-based symptom extraction.
Baselines Struggle to Capture Nuanced Symptoms

All the baselines struggled to detect subtle symptoms, specifically when they were stated in informal language. As a result, GHNLP correctly identified only 50% (1589/3124) of the ground truth symptoms, whereas the LLMs showed better performance with approximately 65% (4133/6248) correct matches (Figure 4). Although ACM achieved a 79% (2470/3124) match with the ground truth, its overall performance was hindered by significant overpredictions. For instance, a PCOS symptom, high testosterone, was mentioned informally in the following post, but only LSE and none of the baselines could identify it:

I had two sets of blood tests taken last year, which showed that my hormones (especially testosterone) were all over the place. My GP said that my test results were enough to confirm PCOS diagnosis.
LSE Outperformed State-of-the-Art LLMs

Figure 4 shows that, while LLMs significantly reduced the number of incorrectly identified symptoms, they still missed a notable number of symptoms. Leveraging the symptom list generated by LSE in the base prompt for generative pretrained transformers increased the number of correctly identified symptoms, enhancing overall performance. The high values for LSE across all metrics in Table 3 indicate that the comprehensive symptom list effectively covered all PCOS lived experience symptoms while filtering out generic ones.

Statistical Significance

We also compared the F1-score of LSE with those of the baselines using the approximated randomization test [28] with a similarity threshold of 0.6 to determine whether the results were statistically significant. As this test does not require specific distributional assumptions, it is particularly suitable for complex metrics such as the F1-score [64]. We conducted pairwise comparisons between symptoms extracted from a post via LSE and each of the baselines. The number of occurrences in the sample set of 1000 posts in which LSE significantly outperformed the baselines at 95% (P<.05) and 90% (P<.10) confidence levels. The observed results suggest that, in terms of the F1-score, LSE significantly outperformed all the other baselines shown in Table S10 in Multimedia Appendix 1.


Analysis of PCOS Symptomatology

Overview

Our study was motivated by the limitations of existing information extraction methods in identifying symptom expressions from colloquial patient narratives. We propose an LSE method incorporating LLMs into the pipeline to capture disease-specific symptoms from social media discourse. We focused on PCOS as a representative case given its highly variable and often underdocumented symptomatology. We evaluated LSE against multiple strong baselines across a manually annotated Reddit dataset, and it demonstrated substantial improvements in precision, recall, and F1-score.

We conducted an in-depth study of PCOS symptomatology to evaluate the characteristics, prevalence, and occurrence patterns of the symptoms, capturing implicit insights from patient-generated data. We also highlighted how capturing symptoms in patient-centric language can bridge gaps between patient experiences and clinical practice, potentially enhancing health care provider communication and guiding clinical decision-making.

Potentially Emerging Symptoms

We identified 28 distinct symptoms from our comprehensive symptom list prominently reported by patients yet currently underrepresented or insufficiently explored in established medical literature and not acknowledged by existing eHealth sources (highlighted in italics in Table 2). Although many of these symptoms currently lack robust clinical evidence linking them directly to PCOS, their frequent occurrence in patient-generated data suggests that patients often experience and associate these symptoms with PCOS. As illustrated in Figure 5, the number of unique users reporting some of these symptoms, such as hypothyroidism and brain fog, demonstrate a consistent upward reporting trend over time. Moreover, some recent studies have provided medical evidence supporting the connection between some of these symptoms and PCOS, such as brain fog, thyroid disorder, low libido, digestive issues, and cravings [65-69]. Therefore, we marked them as potentially emerging symptoms. However, while frequently reported by patients, it is important to recognize that these symptoms might not represent entirely novel manifestations of PCOS. Instead, they may reflect less common but known clinical symptoms, secondary manifestations of the primary disease, or symptoms associated with frequently co-occurring conditions. For example, hypothyroidism, reported by 5366 users, reflects a well-established thyroid-PCOS comorbidity [66], underscoring potential shared pathophysiological mechanisms rather than indicating a novel PCOS symptom. In addition, nausea and cravings, although frequently reported by patients, advocate for a broader and patient-inclusive diagnostic perspective rather than confirming direct causality. These emerging findings reinforce the credibility of patient-reported underdocumented symptoms and underscore the necessity for rigorous clinical investigation to determine their prevalence, pathophysiological connections, and implications for diagnosis and treatment protocols.

Figure 5. Instances of potentially emerging symptoms mentioned by users in the primary dataset, showing notable upward trends over time.
Capturing Symptoms in Patient-Centric Language

LSE captured the lived experience symptoms as expressed by the patients. This approach makes the symptoms more accessible to laypersons compared to clinical terminology and jargon. Some symptoms with unorthodox clinical terms require complex medical tests to diagnose, whereas their effects are often evident in lived experiences, and hence, these symptoms are included in the symptom list. For instance, hyperandrogenism is a symptom that may not be familiar to all users and requires laboratory tests, whereas hirsutism, acne, or hair loss are visible outcomes of the condition that are easily recognized and commonly reported by users, as shown in Table 2.

Identifying Co-Occurrence of Symptoms

We constructed a symptom co-occurrence matrix in which multiple symptoms mentioned by an individual in a single post or comment were considered as co-occurring. Figure 6 illustrates a segment of the matrix highlighting the most frequently mentioned symptom pairs with their number of co-occurrences. The eHealth forums further complemented these co-occurrences. Understanding these co-occurring symptoms raises awareness among the community and helps individuals better cope with their condition. In addition, such analysis can assist health care providers and patients in proactively diagnosing and managing symptoms such as hirsutism, oily skin, and acne, as observed in the following comment:

The best natural remedy I’ve tried is spearmint supplement. I’m a thin cysterbut have most of the other symptoms and it has been the only thing I’ve foundso far that helps with hirsutism, oily skin, acne.
Figure 6. A segment of the symptom co-occurrence matrix, where hirsutism and acne appear as one of the most frequently mentioned symptom pairs.
Association With Chronic Diseases and Comorbidity

We identified 8 chronic diseases from the list of PCOS symptoms that were distinct medical conditions in their own right. For example, chronic conditions such as hypothyroidism and endometriosis were mentioned by 5366 and 2303 unique users, respectively (Table 2). The association with chronic diseases and comorbidity increases the complexity of the disease and its management and significantly impacts social determinants of health, as shown in the following example:

I went to the ER...but they found cysts on my ovaries and mentioned PCOS with possibly Endometriosis. Now, I have no insurance and I’m struggling to find the care. My husband and I want to start a family, but I feel like our chances are slipping away because I can’t afford the help I need. Any advice on where to get financial or medical support would be greatly appreciated.

The financial strain from frequent physician visits and prolonged medications impacts patients’ overall health outcomes. This emphasizes the need for clinical and public health research to find ways to reduce the burden on patients, such as overprescription and cost. This may also involve policy reforms to ensure comprehensive coverage for chronic conditions associated with PCOS.

Enhancing Patient–Health Care Provider Communication

The comprehensive symptom list generated by our lexicon-based approach, encompassing both widely recognized and emerging patient-reported symptoms, provides clinicians with a deeper insight into how patients perceive and articulate their condition. For example, terms such as brain fog, pelvic pressure, and fatigue are not only clinically relevant but also capture the nuanced ways in which patients describe their daily experiences with PCOS. Studies also show that using patient-generated health-related terminology during visits deepens shared understanding and decision‑making [70]. Therefore, clinicians can leverage this enriched vocabulary to enhance patient–health care provider communication, promoting empathy, validating patient experiences, and facilitating more targeted and personalized clinical interviews.

Guide to Clinical Decision-Making

The patient-derived symptom list (Table 2) can serve as a preliminary checklist during consultations to identify overlooked or underdiscussed symptoms that patients might hesitate to share spontaneously due to stigma or uncertainty. Documenting co-occurrence patterns further empowers clinicians to anticipate linked symptoms. A patient presenting with hirsutism, for instance, could be screened for insulin resistance or acne, enabling early intervention and holistic management. To scale these benefits, the symptom list can be integrated into electronic health record systems. This integration allows for auto-completion prompts that reflect patient language while supporting consistent clinical documentation. Such structured integration enhances clinical decision-making through automated reminders and promotes interoperability for population-level research. Acknowledging patient-reported symptoms does not undermine established clinical evidence; rather, it complements traditional clinical frameworks by incorporating patient perspectives into evidence-based medicine.

Limitations

Despite the promising results, this study has several limitations. First, our study focused on a single health condition and a single social media platform, which may introduce platform-specific biases and limit generalizability. Second, both the lexicon and the evaluation sample sizes are relatively small for broader conclusions. Third, our findings are exploratory and hypothesis generating and not yet clinically validated.

We also acknowledge specific implementation limitations. Although the LSE pipeline is designed for modularity and adaptability, applying it to other complex health conditions may require empirical adjustments. Determining optimal clusters for symptom identification and establishing semantic mapping rules for symptom normalization would require condition-specific customization. In this study, we used silhouette scores and inertia analysis to determine the optimal cluster metrics, such as the number of clusters (k). However, we recognize the limitation of distance-based metrics due to the curse of dimensionality, which may not fully capture cluster shape or density nuances. Despite this, our bootstrap stability analysis demonstrated consistently high Adjusted Rand Index (ARI) values across various k settings, indicating that our symptom clusters were robust to moderate changes in the number of clusters and to the random removal of data points. The performance of this framework may vary depending on the extracted lexicon due to linguistic variability and the nature of symptom discourse. Future work should focus on validating this method across diverse conditions, platforms, and languages. In addition, we plan to conduct online surveys and focus groups involving both patients and practitioners to assess the clinical relevance of our findings.

Conclusions

This work focused on extracting and analyzing disease-specific symptoms from individuals’ lived experiences with an application to the PCOS use case. The uniqueness of our work lies in several key aspects. The extraction of a symptom-related lexicon ensures the capture of subtle mentions without overlooking informal expressions on social media. We generated disease-specific symptoms by integrating LLMs into the NLP pipeline rather than relying on traditional NLP methods, which often obtain generic symptoms. In addition, analyzing PCOS symptomatology across varied dimensions revealed unique findings, such as emerging new symptoms expressed in patient-centric language, symptom co-occurrence patterns, and associated comorbidities. The outcomes of this work may help eliminate uncertainties surrounding complex medical conditions and contribute to clinical research and public health policies in a community-informed way.

Acknowledgments

This work was partially supported by the Information and Communication Technology Division of Bangladesh under a postgraduate research fellowship grant. The funding body had no involvement in the study design; data collection, analysis, or experimentation; writing of the report; or decision to submit the paper for publication.

Data Availability

We have publicly released our curated Reddit dataset and human-annotated labeled dataset. To support reproducibility, all source code and resources used in this study have also been made available. The datasets and related materials can be accessed through our GitHub repository [71].

Authors' Contributions

BH conceptualized the study, curated the data, designed the methodology, and organized this study; analyzed the results; and wrote, reviewed, and revised the manuscript. SMP and MEA contributed to conceptualization; supervision; validation of the results; and writing, review, and revision of the manuscript. MFR analyzed the results and contributed to writing, reviewing, and revising the manuscript. RA conceptualized the study, validated the results, and reviewed and revised the manuscript.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Supplementary tables, prompts, and results.

DOC File , 797 KB

  1. Kemp S. Digital 2023: global overview report. DataReportal. Jan 26, 2023. URL: https://datareportal.com/reports/digital-2023-global-overview-report [accessed 2025-05-25]
  2. National health interview survey. Centers for Disease Control and Prevention. URL: https://www.cdc.gov/nchs/nhis/2022nhis.htm [accessed 2025-05-25]
  3. Jelodar H, Wang Y, Orji R, Huang S. Deep sentiment classification and topic discovery on novel coronavirus or COVID-19 online discussions: NLP using LSTM recurrent neural network approach. IEEE J Biomed Health Inform. Oct 2020;24(10):2733-2742. [CrossRef] [Medline]
  4. Chopra S, Zehrung R, Shanmugam TA, Choe EK. Living with uncertainty and stigma: self-experimentation and support-seeking around polycystic ovary syndrome. In: Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 2021. Presented at: CHI '21; May 8-13, 2021; Yokohama, Japan. [CrossRef]
  5. Chen LL, Wilson SR, Lohmann S, Negraia DV. What are you anxious about? examining subjects of anxiety during the COVID-19 pandemic. Proceedings of the Int AAAI Conf Web Soc Media. Jun 02, 2023;17(1):137-148. [CrossRef]
  6. Attai DJ, Cowher MS, Al-Hamadani M, Schoger JM, Staley AC, Landercasper J. Twitter social media is an effective tool for breast cancer patient education and support: patient-reported outcomes by survey. J Med Internet Res. Jul 30, 2015;17(7):e188. [FREE Full text] [CrossRef] [Medline]
  7. Jeon J, Baruah G, Sarabadani S, Palanica A. Identification of risk factors and symptoms of COVID-19: analysis of biomedical literature and social media data. J Med Internet Res. Oct 02, 2020;22(10):e20509. [FREE Full text] [CrossRef] [Medline]
  8. Sendra A, Farré J, Vaagan RW. Seeking, sharing and co-creating: a systematic review of the relation between social support theory, social media use and chronic diseases. Soc Theory Health. May 27, 2019;18:317-339. [CrossRef]
  9. McKillop M, Mamykina L, Elhadad N. Designing in the dark: eliciting self-tracking dimensions for understanding enigmatic disease. In: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. 2018. Presented at: CHI '18; April 21-26, 2018; Montreal, QC. [CrossRef]
  10. Lentscher JA, Decherney AH. Clinical presentation and diagnosis of polycystic ovarian syndrome. Clin Obstet Gynecol. Mar 01, 2021;64(1):3-11. [FREE Full text] [CrossRef] [Medline]
  11. Porta M, Greenland S, Hernán M, dos Santos Silva I, Last JM. A Dictionary of Epidemiology. Oxford, UK. Oxford University Press; 2014.
  12. Romano W, Sharif O, Basak M, Gatto J, Preum SM. Theme-driven keyphrase extraction to analyze social media discourse. Proc Int AAAI Conf Web Soc Media. May 28, 2024;18(1):1315-1327. [CrossRef]
  13. Ohno Y, Aomori T, Nishiyama T, Kato R, Fujiki R, Ishikawa H, et al. Performance improvement of a natural language processing tool for extracting patient narratives related to medical states from Japanese pharmaceutical care records by increasing the amount of training data: natural language processing analysis and validation study. JMIR Med Inform. Mar 04, 2025;13:e68863. [FREE Full text] [CrossRef] [Medline]
  14. Molenaar A, Lukose D, Brennan L, Jenkins EL, McCaffrey TA. Using natural language processing to explore social media opinions on food security: sentiment analysis and topic modeling study. J Med Internet Res. Mar 21, 2024;26:e47826. [FREE Full text] [CrossRef] [Medline]
  15. Zhang T, Schoene AM, Ji S, Ananiadou S. Natural language processing applied to mental illness detection: a narrative review. NPJ Digit Med. Apr 08, 2022;5(1):46. [FREE Full text] [CrossRef] [Medline]
  16. Sharif O, Gatto J, Basak M, Preum SM. Explicit, implicit, and scattered: revisiting event extraction to capture complex arguments. In: Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing. 2024. Presented at: EMNLP 2024; November 12-16, 2024; Miami, FL. [CrossRef]
  17. Guo Y, Ovadje A, Al-Garadi MA, Sarker A. Evaluating large language models for health-related text classification tasks with public social media data. J Am Med Inform Assoc. Oct 01, 2024;31(10):2181-2189. [CrossRef] [Medline]
  18. Agarwal S, Searle T, Ratas M, Shek A, Teo J, Dobson R. Improving extraction of clinical event contextual properties from electronic health records: a comparative study. ArXiv. Preprint posted online on August 30, 2024. [CrossRef]
  19. Bai L, Mulvenna MD, Wang Z, Bond R. Clinical entity extraction: comparison between MetaMap, cTAKES, CLAMP and Amazon comprehend medical. In: Proceedings of the 32nd Irish Signals and Systems Conference. 2021. Presented at: ISSC 2021; June 10-11, 2021; Athlone, Ireland. URL: https://doi.org/10.1109/ISSC52156.2021.9467856 [CrossRef]
  20. Peng J, Zhao M, Havrilla J, Liu C, Weng C, Guthrie W, et al. Natural language processing (NLP) tools in extracting biomedical concepts from research articles: a case study on autism spectrum disorder. BMC Med Inform Decis Mak. Dec 30, 2020;20(Suppl 11):322. [FREE Full text] [CrossRef] [Medline]
  21. Agrawal M, Hegselmann S, Lang H, Kim Y, Sontag D. Large language models are few-shot clinical information extractors. ArXiv. Preprint posted online on May 25, 2022. [FREE Full text] [CrossRef]
  22. Sorbello A, Haque SA, Hasan R, Jermyn R, Hussein A, Vega A, et al. Artificial intelligence-enabled software prototype to inform opioid pharmacovigilance from electronic health records: development and usability study. JMIR AI. 2023;2:e45000. [FREE Full text] [CrossRef] [Medline]
  23. Luo X, Gandhi P, Storey S, Huang K. A deep language model for symptom extraction from clinical text and its application to extract COVID-19 symptoms from social media. IEEE J Biomed Health Inform. Apr 2022;26(4):1737-1748. [FREE Full text] [CrossRef] [Medline]
  24. Dolatabadi E, Moyano D, Bales M, Spasojevic S, Bhambhoria R, Bhatti J, et al. Using social media to help understand patient-reported health outcomes of post-COVID-19 condition: natural language processing approach. J Med Internet Res. Sep 19, 2023;25:e45767. [FREE Full text] [CrossRef] [Medline]
  25. Ding Q, Massey D, Huang C, Grady CB, Lu Y, Cohen A, et al. Tracking self-reported symptoms and medical conditions on social media during the COVID-19 pandemic: infodemiological study. JMIR Public Health Surveill. Sep 28, 2021;7(9):e29413. [FREE Full text] [CrossRef] [Medline]
  26. Babaian T, Xu J. Entity recognition from colloquial text. Decis Support Syst. Apr 2024;179:114172. [FREE Full text] [CrossRef]
  27. Guo M, Ma Y, Eworuke E, Khashei M, Song J, Zhao Y, et al. Identifying COVID-19 cases and extracting patient reported symptoms from Reddit using natural language processing. Sci Rep. Aug 22, 2023;13(1):13721. [FREE Full text] [CrossRef] [Medline]
  28. Goel A, Gueta A, Gilon O, Liu C, Erell S, Nguyen LH, et al. LLMs accelerate annotation for medical information extraction. ArXiv. Preprint posted online on December 4, 2023. [CrossRef]
  29. Wei WI, Leung CL, Tang A, McNeil EB, Wong SY, Kwok KO. Extracting symptoms from free-text responses using ChatGPT among COVID-19 cases in Hong Kong. Clin Microbiol Infect. Jan 2024;30(1):142.e1-142.e3. [FREE Full text] [CrossRef] [Medline]
  30. Li Y, Viswaroopan D, He W, Li J, Zuo X, Xu H, et al. Improving entity recognition using ensembles of deep learning and fine-tuned large language models: a case study on adverse event extraction from VAERS and social media. J Biomed Inform. Mar 2025;163:104789. [FREE Full text] [CrossRef] [Medline]
  31. Polycystic ovary syndrome. World Health Organization. Feb 07, 2025. URL: https://www.who.int/news-room/fact-sheets/detail/polycystic-ovary-syndrome [accessed 2025-05-25]
  32. Escobar-Morreale HF. Polycystic ovary syndrome: definition, aetiology, diagnosis and treatment. Nat Rev Endocrinol. May 2018;14(5):270-284. [CrossRef] [Medline]
  33. Rowlands IJ, Teede H, Lucke J, Dobson AJ, Mishra GD. Young women's psychological distress after a diagnosis of polycystic ovary syndrome or endometriosis. Hum Reprod. Sep 13, 2016;31(9):2072-2081. [CrossRef] [Medline]
  34. Malhotra K, Kempegowda P. Appraising unmet needs and misinformation spread about polycystic ovary syndrome in 85,872 YouTube comments over 12 years: big data infodemiology study. J Med Internet Res. Sep 11, 2023;25:e49220. [FREE Full text] [CrossRef] [Medline]
  35. Naroji S, John J, Gomez-Lobo V. Understanding PCOS-related content across social media platforms-a cross-sectional analysis. J Pediatr Adolesc Gynecol. Apr 2024;37(2):142-148. [CrossRef] [Medline]
  36. Cruz J, Qazi Z, Vasireddy AM, Vilchez-Russell KA, Stuparich MA, Nahas S, et al. Instagram cysters: what’s available on polycystic ovarian syndrome? Fertil Steril. Oct 2022;118(4):e323. [FREE Full text]
  37. Proferes N, Jones N, Gilbert S, Fiesler C, Zimmer M. Studying Reddit: a systematic overview of disciplines, approaches, methods, and ethics. Soc Media Soc. May 26, 2021;7(2). [CrossRef]
  38. Boe B. PRAW: the python Reddit API wrapper. GitHub. 2020. URL: https://github.com/praw-dev/praw/ [accessed 2025-08-13]
  39. Baumgartner J. Pushshift Reddit API Documentation. 2018. URL: https://reddit-api.readthedocs.io/ [accessed 2025-08-27]
  40. Patel D, Timsina P, Raut G, Freeman R, Levin MA, Nadkarni GN, et al. Exploring temperature effects on large language models across various clinical tasks. medRxiv. Preprint posted online on July 22, 2024. [FREE Full text] [CrossRef]
  41. Hu Y, Chen Q, Du J, Peng X, Keloth VK, Zuo X, et al. Improving large language models for clinical named entity recognition via prompt engineering. J Am Med Inform Assoc. Sep 01, 2024;31(9):1812-1820. [CrossRef] [Medline]
  42. Ouyang L, Wu J, Jiang X, Almeida D, Wainwright CL, Mishkin P, et al. Training language models to follow instructions with human feedback. In: Proceedings of the 36th International Conference on Neural Information Processing Systems. 2022. Presented at: NIPS'22; November 28-December 9, 2022; New Orleans, LA. URL: https://dl.acm.org/doi/10.5555/3600270.3602281 [CrossRef]
  43. Leas EC, Ayers JW, Desai N, Dredze M, Hogarth M, Smith DM. Using large language models to support content analysis: a case study of ChatGPT for adverse event detection. J Med Internet Res. May 02, 2024;26:e52499. [FREE Full text] [CrossRef] [Medline]
  44. Kuo HY, Chen SY. Predicting user engagement in health misinformation correction on social media platforms in Taiwan: content analysis and text mining study. J Med Internet Res. Jan 23, 2025;27:e65631. [FREE Full text] [CrossRef] [Medline]
  45. Mohr I, Wuhrl A, Klinger R. CoVERT: a corpus of fact-checked biomedical COVID-19 tweets. In: Proceedings of the 13th Conference on Language Resources and Evaluation. 2022. Presented at: LREC 2022; June 20-25, 2022; Marseille, France. URL: https://aclanthology.org/2022.lrec-1.26.pdf
  46. Herman Bernardim Andrade G, Nishiyama T, Fujimaki T, Yada S, Wakamiya S, Takagi M, et al. Assessing domain adaptation in adverse drug event extraction on real-world breast cancer records. Int J Med Inform. Nov 2024;191:105539. [FREE Full text] [CrossRef] [Medline]
  47. Sarwar TB, Noor NM, Saef Ullah Miah M. Evaluating keyphrase extraction algorithms for finding similar news articles using lexical similarity calculation and semantic relatedness measurement by word embedding. PeerJ Comput Sci. 2022;8:e1024. [FREE Full text] [CrossRef] [Medline]
  48. Devlin J, Chang MW, Lee K, Toutanova K. Bert: pre-training of deep bidirectional transformers for language understanding. ArXiv. Preprint posted online on October 11, 2018. [FREE Full text] [CrossRef]
  49. Wang S, Thompson L, Iyyer M. Phrase-BERT: improved phrase embeddings from BERT with an application to corpus exploration. ArXiv. Preprint posted online on September 13, 2021. [FREE Full text] [CrossRef]
  50. Lee J, Yoon W, Kim S, Kim D, Kim S, So CH, et al. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics. Feb 15, 2020;36(4):1234-1240. [FREE Full text] [CrossRef] [Medline]
  51. Shahapure KR, Nicholas C. Cluster quality analysis using silhouette score. In: Proceedings of the IEEE 7th International Conference on Data Science and Advanced Analytics. 2020. Presented at: DSAA 2020; October 6-9, 2020; Sydney, Australia. [CrossRef]
  52. Moulavi D, Jaskowiak PA, Campello RJ, Zimek A, Sander J. Density-based clustering validation. In: Proceedings of the 2014 SIAM International Conference on Data Mining. 2014. Presented at: SIAM 2014; April 24-26, 2014; Philadelphia, PA. [CrossRef]
  53. Meng Y, Huang J, Wang G, Zhang C, Zhuang H, Kaplan L, et al. Spherical text embedding. In: Proceedings of the 33rd International Conference on Neural Information Processing Systems. 2019. Presented at: NIPS'19; December 8-14, 2019; Vancouver, BC.
  54. Risch J, Möller T, Gutsch J, Pietsch M. Semantic answer similarity for evaluating question answering models. In: Proceedings of the 3rd Workshop on Machine Reading for Question Answering. 2021. Presented at: MRQA 2021; November 10, 2021; Online. [CrossRef]
  55. Liu J, Gupta S, Chen A, Wang CK, Mishra P, Dai HJ, et al. OpenDeID pipeline for unstructured electronic health record text notes based on rules and transformers: deidentification algorithm development and validation study. J Med Internet Res. Dec 06, 2023;25:e48145. [FREE Full text] [CrossRef] [Medline]
  56. Mayo clinic. Mayo Foundation for Medical Education and Research. URL: https://www.mayoclinic.org/ [accessed 2025-05-25]
  57. A Visual Guide to PCOS. WebMD. Jun 24, 2025. URL: https://www.webmd.com/women/ss/slideshow-pcos-overview [accessed 2025-08-27]
  58. Free ICD-10-CM data. ICD10Data.com. URL: https://www.icd10data.com/ICD10CM [accessed 2025-05-25]
  59. Polycystic ovary syndrome (PCOS). Johns Hopkins University. URL: https://www.hopkinsmedicine.org/health/conditions-and-diseases/polycystic-ovary-syndrome-pcos [accessed 2025-05-25]
  60. Polycystic ovary syndrome. National Health Service. URL: https://www.nhs.uk/conditions/polycystic-ovary-syndrome-pcos/ [accessed 2025-05-25]
  61. Diabetes and polycystic ovary syndrome (PCOS). Centers for Disease Control and Prevention. May 15, 2024. URL: https://www.cdc.gov/diabetes/risk-factors/pcos-polycystic-ovary-syndrome.html [accessed 2025-05-25]
  62. Amazon comprehend medical. Amazon Web Services. URL: https://aws.amazon.com/comprehend/medical/ [accessed 2025-05-25]
  63. Cloud healthcare API. Google Cloud. URL: https://cloud.google.com/healthcare-api?hl=en [accessed 2025-05-25]
  64. Yeh A. More accurate tests for the statistical significance of result differences. In: Proceedings of the COLING 2000 Volume 2: The 18th International Conference on Computational Linguistics. 2000. Presented at: COLING 2000; July 31-August 4, 2000; Saarbrücken, Germany. [CrossRef]
  65. Huddleston HG, Jaswa EG, Casaletto KB, Neuhaus J, Kim C, Wellons M, et al. Associations of polycystic ovary syndrome with indicators of brain health at midlife in the CARDIA cohort. Neurology. Feb 27, 2024;102(4). [CrossRef]
  66. Singla R, Gupta Y, Khemani M, Aggarwal S. Thyroid disorders and polycystic ovary syndrome: an emerging relationship. Indian J Endocrinol Metab. 2015;19(1):25-29. [FREE Full text] [CrossRef] [Medline]
  67. Eftekhar T, Sohrabvand F, Zabandan N, Shariat M, Haghollahi F, Ghahghaei-Nezamabadi A. Sexual dysfunction in patients with polycystic ovary syndrome and its affected domains. Iran J Reprod Med. Aug 2014;12(8):539-546. [FREE Full text] [Medline]
  68. Kałużna M, Kompf P, Wachowiak-Ochmańska K, Moczko J, Królczyk A, Janicki A, et al. Are patients with polycystic ovary syndrome more prone to irritable bowel syndrome? Endocr Connect. 2022;11(4):e210309. [FREE Full text] [CrossRef]
  69. Jeanes YM, Reeves S, Gibson EL, Piggott C, May VA, Hart KH. Binge eating behaviours and food cravings in women with polycystic ovary syndrome. Appetite. Feb 01, 2017;109:24-32. [CrossRef] [Medline]
  70. Staehelin D, Dolata M, Stöckli L, Schwabe G. How patient-generated data enhance patient-provider communication in chronic care: field study in design science research. JMIR Med Inform. Sep 10, 2024;12:e57406. [FREE Full text] [CrossRef] [Medline]
  71. Subreddit to symptomatology data and code. GitHib. URL: https://github.com/bush1987/LSE-Symptomatology [accessed 2025-08-02]


ACM: Amazon Comprehend Medical
API: application programming interface
BERT: bidirectional encoder representations from transformers
DBCV: density-based clustering validation
GHNLP: Google Healthcare Natural Language Application Programming Interface
HDBSCAN: hierarchical density-based spatial clustering of applications with noise
JI: Jaccard index
LLM: large language model
LSE: lexicon-based symptom extraction
NLP: natural language processing
PCOS: polycystic ovary syndrome


Edited by A Castonguay; submitted 06.01.25; peer-reviewed by S Hwang, R Sauvayre, S Sivarajkumar; comments to author 26.04.25; revised version received 25.05.25; accepted 09.07.25; published 12.09.25.

Copyright

©Bushra Hossain, Sarah M Preum, Md Fazle Rabbi, Rifat Ara, Mohammed Eunus Ali. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 12.09.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.