Published on in Vol 13 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/71687, first published .
Detecting Redundant Health Survey Questions by Using Language-Agnostic Bidirectional Encoder Representations From Transformers Sentence Embedding: Algorithm Development Study

Detecting Redundant Health Survey Questions by Using Language-Agnostic Bidirectional Encoder Representations From Transformers Sentence Embedding: Algorithm Development Study

Detecting Redundant Health Survey Questions by Using Language-Agnostic Bidirectional Encoder Representations From Transformers Sentence Embedding: Algorithm Development Study

Authors of this article:

Sunghoon Kang1 Author Orcid Image ;   Hyewon Park1 Author Orcid Image ;   Ricky Taira2 Author Orcid Image ;   Hyeoneui Kim1 Author Orcid Image

1College of Nursing, Seoul National University, 103 Daehak-ro, Jongno-gu, Seoul, Republic of Korea

2The Department of Radiological Sciences, David Geffen School of Medicine, University of California, Los Angeles, Los Angeles, CA, United States

Corresponding Author:

Hyeoneui Kim, RN, PhD


Background: As the importance of person-generated health data (PGHD) in health care and research has increased, efforts have been made to standardize survey-based PGHD to improve its usability and interoperability. Standardization efforts such as the Patient-Reported Outcomes Measurement Information System (PROMIS) and the National Institutes of Health (NIH) Common Data Elements (CDE) repository provide effective tools for managing and unifying health survey questions. However, previous methods using ontology-mediated annotation are not only labor-intensive and difficult to scale but also challenging for identifying semantic redundancies in survey questions, especially across multiple languages.

Objective: The goal of this work was to compute the semantic similarity among publicly available health survey questions to facilitate the standardization of survey-based PGHD.

Methods: We compiled various health survey questions authored in both English and Korean from the NIH CDE repository, PROMIS, Korean public health agencies, and academic publications. Questions were drawn from various health lifelog domains. A randomized question pairing scheme was used to generate a semantic text similarity dataset consisting of 1758 question pairs. The similarity scores between each question pair were assigned by 2 human experts. The tagged dataset was then used to build 4 classifiers featuring bag-of-words, sentence-bidirectional encoder representations from transformers (SBERT) with bidirectional encoder representations from transformers (BERT)–based embeddings, SBERT with language-agnostic BERT sentence embedding (LaBSE), and GPT-4o. The algorithms were evaluated using traditional contingency statistics.

Results: Among the 3 algorithms, SBERT-LaBSE demonstrated the highest performance in assessing the question similarity across both languages, achieving area under the receiver operating characteristic and precision-recall curves of >0.99. Additionally, SBERT-LaBSE proved effective in identifying cross-lingual semantic similarities. The SBERT-LaBSE algorithm excelled at aligning semantically equivalent sentences across both languages but encountered challenges in capturing subtle nuances and maintaining computational efficiency. Future research should focus on testing with larger multilingual datasets and on calibrating and normalizing scores across the health lifelog domains to improve consistency.

Conclusions: This study introduces the SBERT-LaBSE algorithm for calculating the semantic similarity across 2 languages, showing that it outperforms BERT-based models, the GPT-4o model, and the bag-of-words approach, highlighting its potential in improving the semantic interoperability of survey-based PGHD across language barriers.

JMIR Med Inform 2025;13:e71687

doi:10.2196/71687

Keywords



Person-generated health data (PGHD) is becoming increasingly important in managing individual health. PGHD encompass health-related information that individuals create and collect outside traditional clinical environments, helping them monitor and manage their well-being [1,2]. Examples of PGHD include biometric data from wearable devices and self-reported information such as patient-reported outcomes. Since PGHD has the potential for continuously capturing health insights beyond health care settings, there is growing interest in leveraging PGHD to support clinical care [3-5]. In parallel, PGHD is increasingly explored as a resource for patient-centered outcomes research [6,7]. However, there are several challenges in the effective use of PGHD, including developing robust data management systems, ensuring data security, deploying it seamlessly into clinical workflows, and maintaining high data quality [5,7,8].

Standardizing survey-based PGHD is a critical step in enabling its broader use [9]. An important aspect of standardization is to identify redundancies in the form of semantic equivalencies. These redundancies may arise because the clarity, tone, tense, directness, and formality of the language can be phrased differently for the same purposeful inquiry depending upon the author. For example, emotional symptoms may be captured by questions such as “Do you feel like withdrawing from family or friends?” or “I don’t really want to talk to people around me.” This variation makes identifying semantically equivalent questions—and thus standardizing survey-based PGHD—a complex task. Efforts such as the Patient-Reported Outcomes Measurement Information System (PROMIS) and the National Institutes of Health (NIH) Common Data Elements (CDE) repository aim to provide standardized health survey questions. PROMIS, a consensus-based item bank designed for managing patient-reported outcomes, offers standardized measures that are applicable across various diseases and clinical settings [10-12]. These measures have helped health care providers across various clinical settings, including pain management [13], orthopedics [14], and primary care [15]; in cancer care [16]; in managing patient symptoms; in tailoring treatments; and in improving communication between patients and clinicians. The NIH CDE repository, through metadata tagging, also plays a key role in standardizing data elements, including health surveys [17,18]. Both PROMIS and the CDE repository are essential for enhancing the interoperability of health data.

In practice, the deployment of PGHD acquisition applications requires that survey questions be drawn from these established standardized resources. Data collected using questions outside of these resources still require additional efforts to achieve standardization. Although previous studies have explored ontology-mediated methods to identify semantically equivalent health questions [10,11], annotating each question with ontology concepts is labor-intensive and lacks scalability as such knowledge sources expand. As a complementary approach, deep learning and transformer-based methods have been applied to semantic textual similarity (STS) tasks in clinical texts, including radiology and pathology reports [19], clinical notes [20-22], and medical question-answer pairs [23]. A range of models has been explored, such as convolutional neural networks [19]; transformer-based architectures such as bidirectional encoder representations from transformers (BERT), robustly optimized BERT approach, and XLNet [20-22]; and the Siamese network [23]. Despite their promising performance, most of these models have been limited to single-language settings—predominantly English [20-22] or Chinese [19,23]. Consequently, cross-lingual STS remains underexplored, highlighting the need for standardization efforts that promote semantic interoperability across languages.

To address these challenges, we developed Standardized PGHD Utilization Resources and Tools (SPURT), which supports the standardization and reuse of survey-based PGHD by identifying semantically equivalent questions and facilitating the storage, retrieval, and sharing of these data. Unlike PROMIS and the NIH CDE repository, SPURT annotates and stores health survey questions in both English and Korean while also detecting semantically redundant questions. This ensures the use of consistent question formats whenever possible. Technically, assessing semantic similarity between texts is well-established and widely applied for managing text resources [24]. However, SPURT faces 2 unique challenges in its assigned task. First, it must effectively assess semantic similarities within or between 2 different languages—English and Korean. Although multilingual embeddings can be used to address this challenge [25,26], they often perform less effectively for low-resource languages such as Korean compared to high-resource languages such as English [27]. One common solution is to translate low-resource languages into high-resource ones before embedding, but this approach risks losing or distorting the original meaning [28,29]. Second, it must ensure computational efficiency for real-time semantic comparisons between questions. Calculating semantic similarity by using large language models such as BERT is computationally expensive, with a time complexity of O (N!). For example, computing the similarity of approximately 10,000 sentence pairs can take around 65 hours using a V100 graphics processing unit [30]. Given that SPURT is designed to be a real-time, reactive data processing tool, achieving reasonable response times is crucial for its functionality.

This study presents the development of a novel algorithm for detecting redundant questions, addressing the challenges outlined above. The algorithm utilizes sentence-BERT (SBERT), a variant of BERT designed for efficient sentence-level semantic similarity calculations [30] along with language-agnostic BERT sentence embedding (LaBSE) [31] to enhance multilingual capability. Sentence-BERT is a model specifically designed for calculating STS between sentences, and LaBSE is an embedding that supports efficient cross-lingual STS by mapping multilingual sentences into a shared embedding space. The SBERT-LaBSE algorithm integrates the strengths of both models and facilitates the identification of semantically equivalent questions across languages.


Corpus Description: The STS Dataset

An STS dataset contains text pairs along with predefined similarity scores that quantify their semantic closeness [32-36]. This study shows an STS dataset that fine-tunes pretrained language models and evaluates our algorithms’ performance in determining the semantic similarity between health-related questions.

We collected English and Korean questions from self-reported questionnaires covering 5 health lifelog domains, that is, diet, physical activity, living environment, stress management, and sleep. English questions (n=1222) were sourced from the NIH CDE repository, PROMIS, and academic publications, while Korean questions (n=963) were gathered from web-based resources provided by public health agencies and hospitals in Korea [17,37-40].

To build the STS dataset, we began by randomly selecting 5 seed questions from each of the 5 health lifelog domains in Korean, resulting in 25 seed questions. For each question, correspondingly similar questions for Korean were identified, resulting in 25 similar seed questions for each language. This correspondence of seed questions was performed to minimize the effects of semantic complexity on algorithm performance. We then randomly selected 30 comparison questions for each seed question, which yielded a total of 1500 question pairs (750 in each language).

The gold standard for semantic similarity between the question pairs was determined by 2 researchers with nursing backgrounds who independently scored the similarity of each question pair, following a standardized scoring protocol (Table 1). The agreement between the researchers, as measured by Cohen κ, varied by the health lifelog domains: 0.91 for diet, 0.72 for living environment, 0.83 for physical activity, 0.86 for sleep, and 1.0 for stress management, with an average Cohen κ of 0.86 across all the health lifelog domains.

Table 1. Scoring protocol for semantic similarity. The seed question was “In the past month, have you ever had chest pain when you were not performing any physical activity?”
ScoreScoring protocolExamples
4Minor differences in word choice from the seed question but takes the same form of responseIn the past month, have you had chest pain when you were not doing physical activity?
3Share the same key topic, although some details may be added, altered, or omitted from the seed questionDo you feel pain in your chest when you do physical activity?
2The key topics are similar but more specific or general than that of the seed questionHas your doctor ever said that you have a heart condition and that you should only perform physical activity recommended by a doctor?
1Does not share the core topic from the seed question or belongs to a completely different health lifelog domainHave you done general conditioning exercises in the past 4 weeks?

Upon completion of this annotation process, we observed that the initial distribution of the similarity scores was imbalanced—skewed heavily toward lower similarity scores. Only 2.3% (7/300) of the pairs received a score of 4, and 4.9% (15/300) received a score of 3. To address this imbalance, we supplemented the dataset with an additional 117 English and 142 Korean question pairs from other sources, chosen to increase the frequency of semantically similar (ie, higher scores) samples in the evaluation STS dataset. These additions brought the final evaluation set to 820 question pairs (410 in each language) with the following distribution: 12.2% (99/810), 30.5% (247/810), 26.8% (217/810), and 30.5% (247/810) for scores 4, 3, 2, and 1, respectively.

Using a similar procedure as described above, we compiled a second English STS dataset for fine-tuning our pretrained language models. This fine-tuning dataset included 938 annotated English question pairs. The fine-tuning set had a distribution of 6.2% (58/938) scoring 4, 14% (131/938) scoring 3, 23.5% (220/938) scoring 2, and 56.4% (529/938) scoring 1.

In total, the STS dataset consisted of 1758 question pairs, broken down into 820 for evaluation testing (410 English and 410 Korean) and 938 in English for classifier model refinement (see Multimedia Appendix 1). The process of constructing the STS dataset is illustrated in Figure 1.

Figure 1. The process of preparing the semantic textual similarity dataset for fine-tuning and evaluation. CDE: common data elements; NIH: National Institutes of Health; PGHD: person-generated health data; PROMIS: Patient-Reported Outcomes Measurement Information System; Rand: random selection; STS: semantic textual similarity.

Semantic Similarity Calculation Algorithms

Overview

We developed 4 classifiers to compare their performance capability for distinguishing the binary task of semantic similarity between STS question pairs. These were (1) the bag-of-words (BoW) model, (2) SBERT with BERT-based embeddings (SBERT-BERT), (3) LaBSE, and (4) the GPT-4o model (GPT-4o). Among these, the SBERT-BERT algorithm was included to serve as a translation-dependent baseline, enabling comparison with multilingual models such as SBERT-BERT and GPT-4o. Model fine-tuning and algorithm development were performed using Python (version 3.11).

BoW Classifier

The BoW algorithm, a traditional language model that represents sentences by their word frequency, serves as the baseline [41]. The BoW model’s vocabulary was derived from the STS dataset, comprising 1349 unique word forms after stop-word removal and lemmatization. Each sentence was represented as a 1349D vector based on the vocabulary. Cosine similarity was used to calculate the semantic distance of the question pairs. For Korean questions, translation to English was performed using the Google Translator application programming interface prior to similarity calculation [42].

The SBERT-BERT Algorithm

The SBERT-BERT large language model was derived from the pretrained BERT-based model, which has 12 layers, a 768D hidden layer, 12 attention heads, and 110 million parameters [30]. SBERT-BERT supports only English. We fine-tuned the pretrained SBERT-BERT model to optimize its performance for identifying semantic equivalency among health questions by using the 938 English question pairs described above. The fine-tuning was performed with a batch size of 32, 8 epochs, and a learning rate of 2e-5, which were deemed optimal after testing various configurations. The AdamW optimizer was used for model optimization [43]. The fine-tuned SBERT-BERT algorithm was then evaluated using the test STS dataset of 410 English question pairs and 410 Korean question pairs. As previously stated, the Korean questions were translated into English using the Google Translator application programming interface to execute the evaluation.

The SBERT-LaBSE Algorithm

The SBERT-LaBSE algorithm differs from SBERT-BERT in that it supports multiple languages within a single embedding space [31]. The pretrained SBERT-LaBSE model was derived from the LaBSE model, which also consists of 12 layers, a 768D hidden layer, 12 attention heads, and 110 million parameters [31]. Fine-tuning was performed in the same manner as for SBERT-BERT. Unlike the other models, SBERT-LaBSE can assess the semantic similarity of English and Korean questions without requiring translation.

The GPT-4o Algorithm

The GPT-4o model, a state-of-the-art large language model, is designed to understand and generate text in multiple languages, including English and Korean [44]. Unlike the SBERT-BERT and SBERT-LaBSE, which rely on fixed embeddings for similarity calculation, the GPT-4o operates as a generative model that dynamically evaluates semantic similarity based on contextual understanding. However, in this study, we utilized GPT-4o in a deterministic manner to predict the score of sentence pairs. Each sentence pair was presented with a specific instruction asking to evaluate the score on a scale from 1 to 4 (Multimedia Appendix 2). Fine-tuning of the GPT-4o model was conducted using the fine-tuning application programming interface from the OpenAI platform [45].

Performance Evaluation

The performance of the similarity calculation algorithms was evaluated as a binary classification problem to simplify interpretation. The 4-point ordinal similarity scores from the STS dataset were converted into binary labels, where scores of 3 and 4 were categorized as similar and scores of 1 and 2 as dissimilar.

Optimal thresholds for predicting similarity were determined for the continuous similarity scores, which ranged from –1 to 1. Precision, recall, and F1-scores were calculated to assess algorithm performance, and the area under the curve for both the receiver operating characteristic and precision-recall curves were examined. The processes used by the 3 algorithms to calculate similarity are illustrated in Figure 2.

Figure 2. Similarity calculation with the 4 algorithms. API: application programming interface; BERT: bidirectional encoder representations from transformers; BoW: bag-of-words; Eng: English; Kor: Korean; LaBSE: language-agnostic bidirectional encoder representations from transformers sentence embedding; SBERT: sentence-bidirectional encoder representations from transformers.

Ethical Considerations

This study does not involve human participants, intervention, or identifiable private information. The analysis was based on publicly available and nonidentifiable health survey questions from open repositories and published sources. As such, it does not fall under the scope of human subject research as defined by the Seoul National University institutional review board. According to Article 2 and Article 8, Paragraph 2 of the Seoul National University institutional review board regulations (regulation 27, effective September 11, 2023), studies that do not involve human participants or human-derived materials are exempt from institutional review board review. Therefore, this study was not submitted for ethical review. No informed consent, compensation, or privacy protection measures were applicable, as no human participants were involved, and no personal data were collected or analyzed.


The performance of the 3 models for classifying similar versus dissimilar question pairs when aggregating the 5 health lifelog domains is summarized in Table 2 and Figure 3. In the zero-shot trials (ie, without the model refining stage), there were minimal differences in performance among the 3 algorithms for both English and Korean questions. All algorithms exhibited higher recall than precision in both languages. After fine-tuning, the SBERT-BERT algorithm showed substantial improvement, particularly for English questions, in which the F1-score increased from 0.65 to 0.96. For Korean questions, the improvement was moderate, with the F1-score progressing from 0.68 to 0.73. In contrast, SBERT-LaBSE demonstrated significant improvements for both languages post fine-tuning. For English questions, the F1-scores increased from 0.66 to 0.98, while for Korean, the F1-scores increased from 0.68 to 0.98. Fine tuning for both SBERT-BERT and SBERT-LaBSE models resulted in noticeable balanced performance between recall and precision. Similarly, GPT-4o exhibited improved performance following fine-tuning, with its F1-scores increasing from 0.69 to 0.79 for the English questions and from 0.67 to 0.79 for the Korean questions. However, the degree of improvement was smaller than that observed in SBERT models.

Table 2. Performance metrics for the 3 algorithms, combining the health lifelog domains.
Performance
metrics
BoWaGPT-4o pretrainedGPT-4o fine-tunedSBERTb with pretrainedSBERT with fine-tuned
BERT-baseLaBSEcBERT-baseLaBSE
English question pairs (n=410)
Accuracy0.61120.66830.84630.63080.59170.97020.9853
Precision0.52790.57530.95900.54510.51110.96680.9818
Recall0.81610.85140.66860.79890.92530.96320.9839
F1-score0.64110.68660.78790.64800.65850.96490.9828
Korean question pairs (n=410)
Accuracy0.66100.65120.84880.66590.68780.75760.9839
Precision0.57320.56200.95200.57600.60540.69290.9818
Recall0.80570.82860.68000.82290.77140.78170.9806
F1-score0.66980.66970.79330.67760.67840.73320.9812

aBoW: bag-of-words.

bSBERT: sentence-bidirectional encoder representations from transformers.

cLaBSE: language-agnostic bidirectional encoder representations from transformers sentence embedding.

Figure 3. Receiver operating characteristic and precision-recall curves for pretrained and fine-tuned embeddings on English and Korean questions, combining the health lifelog domains. AUC: area under the curve; BERT: bidirectional encoder representations from transformers; BoW: bag-of-words; Eng: English; Kor: Korean; LaBSE: language-agnostic bidirectional encoder representations from transformers sentence embedding; PR: precision-recall; ROC: receiver operating characteristic; SBERT: sentence-bidirectional encoder representations from transformers.

Table 3 presents the performance of the 2 SBERT algorithms across the 5 health lifelog domains. For all the health lifelog domains, the fine-tuned SBERT-BERT and SBERT-LaBSE algorithms demonstrated high performance on English questions, with receiver operating characteristic and precision-recall area under the curve values exceeding 0.95 and approaching 0.99. However, the SBERT-BERT algorithm struggled with the English-translated Korean questions, particularly in the physical activity domain. In contrast, the SBERT-LaBSE algorithm consistently delivered strong performance across all the health lifelog domains even for Korean questions.

Table 3. Performance metrics of the sentence-bidirectional encoder representations from transformers–based algorithms with fine-tuned bidirectional encoder representations from transformers and language-agnostic bidirectional encoder representations from transformers sentence embedding models by the health lifelog domains.
Performance
metrics
English question pairs (n=410)Korean question pairs (n=410)
DLa
(n=80)
HLEb
(n=80)
PAc
(n=80)
Sleep
(n=85)
Stress
(n=85)
AllDL
(n=80)
HLE
(n=80)
PA
(n=80)
Sleep
(n=85)
Stress
(n=85)
All
BoWd
Accuracy0.72150.72500.66250.41180.71760.61120.72500.81250.72500.57650.58820.6610
Precision0.70000.63830.59520.41180.62790.52790.65850.79410.62750.00000.50000.5732
Recall0.61760.85710.71431.00000.77140.81610.77140.77140.91430.00000.82860.8057
F1-score0.65630.73170.64940.58330.69230.64110.71050.78260.74420.00000.62370.6698
ROCe AUCf0.72970.74570.68100.58200.76110.69760.76670.79370.79330.59370.66090.7174
PRg AUC0.72500.67180.63010.50250.68340.60360.73940.73730.75190.44980.59850.6267
GPT-4o fine-tuned
Accuracy0.93670.97500.96250.97650.97650.78730.88750.93750.77500.80000.84710.8293
Precision0.89190.97140.94440.94590.94590.74030.96431.00000.86960.95000.95830.7586
Recall0.97060.97140.97141.00001.00000.77010.77140.85710.57140.54290.65710.8800
F1-score0.92960.97140.95770.97220.97220.75490.85710.92310.68970.69090.77970.8148
ROC AUC0.95980.98380.97270.98630.97570.85240.93400.94440.82950.88230.87690.8737
PR AUC0.91380.96720.94180.96290.93540.77570.89690.93440.75100.79690.81820.7862
SBERTh with fine-tuned BERT-basei
Accuracy0.96460.96250.98000.99060.98350.97020.85250.81250.70250.72000.78820.7576
Precision0.96500.95020.97840.98360.98300.96680.80370.73910.61750.60360.71080.6929
Recall0.95290.96570.97710.99430.97710.96320.88000.89140.86290.95430.82290.7817
F1-score0.95850.95710.97700.98870.97990.96490.83840.80620.71760.73760.76220.7332
ROC AUC0.98590.96980.99230.99290.99360.98670.91250.85630.79010.84110.84620.8412
PR AUC0.98580.94800.99250.98700.99180.98000.90080.79690.76400.81090.82440.8134
SBERT with fine-tuned LaBSE
Accuracy0.98480.99000.98750.99060.99060.98530.97750.99750.98500.98590.98350.9839
Precision0.97160.98890.97280.99440.98890.98180.97190.99440.98360.97750.98860.9818
Recall0.99410.98861.00000.98290.98860.98390.97711.00000.98290.98860.97140.9806
F1-score0.98260.98850.98610.98840.98870.98280.97430.99720.98280.98290.97970.9812
ROC AUC0.99650.99290.99870.99890.99790.99680.98930.99760.99620.99710.99300.9951
PR AUC0.99640.99010.99840.99850.99750.99600.98720.99580.99470.99570.99270.9934

aDL: dietary lifestyle.

bHLE: human living environment.

cPA: physical activity.

dBoW: bag-of-words.

eROC: receiver operating characteristic.

fAUC: area under the curve.

gPR: precision-recall.

hSBERT: sentence-bidirectional encoder representations from transformers.

iBERT: bidirectional encoder representations from transformers.

Table 4 presents the optimal cutoff values for the 3 algorithms. The pretrained SBERT-BERT and SBERT-LaBSE models showed considerable variation in the cutoff values across the 5 health lifelog domains. However, after fine-tuning, these variations decreased, indicating that fine-tuning helped stabilize the algorithms. Despite this improvement, the SBERT-LaBSE algorithm still exhibited more variability in the cutoff values across the health lifelog domains compared to SBERT-BERT, suggesting that further calibration may be required for SBERT-LaBSE. Multimedia Appendix 3 provides example question pairs from each health lifelog domain, along with the similarity scores assigned by human reviewers and predicted by the 3 algorithms.

Table 4. Optimal cutoff for algorithms on bag-of-words and pretrained and fine-tuned SBERT-BERT and SBERT–LaBSE in each health lifelog domain.
Health lifelog
domain
Bag-of-wordsSBERTa with pretrainedSBERT with fine-tuned
BERT-basebLaBSEBERT-baseLaBSEc
English question pairs (n=410)
Dietary lifestyle0.28870.62740.53590.63490.6262
Human living environment0.12910.53690.39650.61510.6425
Physical activity0.31620.36670.48220.63040.6202
Sleep0.00000.67900.24560.66170.6574
Stress0.10540.58170.38070.63590.5958
All0.12910.58160.37960.62780.6091
Korean question pairs (n=410)
Dietary lifestyle0.28870.59900.31030.56390.6568
Human living environment0.25820.54750.56030.56390.7138
Physical activity0.14910.47780.60040.56390.6741
Sleep0.93540.48370.92150.56390.6849
Stress0.10910.66470.44810.56390.6586
All0.13360.53200.57530.56390.6531

aSBERT: sentence-bidirectional encoder representations from transformers.

bBERT: bidirectional encoder representations from transformers.

cLaBSE: language-agnostic bidirectional encoder representations from transformers sentence embedding.

Figure 4 illustrates that SBERT-LaBSE effectively determined semantic similarities between the 2 languages, with slightly better performance in identifying the semantic similarities of English questions relative to the Korean seed questions. The complete results of the cross-language semantic similarity analysis are provided in Multimedia Appendix 4.

Figure 4. Performance of the cross-language semantic similarity determination. AUC: area under the curve; BERT: bidirectional encoder representations from transformers; BoW: bag-of-words; Eng: English; Kor: Korean; LaBSE: language-agnostic bidirectional encoder representations from transformers sentence embedding; PR: precision-recall; ROC: receiver operating characteristic; SBERT: sentence-bidirectional encoder representations from transformers.

Principal Findings

This study demonstrates the utility of large language models for determining semantic similarities among health questions to facilitate the standardization of survey-based health data. Our results indicate that the fine-tuned SBERT algorithms were significantly more effective than the traditional BoW approach in identifying semantic similarities for both English and Korean questions. Furthermore, the SBERT-LaBSE algorithm demonstrated superior performance particularly for Korean questions, suggesting that it is a more effective method than the SBERT-BERT algorithm, which relies on English translation, for assessing semantic similarity in non-English texts. Notably, the SBERT-LaBSE algorithm outperformed the GPT-4o algorithm, particularly in Korean. Although it is possible that the full potential of the GPT-4o algorithm was not realized, the results clearly show that for the specific task examined in this study, the fine-tuned SBERT algorithms achieved better performance than GPT-4o, with significantly lower computational costs [46].

The SBERT-LaBSE algorithm’s success with Korean questions can be attributed to its structural design and the limitations of language translation. Structurally, LaBSE aligns semantically equivalent words or sentences from different languages into a unified embedding space, preserving semantic consistency across languages. This allows for more accurate semantic similarity assessments. In contrast, the SBERT-BERT algorithm’s lower performance with Korean questions may be due to meaning loss or distortion during translation, which disrupts semantic comparisons between languages [28,29]. Although previous studies have noted that LaBSE may struggle with subtle, sentence-level nuances, limiting its performance in fine-grained similarity tasks [47], our study shows that the SBERT-LaBSE algorithm effectively captured the meanings in both English and Korean sentences, outperforming the SBERT-BERT model. However, this finding should be validated with a larger and more diverse dataset that includes a broader range of syntactic features.

Limitations

When implemented in the SPURT system with 1835 questions in the comparison space, the SBERT-LaBSE algorithm evaluated the similarity of a new question in just 0.03 seconds. This was achieved on a Naver Cloud Platform server with 8GB RAM and no graphics processing unit [48]. Despite its impressive performance, LaBSE’s 440 million parameters—4 times that of BERT base—make it a resource-intensive option, potentially increasing costs for complex tasks. This resource demand may limit its applicability on resource-constrained devices such as mobile platforms [49]. To address these limitations, future work will explore techniques such as distillation [50] and the use of small language models [51], with the goal of reducing model size while maintaining performance.

This study has some limitations. First, the cutoff values for the similarity scores were not uniformly calibrated across the 5 health lifelog domains, leading to inconsistencies in how similarity scores were interpreted. For example, the SBERT-LaBSE algorithm assigned a similarity score of 0.7 to the dietary question pair “I’ve binge eaten” and “Do you ever overeat?” and identified them as similar. However, the algorithm correctly identified the human living environment questions, that is, “Have you moved in the past 5 years?” and “In the last 5 years, the number of people in this community has?” as dissimilar while assigning the same similarity score of 0.7 to the pair. These inconsistencies may impact the accurate interpretation of similarity scores, highlighting the need for future work to focus on calibrating and normalizing scores across the health lifelog domains to ensure greater consistency. Second, our evaluation was conducted on a small set of English and Korean question pairs. Future studies should explore the feasibility of applying the SBERT-LaBSE algorithm to a broader range of sentence types from diverse domains. Additionally, by incorporating texts from more diverse languages, future research can investigate the algorithm’s potential to overcome language barriers and facilitate semantic interoperability.

Comparison With Prior Work

Previous methods that relied on metadata tagging [17,18] and ontology-mediated annotation [10,11] were effective in providing structured mappings between concepts, facilitating interoperability. However, they struggled with comparing the meanings of survey questions composed in multiple languages and addressing semantically redundant questions. This study leverages fine-tuned large language models such as SBERT-BERT and SBERT-LaBSE to assess semantic similarity. In particular, the fine-tuned SBERT-LaBSE algorithm demonstrates the potential to enhance semantic interoperability by capturing semantic similarities across multiple languages with high performance.

Conclusion

This study highlights the potential of large language models in identifying semantic redundancy in survey-based PGHD collections. Specifically, the SBERT-LaBSE algorithm excelled in classifying semantic similarity across diverse question formats in 2 languages. Our findings demonstrate that SBERT-LaBSE outperforms the traditional BERT-based algorithm, the GPT-4o algorithm, and the conventional BoW approach in both languages, highlighting its capacity to improve semantic interoperability of PGHD across language barriers.

Acknowledgments

We are thankful to our colleagues Eunyeong Lim and Jeongha Kim for their help with the semantic textual similarity dataset preparation. This study was supported in part by a grant from the National Research Foundation of Korea (grant 2022R1A2C201136011).

Conflicts of Interest

None declared.

Multimedia Appendix 1

Comparison of algorithms' performance on the semantic textual similarity dataset.

DOCX File, 64 KB

Multimedia Appendix 2

The instructions used for GPT-4o semantic similarity evaluation.

DOCX File, 15 KB

Multimedia Appendix 3

Example question pairs with the scores from human review and predictions from the 3 algorithms.

DOCX File, 34 KB

Multimedia Appendix 4

Performance metrics in the cross-language semantic similarity analysis.

DOCX File, 27 KB

  1. Shapiro M, Johnston D, Wald J, Mon D. Patient-generated health data. White paper. Official Website of the Assistant Secretary for Technology Policy/Office of the National Coordinator for Health IT. 2012. URL: https://www.healthit.gov/sites/default/files/rti_pghd_whitepaper_april_2012.pdf [Accessed 2025-05-31]
  2. Patient-generated health data. Official Website of the Assistant Secretary for Technology Policy/Office of the National Coordinator for Health IT. URL: https://www.healthit.gov/topic/scientific-initiatives/pcor/patient-generated-health-data-pghd [Accessed 2025-05-08]
  3. Islind AS, Lindroth T, Lundin J, Steineck G. Shift in translations: data work with patient-generated health data in clinical practice. Health Informatics J. Sep 2019;25(3):577-586. [CrossRef] [Medline]
  4. Sanger PC, Hartzler A, Lordon RJ, et al. A patient-centered system in a provider-centered world: challenges of incorporating post-discharge wound data into practice. J Am Med Inform Assoc. May 2016;23(3):514-525. [CrossRef] [Medline]
  5. Adler-Milstein J, Nong P. Early experiences with patient generated health data: health system and patient perspectives. J Am Med Inform Assoc. Oct 1, 2019;26(10):952-959. [CrossRef] [Medline]
  6. Chung AE, Sandler RS, Long MD, et al. Harnessing person-generated health data to accelerate patient-centered outcomes research: the Crohn’s and Colitis Foundation of America PCORnet Patient Powered Research Network (CCFA Partners). J Am Med Inform Assoc. May 2016;23(3):485-490. [CrossRef] [Medline]
  7. Austin E, Lee JR, Amtmann D, et al. Use of patient-generated health data across healthcare settings: implications for health systems. JAMIA Open. Apr 2020;3(1):70-76. [CrossRef] [Medline]
  8. Abdolkhani R, Gray K, Borda A, DeSouza R. Patient-generated health data management and quality challenges in remote patient monitoring. JAMIA Open. Dec 2019;2(4):471-478. [CrossRef] [Medline]
  9. Hussein R, Crutzen R, Gutenberg J, Kulnik ST, Sareban M, Niebauer J. Patient-generated health data (PGHD) interoperability: an integrative perspective. Stud Health Technol Inform. May 27, 2021;281:228-232. [CrossRef] [Medline]
  10. Cella D, Yount S, Rothrock N, et al. The Patient-Reported Outcomes Measurement Information System (PROMIS): progress of an NIH roadmap cooperative group during its first two years. Med Care. May 2007;45(5 Suppl 1):S3-S11. [CrossRef] [Medline]
  11. Cella D, Riley W, Stone A, et al. The Patient-Reported Outcomes Measurement Information System (PROMIS) developed and tested its first wave of adult self-reported health outcome item banks: 2005-2008. J Clin Epidemiol. Nov 2010;63(11):1179-1194. [CrossRef] [Medline]
  12. Pilkonis PA, Choi SW, Reise SP, et al. Item banks for measuring emotional distress from the Patient-Reported Outcomes Measurement Information System (PROMIS®): depression, anxiety, and anger. Assessment. Sep 2011;18(3):263-283. [CrossRef] [Medline]
  13. Amtmann D, Cook KF, Jensen MP, et al. Development of a PROMIS item bank to measure pain interference. Pain. Jul 2010;150(1):173-182. [CrossRef] [Medline]
  14. Brodke DJ, Saltzman CL, Brodke DS. PROMIS for orthopaedic outcomes measurement. J Am Acad Orthop Surg. Nov 2016;24(11):744-749. [CrossRef] [Medline]
  15. Kroenke K, Talib TL, Stump TE, et al. Incorporating PROMIS symptom measures into primary care practice-a randomized clinical trial. J Gen Intern Med. Aug 2018;33(8):1245-1252. [CrossRef] [Medline]
  16. Wagner LI, Schink J, Bass M, et al. Bringing PROMIS to practice: brief and precise symptom screening in ambulatory cancer care. Cancer. Mar 15, 2015;121(6):927-934. [CrossRef] [Medline]
  17. Common data elements repository. National Institutes of Health. URL: https://cde.nlm.nih.gov/home [Accessed 2024-10-14]
  18. NIH strategic plan for data science. National Institutes of Health. URL: https://datascience.nih.gov/nih-strategic-plan-data-science [Accessed 2025-05-09]
  19. Zheng T, Gao Y, Wang F, et al. Detection of medical text semantic similarity based on convolutional neural network. BMC Med Inform Decis Mak. Aug 7, 2019;19(1):156. [CrossRef] [Medline]
  20. Yang X, He X, Zhang H, Ma Y, Bian J, Wu Y. Measurement of semantic textual similarity in clinical texts: comparison of transformer-based models. JMIR Med Inform. Nov 23, 2020;8(11):e19735. [CrossRef] [Medline]
  21. Mahajan D, Poddar A, Liang JJ, et al. Identification of semantically similar sentences in clinical notes: iterative intermediate training using multi-task learning. JMIR Med Inform. Nov 27, 2020;8(11):e22508. [CrossRef] [Medline]
  22. Ormerod M, Martínez Del Rincón J, Devereux B. Predicting semantic similarity between clinical sentence pairs using transformer models: evaluation and representational analysis. JMIR Med Inform. May 26, 2021;9(5):e23099. [CrossRef] [Medline]
  23. Li Q, He S. Similarity matching of medical question based on Siamese network. BMC Med Inform Decis Mak. Apr 6, 2023;23(1):55. [CrossRef] [Medline]
  24. Mihalcea R, Corley C, Strapparava C. Corpus-based and knowledge-based measures of text semantic similarity. Presented at: Proceedings of the Twenty-First National Conference on Artificial Intelligence and the Eighteenth Innovative Applications of Artificial Intelligence Conference; Jul 16-20, 2020; Boston, MA. [CrossRef]
  25. Devlin J, Chang MW, Lee K, Toutanova B. BERT: pre-training of deep bidirectional transformers for language understanding. Presented at: Proceedings of the North American Chapter of the Association for Computational Linguistics; Jun 2-9, 2019; Minneapolis, MN. [CrossRef]
  26. Artetxe M, Schwenk H. Massively multilingual sentence embeddings for zero-shot cross-lingual transfer and beyond. Trans Assoc Comput Linguist. Nov 2019;7:597-610. [CrossRef]
  27. Miao Z, Wu Q, Zhao K, Wu Z, Tsuruoka Y. Enhancing cross-lingual sentence embedding for low-resource languages with word alignment. Presented at: Findings of the Association for Computational Linguistics; Jun 16-21, 2024; Mexico City, Mexico. [CrossRef]
  28. Belinkov Y, Bisk Y. Synthetic and natural noise both break neural machine translation. arXiv. Preprint posted online on Feb 24, 2018. [CrossRef]
  29. Tiyajamorn N, Kajiwara T, Arase Y, Onizuka M. Language-agnostic representation from multilingual sentence encoders for cross-lingual similarity estimation. Presented at: Proceedings of Conference on Empirical Methods in Natural Language Processing; Nov 7-11, 2021; Online and Punta Cana, Dominican Republic. [CrossRef]
  30. Reimers N, Gurevych I. Sentence-BERT: sentence embeddings using siamese BERT-networks. Presented at: Proceedings of Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP); Nov 3-7, 2019; Hong Kong, China. [CrossRef]
  31. Feng F, Yang Y, Cer DM, Arivazhagan N, Wang W. Language-agnostic BERT sentence embedding. Presented at: Proceedings of 60th Annual Meeting of the Association for Computational Linguistics (Volume 1); May 22-27, 2022; Dublin, Ireland. [CrossRef]
  32. Agirre E, Cer DM, Diab MT, Gonzalez-Agirre A. SemEval-2012 task 6: a pilot on semantic textual similarity. Presented at: Proceedings of the International Workshop on Semantic Evaluation; Jun 7-8, 2012; Montréal, Canada. URL: https://aclanthology.org/S12-1051.pdf [Accessed 2025-05-13]
  33. Gao T, Yao X, Chen D. SimCSE: simple contrastive learning of sentence embeddings. Presented at: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing; Nov 7-11, 2021; Online and Punta Cana, Dominican Republic. [CrossRef]
  34. Ham J, Choe YJ, Park K, Choi I, Soh H. KorNLI and KORSTS: new benchmark datasets for Korean natural language understanding. Presented at: Findings of the Association for Computational Linguistics; Nov 16-20, 2020. [CrossRef]
  35. Mutinda FW, Yada S, Wakamiya S, Aramaki E. Semantic textual similarity in Japanese clinical domain texts using BERT. Methods Inf Med. Jun 2021;60(S 01):e56-e64. [CrossRef] [Medline]
  36. Zhang N, Chen M, Bi Z, et al. CBLUE: a Chinese biomedical language understanding evaluation benchmark. Presented at: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1); May 22-27, 2022; Dublin, Ireland. [CrossRef]
  37. HealthMeasures. URL: https://www.healthmeasures.net [Accessed 2024-10-14]
  38. Introduction to Ministry of Food and Drug Safety [Article in Korean]. Republic of Korea’s e-Government. URL: https://www.mfds.go.kr/index.do [Accessed 2024-10-14]
  39. Dietary lifestyle assessment [Article in Korean]. National Cancer Information Center. URL: https://www.cancer.go.kr/lay1/S1T226C228/contents.do [Accessed 2024-10-14]
  40. My dietary assessment [Article in Korean]. Seoul Asan Medical Center Health Screening & Promotion Center. URL: https://health.amc.seoul.kr/health/maintain/check.do [Accessed 2024-10-15]
  41. Salton G, Wong A, Yang CS. A vector space model for automatic indexing. Commun ACM. Nov 1975;18(11):613-620. [CrossRef]
  42. Googletrans-py 4.0.0. The Python Package Index. URL: https://pypi.org/project/googletrans-py [Accessed 2024-10-10]
  43. AdamW documentation. PyTorch. URL: https://pytorch.org/docs/stable/generated/torch.optim.AdamW.html [Accessed 2024-10-10]
  44. GPT-4o system card. OpenAI. URL: https://openai.com/research/gpt-4o-system-card [Accessed 2025-01-19]
  45. Fine-tuning guide. OpenAI. URL: https://platform.openai.com/docs/guides/fine-tuning [Accessed 2025-01-19]
  46. Brown T, Mann B, Ryder N, et al. Language models are few-shot learners. arXiv. Preprint posted online on Jul 22, 2020. [CrossRef]
  47. Wang W, Chen G, Wang H, Han Y, Chen Y. Multilingual sentence transformer as a multilingual word aligner. Presented at: Findings of the Association for Computational Linguistics; Dec 7-11, 2022; Abu Dhabi, United Arab Emirates. [CrossRef]
  48. Naver Cloud Platform. URL: https://www.ncloud.com/product/compute/server#detail [Accessed 2024-10-14]
  49. Mao Z, Nakagawa T. LEALLA: learning lightweight language-agnostic sentence embeddings with knowledge distillation. Presented at: Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics; May 2-6, 2023; Dubrovnik, Croatia. [CrossRef]
  50. Jiao X, Yin Y, Shang L, et al. TinyBERT: distilling BERT for natural language understanding. Presented at: Findings of the Association for Computational Linguistics; Nov 16-20, 2020. [CrossRef]
  51. Wang F, Zhang Z, Zhang X, Wu Z, Mo T, Lu Q, et al. A comprehensive survey of small language models in the era of large language models: techniques, enhancements, applications, collaboration with LLMS, and trustworthiness. arXiv. Preprint posted online on Dec 28, 2024. [CrossRef]


BERT: bidirectional encoder representations from transformers
BoW: bag-of-words
CDE: common data elements
LaBSE: language-agnostic bidirectional encoder representations from transformers sentence embedding
NIH: National Institutes of Health
PGHD: person-generated health data
PROMIS: Patient-Reported Outcomes Measurement Information System
SBERT: sentence-bidirectional encoder representations from transformers
SPURT: Standardized Patient-Generated Health Data Utilization Resources and Tools
STS: semantic textual similarity


Edited by Arriel Benis; submitted 24.01.25; peer-reviewed by Anand Rampadarath, Urte Radvilaite; final revised version received 16.04.25; accepted 20.04.25; published 10.06.25.

Copyright

© Sunghoon Kang, Hyewon Park, Ricky Taira, Hyeoneui Kim. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 10.6.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.