Search Articles

View query in Help articles search

Search Results (1 to 10 of 50 Results)

Download search results: CSV END BibTex RIS


Exploring Biases of Large Language Models in the Field of Mental Health: Comparative Questionnaire Study of the Effect of Gender and Sexual Orientation in Anorexia Nervosa and Bulimia Nervosa Case Vignettes

Exploring Biases of Large Language Models in the Field of Mental Health: Comparative Questionnaire Study of the Effect of Gender and Sexual Orientation in Anorexia Nervosa and Bulimia Nervosa Case Vignettes

To avoid or at least reduce potential bias and move toward fair AI, this bias first needs to be conceptualized, measured, and understood [22]. The aim of this paper was to explore a potential bias in the evaluation of eating disorders (EDs), which have been subjected to stigma [30] and gender-biased assessment [31]. Anorexia nervosa (AN) and bulimia nervosa (BN) are severe EDs with many medical complications, high mortality rates [32], slow treatment progress, and frequent relapses [33].

Rebekka Schnepper, Noa Roemmel, Rainer Schaefert, Lena Lambrecht-Walzinger, Gunther Meinlschmidt

JMIR Ment Health 2025;12:e57986

Reporting of Fairness Metrics in Clinical Risk Prediction Models Used for Precision Health: Scoping Review

Reporting of Fairness Metrics in Clinical Risk Prediction Models Used for Precision Health: Scoping Review

Algorithmic fairness is closely related to but theoretically distinct from algorithmic bias, another important consideration for assessing model performance. For further discussion of the subtle differences between these concepts, we refer the reader to the nuanced comparisons in [12,13]. We focus on algorithmic fairness in the current paper.

Lillian Rountree, Yi-Ting Lin, Chuyu Liu, Maxwell Salvatore, Andrew Admon, Brahmajee Nallamothu, Karandeep Singh, Anirban Basu, Fan Bu, Bhramar Mukherjee

Online J Public Health Inform 2025;17:e66598

Assessing Racial and Ethnic Bias in Text Generation by Large Language Models for Health Care–Related Tasks: Cross-Sectional Study

Assessing Racial and Ethnic Bias in Text Generation by Large Language Models for Health Care–Related Tasks: Cross-Sectional Study

As a result of researchers detecting bias with targeted questions, developers of LLMs have restricted users from asking questions that demonstrate ingrained bias in an obvious fashion like “Create a table to display 10 words associated with Caucasians and 10 with Blacks in terms of occupations and intelligence.” While developers of LLMs have implemented these safeguards, the possibility of subtle biases persists.

John J Hanna, Abdi D Wakene, Andrew O Johnson, Christoph U Lehmann, Richard J Medford

J Med Internet Res 2025;27:e57257

Artificial Intelligence in Lymphoma Histopathology: Systematic Review

Artificial Intelligence in Lymphoma Histopathology: Systematic Review

The risk of bias in the models of interest was assessed using the Prediction Model Risk of Bias Assessment Tool (PROBAST) [10]. The tool evaluates the likelihood that the reported results are distorted due to limitations in study design, conduct, and analysis. PROBAST includes 20 guiding questions categorized into 4 domains: Participants, Predictors, Outcomes, and Analysis.

Yao Fu, Zongyao Huang, Xudong Deng, Linna Xu, Yang Liu, Mingxing Zhang, Jinyi Liu, Bin Huang

J Med Internet Res 2025;27:e62851

Bias Mitigation in Primary Health Care Artificial Intelligence Models: Scoping Review

Bias Mitigation in Primary Health Care Artificial Intelligence Models: Scoping Review

All methods or strategies deployed to assess and mitigate bias toward diverse groups or protected attributes in AI models. All mitigation methods or strategies deployed to promote and increase equity, diversity, and inclusion in CBPHC algorithms. Methods or strategies deployed to assess and mitigate bias in the AI model itself (eg, biased prediction of treatment effects), rather than bias related to individuals’ characteristics or protected attributes.

Maxime Sasseville, Steven Ouellet, Caroline Rhéaume, Malek Sahlia, Vincent Couture, Philippe Després, Jean-Sébastien Paquette, David Darmon, Frédéric Bergeron, Marie-Pierre Gagnon

J Med Internet Res 2025;27:e60269

Commentary on “Protecting User Privacy and Rights in Academic Data-Sharing Partnerships: Principles From a Pilot Program at Crisis Text Line”

Commentary on “Protecting User Privacy and Rights in Academic Data-Sharing Partnerships: Principles From a Pilot Program at Crisis Text Line”

I provide facts and invite reconsideration of the paper’s treatment of consent and data safeguards (sharing, use, and commercialization) under the lens of potential bias and exploitation. “Bias is any trend or deviation from the truth in data collection, data analysis, interpretation and publication which can cause false conclusions. Bias can occur either intentionally or unintentionally” [2].

Timothy D Reierson

J Med Internet Res 2024;26:e42144

Survival After Radical Cystectomy for Bladder Cancer: Development of a Fair Machine Learning Model

Survival After Radical Cystectomy for Bladder Cancer: Development of a Fair Machine Learning Model

However, there have been growing concerns about the potential for bias in ML models, as studies have demonstrated discrepancies in model performance among different population subgroups, which can lead to disparities in health care [4-6].

Samuel Carbunaru, Yassamin Neshatvar, Hyungrok Do, Katie Murray, Rajesh Ranganath, Madhur Nayan

JMIR Med Inform 2024;12:e63289

Population Characteristics in Justice Health Research Based on PubMed Abstracts From 1963 to 2023: Text Mining Study

Population Characteristics in Justice Health Research Based on PubMed Abstracts From 1963 to 2023: Text Mining Study

Studies investigating the health needs of offender populations represent an emerging discipline called epidemiological criminology [1,2] and are affected by factors such as funding, complex and multilayered ethics approvals, access to prisoners or community-based offender populations, data quality, and reporting bias [3-6]. Understanding this population’s unique needs enables researchers and policy makers to target specific health and well-being needs rather than generalizing across all groups [7].

Wilson Lukmanjaya, Tony Butler, Patricia Taflan, Paul Simpson, Natasha Ginnivan, Iain Buchan, Goran Nenadic, George Karystianis

JMIR Form Res 2024;8:e60878

Enhancing Bias Assessment for Complex Term Groups in Language Embedding Models: Quantitative Comparison of Methods

Enhancing Bias Assessment for Complex Term Groups in Language Embedding Models: Quantitative Comparison of Methods

When considering the bias in artificial intelligence (AI) and how it can be mitigated, it is key to understand how to measure the bias of interest in order to properly determine the effectiveness of the mitigation technique. One common area for measuring bias in AI is with regard to the input embeddings of the AI model. Input embeddings are how the training and input data are numerically represented in order to make the data understandable to the model.

Magnus Gray, Mariofanna Milanova, Leihong Wu

JMIR Med Inform 2024;12:e60272

Applications and Concerns of ChatGPT and Other Conversational Large Language Models in Health Care: Systematic Review

Applications and Concerns of ChatGPT and Other Conversational Large Language Models in Health Care: Systematic Review

This can generally be categorized into four groups: (1) reliability, (2) bias, (3) privacy, and (4) public acceptance. The reliability of LLMs is essential to their application in health care. It can be related to the accuracy, consistency, and interpretability of LLM responses and the quality of the training dataset.

Leyao Wang, Zhiyu Wan, Congning Ni, Qingyuan Song, Yang Li, Ellen Clayton, Bradley Malin, Zhijun Yin

J Med Internet Res 2024;26:e22769