Original Paper
Abstract
Background: Machine learning (ML) systems in health care have the potential to enhance decision-making but often fail to address critical issues such as prediction explainability, confidence, and robustness in a context-based and easily interpretable manner.
Objective: This study aimed to design and evaluate an ML model for a future decision support system for clinical psychopathological treatment assessments. The novel ML model is inherently interpretable and transparent. It aims to enhance clinical explainability and trust through a transparent, hierarchical model structure that progresses from questions to scores to classification predictions. The model confidence and robustness were addressed by applying Monte Carlo dropout, a probabilistic method that reveals model uncertainty and confidence.
Methods: A model for clinical psychopathological treatment assessments was developed, incorporating a novel ML model structure. The model aimed at enhancing the graphical interpretation of the model outputs and addressing issues of prediction explainability, confidence, and robustness. The proposed ML model was trained and validated using patient questionnaire answers and demographics from a web-based treatment service in Denmark (N=1088).
Results: The balanced accuracy score on the test set was 0.79. The precision was ≥0.71 for all 4 prediction classes (depression, panic, social phobia, and specific phobia). The area under the curve for the 4 classes was 0.93, 0.92, 0.91, and 0.98, respectively.
Conclusions: We have demonstrated a mental health treatment ML model that supported a graphical interpretation of prediction class probability distributions. Their spread and overlap can inform clinicians of competing treatment possibilities for patients and uncertainty in treatment predictions. With the ML model achieving 79% balanced accuracy, we expect that the model will be clinically useful in both screening new patients and informing clinical interviews.
doi:10.2196/64617
Keywords
Introduction
Overview
As in most areas of human activity, mental health care has seen an increase in the application of artificial intelligence (AI), particularly in assisting the diagnosis and treatment of mental disorders [
]. While the debate between clinical judgment and statistical methods is longstanding in the field of clinical psychology [ ], under appropriate conditions, statistical tools can perform as well as—or better than—clinical judgment in diagnosing psychopathologies [ ]. To be effective, these statistical tools should be context-based and simple for clinicians to use [ , ].When assessing patients for treatment in a mental health setting, clinicians consider various screening instruments in addition to demographic and medical information, assessing the scores related to possible diagnoses and treatments based on the patient’s answers to questions [
]. Patients with complex clinical scenarios present a challenge for clinicians in assessing the various possible symptoms to arrive at a most likely diagnosis [ ]. In diagnosing depression, for example, comorbidity with anxiety disorders and other factors was found to lead to uncertainty in diagnosis in as many as 40% of cases [ ]. The consequences of an uncertain diagnosis may be high rates of treatment-resistant depression [ , ].Early correct intervention may reduce the burden of disease for the individual as well as the society. The global costs of mental disorders were estimated to be US $2493 billion in 2010 and are expected to more than double by 2030 [
]. Wittchen et al [ ] deemed mental disorders “the core health challenge of the 21st century.” In their key recommendations for future research, they state that “clinically sensitive and economically feasible decision algorithms could be explored to determine what types of interventions should be assigned to which type of patient.” Therefore, insight into the underlying factors that characterize a psychopathological diagnosis and the factors that discriminate between alternative diagnoses is valuable and a contribution to solving a larger and urgent health need.Mental health care has benefited from the ability of AI technologies to analyze data, infer classifications from new data, and communicate insights from data [
]. AI has been recognized as an important tool in clinical decision-making for mental health [ ]. In assessment, AI technologies offer the potential to complement the role of clinicians with diagnostic decision support technologies [ , ]. Such diagnostic tools require the AI models to be explainable and interpretable [ ], both in terms of the model itself (global) and the individual model predictions (local). Interpretability refers to how a model arrives at a decision, and explainability refers to why the decision is reached [ ]. While explainability in AI has been a topic of interest since the 1980s, the recent accelerations in the widespread use of AI have resulted in guidelines on the responsible use of AI. Responsible AI requires a robust model that conveys the confidence and degree of uncertainty in prediction [ - ]. Although confidence and uncertainty estimation are considered a necessity for medical AI models to enable safe clinical deployment [ ], it has been claimed that most machine learning (ML) methods in the recent medical literature neglect the important issue of model uncertainty [ ].This paper presents the design of an ML model for psychopathological mental health treatment tailored to the omnipresent questionnaire-type screening instruments. The AI model described here is based on a novel hierarchical structure that aims to meet the needs of clinical interpretability while accounting for decision confidence and uncertainty. The presented approach facilitates the questionnaire instruments that are typically used in mental health care to assess symptom severity, but in a data-driven way [
], while leveraging any other information known about respondents.This paper first lays out the rationale for explainable AI (XAI) methods in medical models, introducing the important issues of model confidence and uncertainty. Typical ways in which model confidence and uncertainty can be quantified are introduced with an emphasis on Monte Carlo dropout (MCD), the method used in this paper. The model structure is then introduced and shown to provide interpretability by design with confidence and uncertainty estimation provided through the use of MCD. Next, model training and validation on clinical data from a Danish web-based mental health treatment service are detailed and results presented. Finally, the implications of the work are discussed.
Background
After initial interest in the explainability of AI models in the 1980s and 1990s, recent advances in AI and ML, and the increased use of such technologies in safety-critical, socioeconomically and medically impactful applications have driven a renewed interest in XAI [
]. The explainability of AI models is generally understood as the ability for a wider range of users to understand the outputs of a given AI or ML model [ ]. The concept of interpretability is closely related and refers to the ability of a person to understand how a given set of inputs results in a model output [ ].As an extension to the explainability of ML models, recent endeavors have put more emphasis on responsible or trustworthy AI [
], with confidence and robustness being important technical prerequisites. The confidence of an ML algorithm is defined as the model’s own assessment that the provided inferences are correct. Robustness is defined as the ability of a model to yield consistent results in the presence of uncertainty [ ]. In practice, robustness can be evaluated by perturbing the model structure to induce distribution shifts [ ], introducing epistemic uncertainty. According to XAI frameworks, robustness extends explainability while increasing stability and user satisfaction [ ]. Such measures of confidence and robustness are important for a model aimed at medical decision support systems (DSSs), such as the one presented in this paper [ ].ML Methods in Relation to XAI
Logistic regression is a classification model used in ML that applies a linear combination of the data X, with weights W, to discriminate between classes of the dependent variable Y. It models the probability of an outcome using a logistic (sigmoid) function:
(1)
Since the equation consists of a linear combination of the features transformed by a sigmoid function, each weight of the model directly corresponds to the weight of a feature in determining the output. Because of this linear nature, XAI deems logistic regression to be a white-box model [
, ]. White-box models are also referred to as having the property of transparency [ ]. In contrast, gray-box models and black-box models are categories of AI models with less ability to inspect the inner workings of the model (less interpretability and less transparency), but typically with more power to discriminate, and are often used in complex applications such as computer vision. This property of powerful models being black-box is often referred to as the interpretability-performance trade off [ ]. Other common white-box ML models include decision trees and K-nearest neighbor models [ ]. Deep neural networks are an example of a black-box model due to their many stacked nonlinear layers, typically consisting of millions of weights, which make their operation uninterpretable. Post hoc techniques add interpretability as an additional method or algorithm to the AI model [ ]. One such approach is to add an additional, surrogate model that simplifies relevant sections of the AI model to be explained. Local interpretable model-agnostic explanations (LIME) [ ] is a method that can explain the prediction of any classifier by building a local model that aims to approximate the original model and its predictions. The new model is chosen to be an interpretable model (such as logistic regression). This allows the user to better understand the relationship between a single input and the corresponding output prediction by providing a local explanation and is, therefore, an explanation by simplification. However, LIME has some drawbacks: it does not inherently provide global explanations of the dataset as a whole and it can experience inconsistency because it requires multiple perturbations of input data and retraining local surrogate models [ ]. An alternative XAI method that ranks the importance of features used by the model is Shapley additive explanations (SHAP) [ ]. SHAP is a method that calculates an additive feature importance score, which can provide local explanations for individual predictions as well as global explanations for the whole dataset. However, SHAP may experience sensitivity to feature perturbations and to feature correlations and interactions [ ]. As a whole, a drawback of post hoc explainability techniques is that they approximate the model’s decision-making rather than directly reflecting its internal structure.Since explaining black-box models requires post hoc techniques, SHAP and LIME have been applied to deep learning models in medical applications, for example, to aid the explainability of complex convolutional neural network medical diagnostic models for chest X-ray classification for COVID-19 infection [
]. SHAP has also been applied as a backend explanation module for a neural network predictor of COVID-19 infection diagnosis in eHealth record data [ ], providing both local and global explanations. Post hoc XAI techniques, such as these, represent the vast majority of XAI use cases in health care [ ]. To overcome the drawbacks of post hoc methods, our new approach offers an inherently interpretable model for predicting mental health treatment outcomes. Its transparency stems from a logistic regression framework, which, as described earlier, is interpretable as it is possible to make a direct link between the value of an input parameter, its associated weight, and the prediction made by the model. Moreover, by using a hierarchical structure that moves from questions to scores to classification predictions, this design enables users to understand how and why the model makes decisions, both at the questionnaire score level and the individual question level.Responsible AI
As previously stated, responsible AI requires an understanding of its own confidence and robustness in inferences. The advent of deep learning has resulted in a step change in the performance of models and the associated practical use of AI. However, this has come with disadvantages. Deep learning models can be confidently wrong, especially in case of a shift in the input data, and in recent years, such overconfidence in incorrect inferences has proved fatal, as illustrated by serious accidents caused by machine vision models in cars [
]. Most ML algorithms provide point-estimate predictions, meaning they provide an inference with, at best, a scalar estimate of the model’s confidence in the inference. More recently, methods based on the application of Bayesian statistics have been proposed [ ]. In such methods, rather than estimating a point estimate, the models provide a probabilistic inference modeled by a probability distribution. This distribution can then be used for a point-estimate inference (eg, by taking the mode of the distribution as the inference) and can also be used to assess the model’s uncertainty in this inference by considering the shape of the probability distribution. Distributions strongly peaked around the inference indicate low uncertainty; very wide and flat distributions around the inference indicate high uncertainty.These models often also provide probabilistic information on the effect of model predictions in the face of perturbations in the input features. For example, Cardelli et al [
] showed that for Bayesian inference with Gaussian processes, a class of algorithms overlapping with deep neural networks that have inherent uncertainty estimates [ ], the model’s uncertainty in response to perturbations in the input space can be upper-bounded, thus providing a measure for the model’s robustness.If a full Bayesian probability distribution is established for all parameters in the model, with statistics either learned from data or based on existing knowledge, it can then be used to investigate the confidence and robustness of the model. However, a significant drawback of this approach is that it can be compute-intensive [
].An alternative method called MCD is inspired by using the dropout layers in deep neural networks to improve the ability of a network to “generalize,” or provide good quality predictions on yet unseen data [
]. When used to allow the model to generalize well, dropout sets weights of the model to 0 randomly during training only [ ]. The result is that the trained model does not become overreliant on any one (potentially spurious) characteristic of the data. In contrast, MCD also uses dropout when new predictions are made. It does so by obtaining repeated inferences from the trained model, but with different weights dropped out in every iteration of the inference. The result is a distribution of inferences rather than a point-estimate. Such inference distributions can be shown to be equivalent to those obtained by a fully Bayesian model for an important family of probabilistic models, called Gaussian process models, when dropout is applied before every weight layer [ ]. Consequently, MCD results in a Bayesian approximation by randomly switching off connections in the network during inference.MCD has been applied to estimate uncertainty in medical imaging AI convolutional neural network models where the model perturbations allow the estimation of epistemic uncertainty in the prediction of potentially cancerous skin lesions [ at each iteration, to yield the sample distribution y*. The mean value of class c (µc) of the sample is the prediction probability of the class, while the variance of class c (σc2) is the uncertainty measure:
(2)
(3)
Methods
Overview
This paper presents a methodology to obtain explainable, interpretable models that are robust and provide a reliable confidence score for use with an important class of data in mental health, namely that of sum-score questionnaires. The model uses the answers to individual questions as inputs, but also explicitly represents the traditionally used sum-scores of such questionnaires in the model. Hence, explainability and interpretability are provided both at the question level and at the sum-score level. To also provide robustness and a measure of model confidence in its predictions, MCD is applied as described in the previous section.
The ability of the proposed model to provide interpretable predictions as well as estimates of confidence is demonstrated with a use case aimed at providing clinicians with treatment recommendations for several mental disorders.
Metrics
Precision, recall, F1-score, balanced accuracy score, weighted F1-score, and area under the curve (AUC) are relevant metrics for evaluating the predictive model’s classification performance. Precision measures the proportion of correctly predicted treatment decisions among all predictions for a given class. A high precision score ensures that the model minimizes incorrect treatment assignments. Recall (or sensitivity) assesses the proportion of correctly predicted treatments among all actual instances of that class, reflecting the model’s ability to identify the correct treatment when needed. F1-score is the harmonic mean of precision and recall, balancing both metrics to provide a single performance measure, particularly useful in cases of class imbalance. The balanced accuracy score is the average of recall across all classes, accounting for imbalanced class distributions by ensuring each class contributes equally to the overall performance evaluation. Weighted F1-score adjusts the F1-score by considering class frequencies, preventing minority classes from being overshadowed in imbalanced datasets. The AUC measures the model’s ability to distinguish between different treatment classes, with higher values indicating better discrimination. During validation these metrics ensure reliable treatment predictions, minimize incorrect classifications, and balance performance across all treatment options, reducing potential clinical risks.
Data
A dataset comprising patient questionnaire answers and demographics was used to train and validate the model. The data were gathered between November 14, 2019, and December 31, 2022, from patients of the “Internetpsykiatrien” web-based treatment services delivered by the Centre for Digital Psychiatry, Denmark. The center delivers routine care internet-delivered cognitive behavioral therapy with nationwide coverage [
]. The data consisted of answers to the Patient Health Questionnaire-9 (PHQ-9) [ ], the Generalized Anxiety Disorder Questionnaire-7 (GAD-7) [ ], the Social Interaction Anxiety Scale (SIAS) [ ], the Panic Disorder Severity Scale (PDSS) [ ], and the Fear Questionnaire [ ]. Additional features included demographic information and a brief medical history, such as previously diagnosed conditions.The ground truth of the dataset was the choice of treatment selected by the psychologists in the clinic for each patient. The treatment decision was reached using multiple steps. First, clinical psychologists or psychologists under the supervision of a clinical psychologist, assessed patients using questionnaires to determine the optimal treatment. This assessment used the same questionnaire and demographic data that was used by the model, except for free text data, which were available to the psychologists, but not to the model. Second, they performed a clinical assessment interview, including the use of the semistructured interview, Mini Neuropsychiatric Interview. Any ambiguities were discussed at a clinical conference.
A total of 91 input features were selected for inclusion in the models (
), with data completeness as the primary inclusion criterion. The dataset (N=1068) was randomly split into a training set (n=801, 75%) and a test set (n=267, 25%). The training data consisted of 4 prediction class categories as follows: (1) depression (n=269, 33.6%); (2) panic (n=269, 33.6%); (3) social phobia (n=193, 24.1%); and (4) specific phobia (n=70, 8.7%). The test set class categories were stratified to match the training set in proportion: 33.7% (90/267); 33.3% (89/267); 24% (64/267); and 9% (24/267) for the 4 classes, respectively.The percentages of training or test set data per class were: 34% (269/801); 34% (269/801); 24% (193/801); and 9% (70/801) (rounded), respectively. Given this class distribution, we opted against upsampling the underrepresented classes (social phobia and specific phobia) or downsampling the overrepresented classes (depression and panic). While upsampling could mitigate class imbalance, it risks introducing synthetic patterns that may not generalize well to real-world data. In addition, given the relatively small sample size of the latter category (70/801, 9% in training), aggressive upsampling could lead to overfitting, where the model learns artifacts of the duplicated instances rather than underlying patterns in the data.
Conversely, downsampling depression and panic would reduce the total available training data and potentially discard valuable information, impairing the model’s ability to capture the full variation within these classes. Furthermore, since the dataset represents naturally occurring clinical distributions, artificial class balancing could distort the real-world relevance of model predictions. Instead, we opted to maintain the original class distribution and address class imbalance using performance metrics robust to class imbalance, such as balanced accuracy, weighted F1-score, and AUC, ensuring that model evaluation accounts for differences in class representation.
ID | Feature | Description | Feature type and values |
1 | Education | Years of education |
|
2 | AlcoholIntake | Weekly alcohol use (units) |
|
3 | UseOfDrugs | Use of euphoric drugs or pills |
|
4 | CurrentMedication | Current psychopharmacological treatment |
|
5-13 | Phq{One: Nine} | All 9 questions of PHQ-9a [ | ] over the previous 2 weeks scored on a 4-point Likert scale
|
14 | PhqTen | Assessing to what extent the symptoms reported using PHQ-9 have affected work, household tasks, and relationships scored on a 4-point Likert scale |
|
15-21 | Gad{One:Seven) | All 7 questions of GAD-7b [ | ] scored on a 4-point Likert scale
|
22 | HasPhobia | Binary categorical variable of self-reported previous phobia diagnosis |
|
23-46 | Fq{One:TwentyFour} | All 24 questions of the Fear Questionnaire [ | ] Items 1-17 assess avoidance. Item 18 assesses phobic symptom severity. Items 19-24 assess how reported phobia affects the respondent.
|
47-66 | Sias{One:Twenty} | SIASc questionnaire [ | ] assesses the respondent’s social interaction anxiety using 20 statements, which the respondent indicates agreement with (“... characteristic or true of me”) on a 5-item Likert scale
|
67-73 | Pdss{One:Seven} | PDSSd questionnaire [ | ]
|
74,75,76 | MansaTen, MansaTwentyFour, MansaTwentyFive | Questions 10, 24, and 25 of the 25-item MANSAe questionnaire [ | ]
|
77-80 | DAD | Previous diagnosis one-hot encoded using dummy variables DAD0-DAD3 |
|
81-87 | INC | Describing source of income of the respondent. One-hot encoded using dummy variables INC_1 to INC_7 |
|
88-91 | APP | Treatment applied for one-hot encoded using dummy variables APP_1 to APP_4 |
|
GT | TreatmentType | Treatment decision (ground truth) |
|
aPHQ-9: Patient Health Questionnaire-9.
bGAD-7: Generalized Anxiety Disorder Questionnaire-7.
cSIAS: Social Interaction Anxiety Scale.
dPDSS: Panic Disorder Severity Scale.
eMANSA: Manchester Short Assessment of Quality of Life.
Ethical Considerations
Data were extracted from Internetpsykiatrien after approval from The Regional Council in Southern Denmark. Since this was a secondary data analysis, separate informed consent was not required. However, the Regional Committees on Health Research Ethics for Southern Denmark were informed about the study and were provided with the case number S-20232000-65. In accordance with Danish national ethical guidelines, no additional ethics approval was needed. The study was reported to the Danish Data Protection Agency. Data was anonymized before being provided to the researchers. No compensation was provided.
Model
The proposed model was designed to interpret the traditional sum-score questionnaires used in mental health treatment in a data-driven, explainable fashion. Sum-score questionnaires offer several prespecified answers to each question, with each answer associated with a numeric, typically integer, score. These scores are then summed and in many cases a threshold is then applied, which indicates whether the respondent demonstrates clinically significant symptoms for the disorders tested by the instrument. Hence, such sum-scores play an important role in clinical decision-making. ML models such as logistic regression strongly align with this principle while also offering inherent interpretability due to the nature of the output activation being related to a weighted sum of the inputs. Such inherent explainability is appealing; however, the complexity of the information conveyed when models contain many weights can lead to cognitive overload, hampering decision-making [
] and explainability [ ].To improve the model’s explainability while also retaining its inherent interpretability, a hierarchical model was proposed. As can be seen in
, this model mimics the sum-score models in the hidden layer of the model, in which the responses of each individual instrument are multiplied by their respective weight and then summed to yield pseudosum scores. These scores differ from normal sum-scores by applying a learned weighting to each answer before summing. These pseudosum scores incorporate a rectified linear unit activation function to ensure a linear sum while also ensuring the sum cannot be negative (due to the model learning negative weights during training). These scores and their weights allow inherent interpretability on a per-score basis, which will be familiar to clinical users of traditional questionnaire instruments. These pseudosum scores are then weighted before being combined in the output layer with the weighted sum-score of all other used instruments. In addition to the pseudosum scores, any other features with potentially discriminative power, such as sociodemographical information, can be added. All weights have dropout applied during prediction, consistent with the use of MCD as described previously. A softmax-type activation function is used in the output layer. The softmax activation function ensures that the sum of all the outputs of the model equals 1. Hence, each of the outputs can be interpreted as the probability that the corresponding treatment should be recommended. This output probability is illustrated as p(depression), as shown in . Note that the output layer will typically contain 1 output node for each of the treatments considered by the model and that this number is not necessarily the same as the number of screening instruments used as inputs to the model.With this specific architecture, the model achieved three important goals as follows: (1) the model allows the nonlinear combination of multiple instruments to achieve a treatment prediction, (2) the model allows the explanation of its predictions based on pseudosum score results and thus provides score-level explainability and trust in the model, and (3) at the same time, the model allows for the analysis of model predictions based on the input features, enabling users to understand which particular responses to each questionnaire drive the model predictions. These insights may provide therapists with valuable information on how to personalize the therapy.
The model’s structure inherited its interpretability from its foundation in logistic regression, while confidence and robustness of the model were addressed using the aforementioned MCD. As previously stated, MCD can be shown to approximate the probabilistic results obtained by Bayesian approaches for a class of models called Gaussian Processes. The latter are a generalization of generalized linear models, one of which is the logistic regression model chosen in our approach for its explainability. The hierarchical logistic regression model has dropout applied before every weight layer, since the Bayesian properties of MCD apply for neural networks of arbitrary depth and nonlinearities, as long as dropout is applied in this way [
]. While it is usual for dropout to be applied to neural networks to fulfill the function of regularization during training time [ ], as explained, MCD also applies dropout during inference. The probability of a node being dropped, the dropout rate, is a parameter of the dropout layers. In MCD, the dropout rate is typically set heuristically. Extensions to the technique, such as learnable Bernoulli dropout [ ], allows the dropout rate to be jointly optimized in the model training phase. When MCD is used in this way, to approximate a Bayesian network, it should not be seen as a regularization method [ ]. Therefore, L1 regularization is applied during training to reduce overfitting. L1 regularization attempts to achieve good generalization (ie, good performance on yet unseen data) by forcing the influence of unimportant features to 0. L2 regularization can be applied as an alternative, fulfilling the regularization function without enforcing sparsity in the weights. Further justification for the choice of L1 regularization is provided in the Results section. shows the model details.
Prediction Explainability and Confidence
The distribution of the softmax output classes over the N samples
, are visualized using Violin plots of the output class probability distribution over the N samples of a single patient prediction. A prediction for a single patient is shown in and . The outer, colored, “violin” shape visualizes the probability distributions over the N samples of the single patient example, while the embedded box-plots illustrate the medians and IQR of the model’s predictions. The confidence and uncertainty of the prediction can be understood by observing the shape and overlapping of the violin plots. provides an example of depression being predicted with high likelihood and nonoverlapping with the other 3 classes, whereas the predictions for all 3 other disorders have a low median probability. Predictions with low confidence (ie, low probabilities) and high uncertainty (ie, large probability variation), where more than 1 class may credibly be predicted, resulted in outputs, such as shown in . This is an example of the relation between XAI concepts [ ], using visualization to enhance explainability, contributing to improved interpretability, and the interpretability of the model leading to validation of the explanation.


Prediction Interpretability and Robustness
Interpretability at the model level is aided by examining the learned weights after training. Since the scores are weighted linear sums of the question inputs WTX, the learned weights (W) are informative of the importance of each of the question features in determining the score used in the output class predictions in the softmax layer. The softmax layer implements a multiclass logistic regression function for each class (subscript c) and score (subscript s):
(4)
Where σ represents the sigmoid (or logistic) function, Xs is the input coming from the score layer or demographic features, Wc represents the weights in the softmax layer for class c, and Bc represents a bias value for each class.
Determining feature importance is an important aspect of explainability. At the model level, applying L1 regularization to logistic regression tends to force the weights of unimportant features to zero, leaving more important features as nonzero [
]. However, it is important to be aware that collinear features may not be equally important for a model since they can be accounted for by just one of the collinear features, with the others set to 0. Constraining trained weights to be nonnegative in the score layer aids interpretability since traditionally, questions are summed to produce a score. Therefore, the relative weights play a useful role in determining feature importance, provided that the inputs have been normalized to the same scale during training and prediction.At the model level, standardized coefficients obtained by multiplying the original coefficients by the SD of the corresponding feature, provides a scale-free measure of influence, allowing the comparison of the relative importance of different features [
].At the prediction level, for a specific patient record, the weights can be interpreted as the feature importance, given the normalized input range and positive weight constraint.
Results
Overview
The model was constructed using the TensorFlow library for Python (Python Software Foundation) and trained over 2000 epochs with a batch size of 32. The Adam optimizer was used at default setting, with L1 regularization of 0.001 and a dropout rate of 0.05 [
].During development, both L1 and L2 regularization were considered, with evaluations involving 5-fold cross-validation on the training set. The optimal balanced accuracy score for both L1 and L2 corresponded to a regularization factor of 0.001 for both, with L1 and L2 achieving similar cross-validated mean balanced accuracy scores. Since L1 regularization tends to lead to a sparse solution by forcing unrequired weights to 0, it may be favored over L2 for transparent interpretability. However, when there are multiple near-equivalent predictors (eg, correlated measures), small fluctuations in initialization can lead the model to “select” a slightly different subset of features for its nonzero coefficients, eliminating one correlated feature, for example, while retaining the other. In consistency tests over 30 training runs, consistency in the top features was observed, although the order of importance may switch the top positive weight. In the depression class, for example, the top positive feature was PHQ-9 in 70% (21/30) of the runs and DAD_1.0 (previously diagnosed with depression) in 30% (9/30) of the runs. In the panic class, PDSS was always the top positive feature, with GAD-7 taking second place in 40% (12/30) of the runs and fq_total in 30% (9/30) of the runs. Examining the softmax weights, 80% (24/30) of the cumulative sum of the weights for each class were represented in the top 6 factors for L1 regularization, whereas for L2, up to 11 factors were required to represent the 80% (24/30). Therefore, L1 regularization was adopted for interpretability since: (1) both L1 and L2 regularization demonstrated similar accuracy; (2) L1 regularization enforces sparsity, requiring fewer weights to explain the top 80% (24/30); (3) the top weights are consistently in the top 3; and (4) the model transparency ensures the same weights are used in both prediction and explanation.
Following training, the model was evaluated on the held-out test set, achieving a balanced accuracy score of 0.79, indicating good generalization. The validation metrics (
) indicate strong predictive performance across all 4 classes, with F1-scores ranging from 0.71 to 0.86, which indicates a balanced trade off between precision and recall. Specific phobia achieved the highest precision (0.84) and the highest recall (0.88), suggesting that the model is highly accurate in predicting this class when it does make a prediction. However, social phobia had the lowest recall (0.70) compared to other classes, indicating some difficulty in correctly identifying all cases of this condition. The AUC values, all >0.90, suggest strong discrimination between classes. The aggregated metrics demonstrate an overall balanced accuracy score of 0.79 and a weighted F1-score of 0.78, confirming that class imbalance did not severely impact overall model performance. The use of these weighted metrics ensures that smaller classes, such as specific phobia, contribute proportionally without distorting the results.The confusion matrix (
) details the number of treatment recommendations proposed by the model, with correct classifications appearing on the main diagonal of the table. Off-diagonal elements quantify the number of cases in which the model chose a different treatment than the clinician.Patients assessed with social phobia were misclassified as depression in 4.9% (13/267) cases and as panic disorder in 1.9% (5/267) cases, suggesting that the model struggles to distinguish social phobia from other anxiety-related conditions. This was consistent with the lower recall observed for social phobia in
, indicating that some patients with social phobia were incorrectly classified into other categories.For specific phobia, the model correctly identified 7.9% (21/267) cases, but 1.1% (3/267) cases were misclassified as panic disorder (2/267, 0.7% cases) or social phobia (1/267, 0.4% case). Given that specific phobia was the least represented category in the dataset, the model’s relatively strong performance in this class may be influenced by its high specificity (0.98,
), which suggests that when the model does predict specific phobia, it does so with a high degree of accuracy. However, the smaller sample size in this class may also lead to performance variability.The model generally performed well in classifying depression and panic disorder, with high agreement between model predictions and clinician labels (72/267, 27% and 69/267, 25.8% correct classifications, respectively). However, 3.7% (10/267) panic disorder cases were misclassified as depression, indicating some overlap in how these conditions were represented in the feature space or possible comorbidities.
Treatment | Precision | Recall (sensitivity) | F1-score | Specificity | Per-class balanced accuracy | AUCa |
Depression | 0.76 | 0.81 | 0.78 | 0.87 | 0.84 | 0.93 |
Panic | 0.82 | 0.77 | 0.79 | 0.91 | 0.84 | 0.92 |
Social phobia | 0.71 | 0.70 | 0.71 | 0.91 | 0.81 | 0.91 |
Specific phobia | 0.84 | 0.88 | 0.86 | 0.98 | 0.93 | 0.98 |
aAUC: area under the curve.
Clinician proposed treatment (ground truth) | Model proposed treatment, n (%) | |||
Depression | Panic disorder | Social phobia | Specific phobia | |
Depression | 72 (27) | 8 (3) | 9 (3.4) | 0 (0) |
Panic | 10 (3.7) | 69 (25.8) | 8 (3) | 3 (1.1) |
Social phobia | 13 (4.9) | 5 (1.9) | 45 (16.8) | 1 (0.4) |
Specific phobia | 0 (0) | 2 (0.7) | 1 (0.4) | 21 (7.9) |
Model Level (Global) Explainability of Classes From Score Totals and Demographics
Examining the learned weights of the softmax layer that predicts the class probabilities from the score and demographic features is informative to interpret their relative importance at the model level. Positive weights contribute toward a more probable class prediction, while negative weights detract. Examining the top 6 positive and negative contributions, the depression class (
), is mostly influenced by the PHQ-9 score, application-related variables (APP_4.0, APP_3.0), and the previous diagnosis of depression (DAD_1.0). The Fear Questionnaire Global Phobia rating (question 18) detracts from this class prediction. The panic class ( ) is most influenced positively by the PDSS score and the Fear Questionnaire total score (fq_total) and most negatively by the use of euphoric drugs or pills (UseOfDrugs), so drug use detracts from this class probability. The social phobia class ( ) is most influenced by the SIAS score and the Fear Questionnaire Global Phobia rating (fq_global). Having been previously diagnosed with depression (DAD_1.0) works against this class. The specific phobia class ( ) is most influenced by the Fear Questionnaire main level of avoidance (item 1), and the Global Phobia rating (fq_global), while having applied for both anxiety and depression treatment (APP_3.0) counts against this class.



Model Level (Global) Explainability of Score Totals
As explained, standardized coefficients obtained by multiplying the original coefficients by the SD of the corresponding feature in the dataset provide a scale-free measure of influence, allowing the comparison of the relative importance of different features. The standardized coefficients are shown in
- . Examining the PHQ-9 standardized coefficients ( ), questions 1 and 2, the main cognitive factors of the questionnaire [ ] are the most important, with all the 5 cognitive factors being represented. The importance of the GAD-7 ( ) is almost exclusively represented by questions 4 (“Trouble relaxing”) and 2 (“Not being able to stop or control worrying”). The most important SIAS questions ( ) were 6 and 1. Question 3, (“How often do you feel anxious about the possibility of having a panic attack?”) was the most important in the PDSS ( ). In the Fear Questionnaire anxiety or depression subscale ( ), question 23, “Feeling you or your surroundings are strange or unreal,” was the most important. The Total Phobia Score from the Fear Questionnaire was defined as the sum of the subscale scores for agoraphobias, blood injury phobia, and social phobia. In this case, only the learned agoraphobia subscore was weighted at nonzero ( ). Question 5 was the most relevant question in the Fear Questionnaire agoraphobia subscale ( ).






Patient Level (Local) Explainability
The relevant questionnaire total score summaries of a test patient (patient 1) are provided here. The PHQ-9 (15) score indicated moderately severe depression [
]. The PDSS score (13) indicated “moderately ill” panic disorder [ ]. The GAD-7 score (17) indicated severe anxiety [ ]. The SIAS score (22) was lower than the threshold for an indication of social phobia [ ]. The clinical assessment (ground truth) of an appropriate treatment was panic. Other scores included Fear Questionnaire total score (6), global score (0), and main score (0).Examining a local explanation at the individual patient level for patient 1 (
), the predicted probability distribution of panic (the prescribed treatment), overlapped with depression. The importance was determined from the weighted inputs to the softmax layer, combining both the patient record and the learned weights. Examining the importance associated with the prediction in the softmax layer for the panic class shows the top 6 positive factors ( ), and the top 6 negative factors ( ). Positive values make the class more probable, while negative values make the class less probable. It should be noted that the positive and negative y-values for each class are derived from the same weight layer and, therefore, are suitable for comparative analysis. In the patient 1 example, the PDSS questionnaire score explained most of the probability for the panic class, while the SIAS questionnaire detracted from the probability of the panic class, but to a lesser degree.The model can also be used to explain any of the class probabilities. It is not limited to explaining the highest probability prediction. Hence, the class prediction for depression can be examined in the same way. This showed the PHQ-9 score as the most important (
), followed by APP_3 (the patient applied for both anxiety and depression treatment). The PDSS score showed a negative influence on this class ( ), which was important but lower than the positive importance of PHQ-9. It is interesting to note that the negative importances for the depression class ( ) were larger than the negative importances for panic, the highest probability class ( ). This goes some way to explain why panic was predicted to be more probable than depression.The hierarchical model also allowed a closer examination of the individual questions. For example, an examination of the PHQ-9 questions that contributed to the probability of the depression class (
), showed PHQ-9 questions 1 and 2 as having the most influence on the model prediction for this class, for this patient.





Discussion
Overview
Uncertainty and lack of confidence in diagnosing psychopathological conditions can present a considerable challenge due to comorbidities and overlapping symptoms [
]. The diagnosis of social anxiety disorder, for example, must be differentiated from major depression, panic disorder, agoraphobia, and generalized anxiety disorder [ , ]. Furthermore, maintaining human involvement in patient assessments involving AI is critical [ ]. Therefore, explaining predictions and enabling clinicians to explore the clinical factors associated with the predictions, has been a central aspect of this work.Principal Findings
We have demonstrated the design and quantitative evaluation of a hierarchical model that aims to aid diagnosis and meet the requirements of providing explainable predictions that are context-based and easy for clinicians to use in questionnaire-based assessments [
]. Issues of confidence and robustness that are important for ML models aimed at medical DSSs [ ] have been addressed.As the model achieved a balanced accuracy score of 0.79 and AUC >0.90, it is expected that it will be clinically useful in both screening a new patient and informing a clinical interview. In patient screening, for example, the graphical interpretation of the class probabilities, their spread, and their overlap can inform the clinician of competing treatment possibilities for the patient. Even in the case where the treatment prediction is unambiguous and agrees with the clinician’s assessment, the model may serve as a confirmatory second opinion. At the clinical interview stage, the specific questions that are deemed important for the class prediction may inform the clinician of which areas to probe further in the clinical interview.
Limitations
ML models extract knowledge from data, limiting their applicability to use in scenarios that are statistically similar to the training scenario. The models reported in this study were trained on data obtained from the Internetpsykiatrien web-based treatment with nationwide coverage in Denmark [
]. Hence, while similar performance for individuals presenting to this service in Denmark can be expected, performance in a different country or otherwise different setting cannot be guaranteed. This can be seen as a limitation of ML models, but also as a strength; retraining the model described in this study is cheap. Hence, it is relatively straightforward for any regional mental health service to train this model on local data and obtain predictions that are sensitive to regional sociodemographics and patient characteristics. Moreover, patient characteristics may well change over time, and such temporal changes can be addressed by retraining the model periodically.The ground truth is based on clinical judgment and is, therefore, influenced by the subjectivity of the clinicians, interrater variability, and possibly by group dynamics in decision-making clinical conferences. Clinicians have access to free-text provided by the patient, which the model does not have. A future direction may involve the development of a hybrid model that makes use of all the relevant questionnaire and demographic and free-text data for each patient. An additional limitation of the ground truth is that the assessment relies on structured questionnaires and the Mini Neuropsychiatric Interview, which may not capture all relevant patient characteristics, leading to potential gaps in the ground truth.
During training, L1 regularization adjusts the standardized coefficients and model weights toward 0 for features deemed unimportant or redundant, for example, due to having a high correlation with another feature. Consequently, features that clinicians consider relevant may be excluded by the model and thus do not appear in the explanations of predictions. However, it is common for clinicians to prioritize the importance of certain features over others in a diagnostic context [
]. It is anticipated that clinicians will find the specificity of the explanations at the question level to be clarifying and informative, complementing their own clinical judgment. Moreover, L1 regularization may result in model instability in the sense that small changes in the training set may result in a different selection of nonzero coefficients (and thus important features). Hence, a retrained model may yield different explanations to the original model for similar individuals, and this could be confusing to clinicians. However, it should be emphasized that the feature weights that explain the predictions are the same feature weights that the model uses to make the predictions, due to the model’s transparency. However, using L1 regularization is not a prerequisite for the use of the proposed approach. L2 regularization, which attempts to achieve good generalization by keeping model weights small (rather than forcing a subset to 0 such as in L1 regularization) can also be used. The result would be that the model contains mostly nonzero weights, and thus, the explainability of model decisions becomes more complex. However, clinicians could choose to focus on the most important features (ie, those with the highest influence on the model). Visualizations, such as those in - , could be constrained to the N most important features with N chosen by the clinician.The clinical use of the ML model would require the development of a DSS so that clinicians can view and explore the model predictions and explanations when provided with patient records. The lack of such a DSS may be seen as a limitation in demonstrating the clinical effectiveness of the model. While this study does not empirically demonstrate that the aims of clinical interpretability have been met, further work is planned to develop the DSS and study the clinical usefulness of the system incorporating the model.
Overall, the results highlight the key challenges in differentiating social phobia from other anxiety disorders and suggests that future improvements could focus on enhancing feature differentiation for overlapping categories, particularly social phobia and panic disorder.
Future Work
The incorporation of this model into a DSS for mental health treatment prediction will be an important future step for this work. Once developed, assessing its use in a clinical setting will help to validate it for future clinical use. The model itself could be extended to use more input features, especially those that have a potential to enhance feature differentiation for overlapping categories (particularly social phobia and panic disorder). Additional hybrid features may include free-text data input for each patient. Output enhancements can also be foreseen to further build on the model’s explainability. For example, a text output of the model’s prediction, competing predictions, and explanations in written form could form an important input to clinical notes and provide material for clinical conferences.
Conclusions
This paper presents the design of a model aimed at a DSS for psychopathological mental health treatment. Its basis in standard screening questionnaires will allow seamless adoption in a clinical setting in the future. The AI model incorporated in the system is based on a novel hierarchical structure that aims to meet the needs of clinical interpretability while accounting for decision confidence and uncertainty that are lacking in previous work.
Acknowledgments
This work has been supported by the Innovation Fund, Denmark. The authors wish to thank the clinical staff, administrative staff, and patients at the Centre for Digital Psychiatry, Denmark.
Authors' Contributions
AK conceived the probabilistic model; designed, trained, and verified the models; and drafted the paper. PVdV conceived the hierarchical score model, made significant contributions to the paper’s Introduction and Methods sections, and improved all sections of the paper. EMG made contributions to the explainable artificial intelligence literature description. EKJ and KM made contributions to the Data and Introduction sections as well as provided the clinical input, data, infrastructure, and support for the project. All authors reviewed the completed paper.
Conflicts of Interest
None declared.
Details of the model layers and structure.
DOCX File , 21 KBReferences
- Minerva F, Giubilini A. Is AI the future of mental healthcare? Topoi (Dordr). May 31, 2023;42(3):1-9. [FREE Full text] [CrossRef] [Medline]
- Meehl PE. Clinical Versus Statistical Prediction: A Theoretical Analysis and a Review of the Evidence. Brattleboro, VT. Echo Point Books and Media; 2013.
- Ægisdóttir S, White MJ, Spengler PM, Maugherman AS, Anderson LA, Cook RS, et al. The meta-analysis of clinical judgment project: fifty-six years of accumulated research on clinical versus statistical prediction. Counsel Psychol. May 01, 2006;34(3):341-382. [FREE Full text] [CrossRef]
- Katsikopoulos K, Machery E, Pachur T, Wallin A. The search for models of clinical judgment: fast, frugal, and friendly in Paul Meehl’s spirit. J Behav Decis Mak. 2004. [FREE Full text]
- Makivić I, Kragelj A, Lasalvia A. Quantitative needs assessment tools for people with mental health problems: a systematic scoping review. Curr Psychol. Mar 12, 2024;43(22):19996-20013. [FREE Full text] [CrossRef]
- Morrison J. Diagnosis Made Easier: Principles and Techniques for Mental Health Clinicians. New York, NY. Guilford Press; 2024.
- Mogi T, Toda H, Yoshino A. Clinical characteristics of patients with diagnostic uncertainty of major depressive disorder. Asian J Psychiatr. Dec 2017;30:159-162. [FREE Full text] [CrossRef] [Medline]
- Souery D, Papakostas GI, Trivedi MH. Treatment-resistant depression. J Clin Psychiatry. 2006;67 Suppl 6:16-22. [Medline]
- Bloom DE, Cafiero ET, Jané-Llopis E, Abrahams-Gessel S, Bloom LR, Fathima S, et al. The global economic burden of noncommunicable diseases. World Economic Forum. 2011. URL: https://www3.weforum.org/docs/WEF_Harvard_HE_GlobalEconomicBurdenNonCommunicableDiseases_2011.pdf [accessed 2024-05-30]
- Wittchen HU, Jacobi F, Rehm J, Gustavsson A, Svensson M, Jönsson B, et al. The size and burden of mental disorders and other disorders of the brain in Europe 2010. Eur Neuropsychopharmacol. Sep 2011;21(9):655-679. [FREE Full text] [CrossRef] [Medline]
- D'Alfonso S. AI in mental health. Curr Opin Psychol. Dec 2020;36:112-117. [FREE Full text] [CrossRef] [Medline]
- Graham S, Depp C, Lee EE, Nebeker C, Tu X, Kim HC, et al. Artificial intelligence for mental health and mental illnesses: an overview. Curr Psychiatry Rep. Nov 07, 2019;21(11):116. [FREE Full text] [CrossRef] [Medline]
- Martin-Key NA, Spadaro B, Funnell E, Barker EJ, Schei TS, Tomasik J, et al. The current state and validity of digital assessment tools for psychiatry: systematic review. JMIR Ment Health. Mar 30, 2022;9(3):e32824. [FREE Full text] [CrossRef] [Medline]
- Thakkar A, Gupta A, De Sousa A. Artificial intelligence in positive mental health: a narrative review. Front Digit Health. 2024;6:1280235. [FREE Full text] [CrossRef] [Medline]
- Ghassemi M, Oakden-Rayner L, Beam A. The false hope of current approaches to explainable artificial intelligence in health care. Lancet Digit Health. Nov 2021;3(11):e745-e750. [FREE Full text] [CrossRef]
- Ali S, Abuhmed T, El-Sappagh S, Muhammad K, Alonso-Moral JM, Confalonieri R, et al. Explainable Artificial Intelligence (XAI): what we know and what is left to attain Trustworthy Artificial Intelligence. Inf Fusion. Nov 2023;99:101805. [FREE Full text] [CrossRef]
- Singh RK, Gorantla R, Allada SG, Narra P. SkiNet: a deep learning framework for skin lesion diagnosis with uncertainty estimation and explainability. PLoS One. 2022;17(10):e0276836. [FREE Full text] [CrossRef] [Medline]
- Barredo Arrieta A, Díaz-Rodríguez N, Del Ser J, Bennetot A, Tabik S, Barbado A, et al. Explainable Artificial Intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI. Inf Fusion. Jun 2020;58:82-115. [FREE Full text] [CrossRef]
- Kompa B, Snoek J, Beam AL. Second opinion needed: communicating uncertainty in medical machine learning. NPJ Digit Med. Jan 05, 2021;4(1):4. [FREE Full text] [CrossRef] [Medline]
- Gunning D, Vorm E, Wang JY, Turek M. DARPA's explainable AI (XAI) program: a retrospective. Appl AI Letters. Dec 16, 2021;2(4):e61. [FREE Full text] [CrossRef]
- Explainable AI: the basics. The Royal Society. Nov 2019. URL: https://royalsociety.org/-/media/policy/projects/explainable-ai/AI-and-interpretability-policy-briefing.pdf [accessed 2024-05-30]
- TechDispatch #2/2023 - explainable artificial intelligence. European Data Protection Supervisor. 2023. URL: https://www.edps.europa.eu/data-protection/our-work/publications/techdispatch/2023-11-16-techdispatch-22023-explainable-artificial -intelligence_en [accessed 2024-05-30]
- Ethics guidelines for trustworthy AI. European Commission. Apr 8, 2019. URL: https://digital-strategy.ec.europa.eu/en/library/ethics-guidelines-trustworthy-ai [accessed 2024-05-30]
- Subbaswamy A, Adams R, Saria S. Evaluating model robustness and stability to dataset shift. arXiv. Preprint posted online on October 28, 2020. [FREE Full text]
- Begoli E, Bhattacharya T, Kusnezov D. The need for uncertainty quantification in machine-assisted medical decision making. Nat Mach Intell. Jan 07, 2019;1(1):20-23. [FREE Full text] [CrossRef]
- Wanner J, Herm LV, Heinrich K, Janiesch C. Stop ordering machine learning algorithms by their explainability! An empirical investigation of the tradeoff between performance and explainability. In: Proceedings of the Conference on e-Business, e-Services and e-Society. 2021. Presented at: I3E 2021; September 1-3, 2021; Galway, Ireland. URL: https://doi.org/10.1007/978-3-030-85447-8_22/FIGURES/3 [CrossRef]
- Ribeiro MT, Singh S, Guestrin C. "Why should I trust you?": explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2016. Presented at: KDD '16; August 13-17, 2016; San Francisco, CA. URL: https://doi.org/10.1145/2939672.2939778 [CrossRef]
- Zhang Y, Song K, Sun Y, Tan S, Udell M. "Why should you trust my explanation?" Understanding uncertainty in lime explanations. arXiv. Preprint posted online on April 29, 2019. [FREE Full text]
- Lakkaraju H, Kamar E, Caruana R, Leskovec J. Interpretable and explorable approximations of black box models. arXiv. Preprint posted online on July 4, 2017. [FREE Full text]
- Alvarez-Melis D, Jaakkola TS. On the robustness of interpretability methods. arXiv. Preprint posted online on June 21, 2018. [FREE Full text]
- Abeyagunasekera SH, Perera Y, Chamara K, Kaushalya U, Sumathipala P, Senaweera O. LISA : enhance the explainability of medical images unifying current XAI techniques. In: Proceedings of the IEEE 7th International conference for Convergence in Technology. 2022. Presented at: I2CT 2022; April 07-09, 2022; Mumbai, India. URL: https://doi.org/10.1109/I2CT54291.2022.9824840 [CrossRef]
- Singh Dhaliwal B, Imran R, Leung CK, Evan EW. Trustworthy explanations for knowledge discovered from e-health records. In: Proceedings of the IEEE International Conference on E-health Networking, Application & Services. 2022. Presented at: HealthCom 2022; October 17-19, 2022; Genoa, Italy. URL: https://doi.org/10.1109/HEALTHCOM54947.2022.9982786 [CrossRef]
- Chaddad A, Peng J, Xu J, Bouridane A. Survey of explainable AI techniques in healthcare. Sensors (Basel). Jan 05, 2023;23(2):634. [FREE Full text] [CrossRef] [Medline]
- Ovadia Y, Fertig E, Ren J, Nado Z, Sculley D, Nowozin S, et al. Can you trust your model's uncertainty? Evaluating predictive uncertainty under dataset shift. arXiv. Preprint posted online on June 6, 2019. [FREE Full text]
- Ghahramani Z. Probabilistic machine learning and artificial intelligence. Nature. May 28, 2015;521(7553):452-459. [FREE Full text] [CrossRef] [Medline]
- Cardelli L, Kwiatkowska M, Laurenti L, Patane A. Robustness guarantees for Bayesian inference with gaussian processes. Proc AAAI Conf Artif Intell. Jul 17, 2019;33(01):7759-7768. [FREE Full text] [CrossRef]
- Winter S, Campbell T, Lin L, Srivastava S, Dunson DB. Machine learning and the future of Bayesian computation. arXiv. Preprint posted online on April 21, 2023. [FREE Full text]
- Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R. Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res. 2014;15(1):1929-1958.
- Gal Y, Ghahramani Z. Dropout as a Bayesian approximation: representing model uncertainty in deep learning. arXiv. Preprint posted online on June 6, 2015. [FREE Full text]
- Wang G, Li W, Aertsen M, Deprest J, Ourselin S, Vercauteren T. Aleatoric uncertainty estimation with test-time augmentation for medical image segmentation with convolutional neural networks. Neurocomputing (Amst). Sep 03, 2019;335:34-45. [FREE Full text] [CrossRef] [Medline]
- Mathiasen K, Riper H, Andersen TE, Roessler KK. Guided internet-based cognitive behavioral therapy for adult depression and anxiety in routine secondary care: observational study. J Med Internet Res. Nov 28, 2018;20(11):e10927. [FREE Full text] [CrossRef] [Medline]
- Kroenke K, Spitzer RL, Williams JB. The PHQ-9: validity of a brief depression severity measure. J Gen Intern Med. Sep 2001;16(9):606-613. [FREE Full text] [CrossRef] [Medline]
- Spitzer RL, Kroenke K, Williams JB, Löwe B. A brief measure for assessing generalized anxiety disorder: the GAD-7. Arch Intern Med. May 22, 2006;166(10):1092-1097. [FREE Full text] [CrossRef] [Medline]
- Mattick RP, Clarke JC. Development and validation of measures of social phobia scrutiny fear and social interaction anxiety. Behav Res Ther. Apr 1998;36(4):455-470. [FREE Full text] [CrossRef] [Medline]
- Houck PR, Spiegel D, Shear M, Rucci P. Reliability of the self-report version of the panic disorder severity scale. Depress Anxiety. 2002;15(4):183-185. [FREE Full text] [CrossRef] [Medline]
- Marks I, Mathews A. Brief standard self-rating for phobic patients. Behav Res Ther. 1979;17(3):263-267. [CrossRef] [Medline]
- Priebe S, Huxley P, Knight S, Evans S. Application and results of the Manchester Short Assessment of Quality of Life (MANSA). Int J Soc Psychiatry. 1999;45(1):7-12. [FREE Full text] [CrossRef] [Medline]
- Boluki S, Ardywibowo R, Dadaneh SZ, Zhou M, Qian X. Learnable bernoulli dropout for Bayesian deep learning. arXiv. Preprint posted online on February 12, 2020. [FREE Full text]
- Jospin LV, Laga H, Boussaid F, Buntine W, Bennamoun M. Hands-on Bayesian neural networks—a tutorial for deep learning users. IEEE Comput Intell Mag. May 2022;17(2):29-48. [FREE Full text] [CrossRef]
- Saarela M, Jauhiainen S. Comparison of feature importance measures as explanations for classification models. SN Appl Sci. Feb 03, 2021;3(2):1. [FREE Full text] [CrossRef]
- Menard S. Six approaches to calculating standardized logistic regression coefficients. Am Stat. Aug 2004;58(3):218-223. [FREE Full text] [CrossRef]
- Forsell E, Jernelöv S, Blom K, Kaldo V. Clinically sufficient classification accuracy and key predictors of treatment failure in a randomized controlled trial of internet-delivered cognitive behavior therapy for insomnia. Internet Interv. Sep 2022;29:100554. [FREE Full text] [CrossRef] [Medline]
- Boothroyd L, Dagnan D, Muncer S. PHQ-9: one factor or two? Psychiatry Res. Jan 2019;271:532-534. [FREE Full text] [CrossRef] [Medline]
- Keough ME, Porter E, Kredlow MA, Worthington JJ, Hoge EA, Pollack MH, et al. Anchoring the panic disorder severity scale. Assessment. Jun 2012;19(2):257-259. [FREE Full text] [CrossRef] [Medline]
- Sapra A, Bhandari P, Sharma S, Chanpura T, Lopp L. Using generalized anxiety disorder-2 (GAD-2) and GAD-7 in a primary care setting. Cureus. May 21, 2020;12(5):e8224. [FREE Full text] [CrossRef] [Medline]
- Peters L. Discriminant validity of the Social Phobia and Anxiety Inventory (SPAI), the Social Phobia Scale (SPS) and the Social Interaction Anxiety Scale (SIAS). Behav Res Ther. Sep 2000;38(9):943-950. [CrossRef] [Medline]
- Moutier CY, Stein MB. The history, epidemiology, and differential diagnosis of social anxiety disorder. J Clin Psychiatry. 1999;60 Suppl 9:4-8. [Medline]
- Olawade DB, Wada OZ, Odetayo A, David-Olawade AC, Asaolu F, Eberhardt J. Enhancing mental health with artificial intelligence: current trends and future prospects. J Med Surg Public Health. Aug 2024;3:100099. [FREE Full text] [CrossRef]
- Arroll B, Goodyear-Smith F, Crengle S, Gunn J, Kerse N, Fishman T, et al. Validation of PHQ-2 and PHQ-9 to screen for major depression in the primary care population. Ann Fam Med. 2010;8(4):348-353. [FREE Full text] [CrossRef] [Medline]
Abbreviations
AI: artificial intelligence |
AUC: area under the curve |
DSS: decision support system |
GAD-7: Generalized Anxiety Disorder Questionnaire-7 |
LIME: local interpretable model-agnostic explanations |
MCD: Monte Carlo dropout |
ML: machine learning |
PDSS: Panic Disorder Severity Scale |
PHQ-9: Patient Health Questionnaire-9 |
SHAP: Shapley additive explanations |
SIAS: Social Interaction Anxiety Scale |
XAI: explainable artificial intelligence |
Edited by C Lovis; submitted 22.07.24; peer-reviewed by P Bjurner, A Peimankar, A Dasgupta; comments to author 25.01.25; revised version received 08.02.25; accepted 16.02.25; published 26.03.25.
Copyright©Anthony Kelly, Esben Kjems Jensen, Eoin Martino Grua, Kim Mathiasen, Pepijn Van de Ven. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 26.03.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.