Abstract
Background: Efforts are being made to integrate digital health technologies into clinical care for multiple sclerosis (MS) to improve patient monitoring. Efficiently probing how they might impact clinical care could streamline digital tool development. The Floodlight digital tool, comprising 5 smartphone sensor–based tests, was used to generate health-related data on patient function and symptoms in a clinical simulation.
Objective: The study had 3 objectives: (1) assess the utility of simulated clinical encounters as a research methodology for exploring the introduction of digital health technologies into clinical practice in MS, (2) confirm the fidelity of the simulated environment and patient cases developed and understand what metrics (eg, workflow, comprehensive evaluation) could be generated, and (3) generate insights into the utility of digitally collected data, including usability, clinical decision contribution, and impact on workflows, in clinical practice.
Methods: A total of 2 patient cases consisting of clinical, radiological, and digital health data were developed with clinician input. US-based neurologists prepared for and conducted 2 simulated teleconsultations each, with an actor briefed on case profiles. Floodlight data were available, via the Floodlight MS™ Health Care Professional Portal, for 1 of the 2 consultations. Participant neurologists completed interviews and surveys assessing the fidelity of the cases presented, user experience and workflow metrics, patient concerns identified, care decisions made, and confidence in making decisions.
Results: All 10 neurologists indicated that the simulations were high-fidelity representations of real consultations. Using the Floodlight technology for the first time, median time taken to prepare for and conduct the consultation was ~1.7‐2 minutes longer, with slightly greater mental effort reported by participants, compared with not using the tool. The Floodlight MS Health Care Professional Portal scored an “above average” 79 on the System Usability Scale and an “acceptable” Net Promoter Score of 10. In total, 6 of the 10 neurologists “strongly agreed” that it was easier and quicker to identify patient concerns when they had access to the patient-generated Floodlight data to prepare for their encounters than when they did not. Overall, more care and management decisions were taken when the digital tool was used (37 vs 29). Of those 37 decisions, Floodlight data were reported as a trigger for 20 decisions, always in combination with other elements including patient history (20/20) and clinical exam findings (9/20).
Conclusions: These findings advance our understanding of clinical simulation as a method for evaluating digital tools and other innovative technologies for MS care. High-fidelity patient cases could be provided for the mock teleconsultations, and the simulated clinical environment was useful for evaluating usability and utility of a new digital tool—yielding preliminary evidence on how digital data could be accessed and utilized by neurologists to support routine MS care.
doi:10.2196/67845
Keywords
Introduction
Background
Multiple sclerosis (MS) is a central nervous system disorder marked by significant heterogeneity in its clinical presentation and characterized by disease progression from onset [,]. MS, often commencing with subtle signs and advancing to substantial disability, manifests through a wide spectrum of symptoms, ranging from sensory disturbances and motor deficits to cognitive impairments and emotional challenges, profoundly impacting patients, their families, and society [,].
To date, monitoring in MS has been reliant upon brief, infrequent clinical assessments (typically occurring once or twice annually, or at the time of a relapse), without sufficiently sensitive and objective (ie, free from rater-dependent variability) assessments of progression available to health care professionals (HCPs) []. Moreover, due to time constraints, standard assessments of disability in MS, such as the Expanded Disability Status Scale (EDSS) [], which requires ~30 minutes to conduct [], are not always performed in clinical appointments. Remote consultations (teleconsultations) are a recognized alternative to clinic visits [] and have been adopted by the MS community, particularly during the COVID-19 pandemic [,].
Digital health technologies offer greater access to care in MS and opportunities to improve frequent monitoring [], enabling the capture of ongoing subtle or subclinical progression of MS [,-] in specific domains such as cognition (which can often be missed). These technologies can be remotely administered and are patient-centered and operated, allowing longitudinal assessment of multiple functional domains in a naturalistic environment [,]. Furthermore, they offer the opportunity to improve the involvement of patients in their disease or treatment course [,] and to enhance conversations between HCPs and people with MS []. By providing objective data and bridging the gap of information in between clinical visits, sensor-based digital assessments may ultimately allow HCPs to make more informed decisions, evaluate treatment effects, and act earlier on indications of disease progression [] or symptomatic exacerbation prior to scheduled clinical visits.
Despite the potential benefits, evaluating, testing, and optimizing the integration of new and innovative solutions into health care settings can itself be a time- and resource-consuming process. Hence, there is growing demand for agile methods to rapidly learn how digital solutions can play a role in clinical practice []. Clinical simulation, a low-cost and scalable methodology whereby data are collected from users as they perform tasks in realistic clinical scenarios, can rapidly generate new insights. For example, they allow the exploration of the possible impact of novel methodologies without the risks inherent in deploying them in real clinical settings (eg, of misdiagnoses [,]) or of the expansion of technology to new user groups or in new clinical contexts []. The preliminary insights generated from simulation models can support the refinement of technologies before rollout or scaling up []. To confirm that the insights gained from these clinical simulations are meaningful, they must be of high fidelity and account for a number of variables (including behavioral ones) and the differences inherent among real-world practices [,].
Aims and Objectives
This feasibility study was designed to evaluate clinical simulation as a method for assessing the preparation and conduct of teleconsultations. This included evaluating the fidelity of the simulated environment (compared with real practice) and the patient cases developed. An additional aim was to assess the feasibility of using data derived from smartphone-based remote assessment technology in combination with MS teleconsultations, including their usability and impacts on clinical workflows (eg, preparation time and time to identifying primary patient concerns), care decisions, and disease management. The Floodlight technology [] was utilized as the digital tool for evaluation. Accordingly, the aim of this study was to evaluate the potential impact of integrating Floodlight MS into clinical practice, generating insights that may also be relevant to similar digital health tools. As it is expected that any new test or technology will initially impose a higher mental load on clinicians, the study also aimed to quantify the extent of this increase.
Floodlight MS comprised 5 smartphone sensor–based tests (hereafter referred to as “smartphone-based tests”) presented on an app, measuring information processing speed (Cognitive Test []), hand function (Pinching Test [] and Draw a Shape Test []), and gait (U-Turn Test [] and 2-Minute Walk Test [,]), as well as a Patient Journal including 2 patient-reported outcome (PRO) measures (Daily Status Questions and a Symptom Tracker []). The data were consolidated in a dashboard for HCPs (Floodlight MS HCP Portal), which was accessible between clinical visits [], and constituted the focus of the technological assessment in this research.
Methods
Recruitment of Neurologists
US-based neurologists were recruited from Prova Health’s existing network, professional communities (eg, LinkedIn), and academic publications, and through snowball sampling (where research participants recruit future participants), and they were sent email invitations to take part in the study. Neurologists were required to have managed a minimum of 40 people with MS per month and to have had no previous experience using the Floodlight technology. Convenience sampling was utilized to recruit primarily academic neurologists, as they were expected to give more granular feedback on the technology. There was no defined sample size due to the exploratory nature of this feasibility study.
Patient Case Overview
A total of 2 patient cases were presented to each participant neurologist. Radiology reports and clinical letters were drafted by qualified radiologists and medical doctors to describe the cases and reflect the real magnetic resonance imaging and clinical data in a realistic narrative. For additional authenticity, synthetic social and family histories consistent with typical MS presentations were included. All cases were validated by 2 expert neurologists.
The goal was to achieve high-fidelity patient cases that were representative of real-world cases while minimizing the variation inherent among people with MS. Patient cases included (experimental) or did not include (control) Floodlight data, thus simulating real-world scenarios in which neurologists manage people with MS with or without access to the Floodlight technology (see the Patient Case Development section for more details).
Patient Actor
A professional actor was recruited to play the patients’ roles during the teleconsultations. They were recruited from a company with 25 years of experience in providing trained actors to simulate patients with medical, surgical, and psychiatric conditions. The patient roles were based on generated behavioral profiles, which were used to inform the biographies for the 2 patient cases. The actor was provided with detailed scripts developed by a medical doctor and validated by neurologists to ensure [] the consistency and correctness of answers.
Conduct of Clinical Simulations
Each clinical simulation session lasted approximately 2.5 hours and was conducted via Zoom. Medically qualified team members from Prova Health served as moderators, facilitating introductions (~5 min), reviewing the simulation session guide with the neurologists (~5 min), and providing training on the Floodlight technology (~30 min). After the simulated teleconsultations had begun, the moderators took notes with their cameras and microphones switched off. Teleconsultations were designed to replicate those that the participant neurologists would perform in their clinical practices. The individual simulated teleconsultation sessions for each of the 2 presented patient cases consisted of preparation (~5 min) and conduct (~30 min) phases, both led by the participant neurologist. Participant neurologists were instructed to evaluate the relevant information and recommend an appropriate management plan. In the preparation phase, clinical information provided with or without Floodlight data was reviewed for each patient case. The “patient” (played by the professional actor) then joined the call. Following the conclusion of each simulated teleconsultation, participant neurologists completed a short survey (~5 min) to collect feedback on the examined case. At the end of both sessions, the moderator shared a link to the postsimulation survey (~7 min) and conducted a brief interview (~5 min) with each participant neurologist.
Patient Case Development
The 2 patient cases presented to each participant neurologist were developed using deidentified clinical and Floodlight data (see the Patient Cases Selection Criteria section). For both cases, the participant neurologists reviewed data from 3 recent clinical visits, including patient demographics, symptoms, diagnosis, treatment and medical history, and radiology reports based on original magnetic resonance imaging metrics.
The clinical data presented in the letters included EDSS scores, Symbol Digit Modalities Test (SDMT) scores, 9-Hole Peg Test times, and Timed 25-Foot Walk times. The main clinical traits emerging from clinical case narratives are presented in . It should be noted that although both patient cases presented gait impairment, Patient Case B was designed to show significant worsening of gait impairment in the last year.
Floodlight data visualized in the Floodlight MS HCP Portal were presented for 1 of the 2 cases reviewed by each neurologist. The orders of the patient cases and of the interventions were randomized before the participant neurologists started the teleconsultations ().
| Simulation patient case | Main clinical traits emerging from clinical case narratives and actor’s interpretation |
| Patient Case A |
|
| Patient Case B |
|
aSDMT: Symbol Digit Modalities Test.
b9HPT: 9-Hole Peg Test.
cT25FW: Timed 25-Foot Walk.

Data Sources
Clinical data from the CONSONANCE study (NCT03523858) were used in conjunction with Floodlight performance and symptom tracker data collected over 12 consecutive months. CONSONANCE is an ongoing, phase 3b, 4-year, open-label study assessing the effectiveness and safety of ocrelizumab in adults with progressive multiple sclerosis.
Participants enrolled in the CONSONANCE study performed smartphone-based tests using a precursor to Floodlight MS for clinical research purposes, providing a battery of digital assessments and digital PROs. Data from the Pinching Test, the Draw a Shape Test, the U-Turn Test, and the 2-Minute Walk Test (all performed daily) were utilized, alongside data from a version of the Cognitive Test (performed weekly; involving matching a sequence of symbols to digits within 90 s).
Data from 2 PRO measures collected via the Floodlight Patient Journal were also used. These were Daily Status Questions (“How are you feeling physically now?” and “How is your mood now?”) and a biweekly Symptom Tracker to track symptoms related to cognition (attention deficit, memory loss, brain “fog” sensation, mental fatigue), upper extremity function (sensory disturbance, muscle weakness, clumsiness, muscular spasm, pain), lower extremity function (sensory disturbance, muscle weakness, clumsiness, muscular spasm, pain), gait (trouble walking, lack of balance, walking fatigue), and bowel or bladder dysfunction (incontinence, bladder or urinary problems, constipation), as well as the time that any new or worsening symptoms manifested and whether or not this was perceived to be a relapse. All Floodlight data collected in the CONSONANCE trial and used in this study were imported into the Floodlight MS HCP Portal (version 1.5 at the time of the research).
Patient Cases Selection Criteria
Patient profiles were selected from the CONSONANCE dataset based on selected criteria that indicated a clinical performance decline observed in the previous 6‐12 months, based on standard clinical assessments of cognitive function, hand motor function, or gait capacity.
For Patient Case A, the key selection criteria included cognitive function decline (a decrease in SDMT score of ≥20% in the previous 12 mo) and hand motor function decline (increase in 9-Hole Peg Test time of ≥20% in the previous 6‐12 mo). For Patient Case B, the criteria included gait capacity decline (increase in Timed 25-Foot Walk time of ≥20% in the previous 6‐12 mo) as well as 2 consecutive reports of fatigue and bladder issues (score of 3‐6; some of the most impactful “invisible” symptoms of MS sourced via the SymptoMScreen). Some of the symptoms experienced by people living with MS may not be externally evident or easy to report and quantify, such as fatigue, mood disorders, cognitive impairments, pain, bladder or bowel dysfunction, sexual dysfunction, and vision changes. These symptoms can be difficult for the patient to report and articulate during a clinical encounter (eg, due to subjectivity bias, poor recall, or stigma), and for physicians to capture (limitations in available measurement tools, ample reliance on patient recall on the visit day) []. The clinical performance decline, together with reported signs of “invisible” symptoms, provided impetus for a care decision to be taken during the clinical simulation.
Data Extraction
Clinical data from 3 recent in-person clinical visits were extracted to present to the participant neurologists (). No clinical data from the final in-person visit (the time point corresponding to the day of the teleconsultation in the simulation) were provided. Of note, cognitive decline had been confirmed at this visit for the patient in Patient Case A (as measured by the SDMT); these data were not provided to the participant neurologists in order to mimic a common clinical scenario in MS, where evident patient deterioration may precede formal confirmation through standard clinical assessments.

To build the patient cases, MS phenotype and date of diagnosis; clinical parameters, including recent disease history (eg, relapses, EDSS evolution); symptom reports; and current and prior treatments (eg, disease-modifying therapies) were also extracted. Additional “synthetic” data were used to construct demographic (age, sex) and radiology reports.
Floodlight data from the prior 12-month period for each clinical profile were also extracted to be provided for 1 of the 2 cases presented to each participant neurologist.
Study Outcomes and Data Collection
Study outcomes were either moderator-reported or participant neurologist–reported (via participant surveys and interviews; ). Measures included questions regarding the fidelity of the simulated cases developed; use of the digital tool (Floodlight MS), including user experience metrics for data visualization, overall experience (Net Promoter Score), and usability (System Usability Scale []); and workflow metrics, including time for case preparation, teleconsultation time, and mental effort required (Paas scale []). Confidence in approaching clinical consultations was also measured. Additionally, moderators recorded the time taken for the identification of patient concerns, the recognition of “invisible” MS symptoms, and the care decisions prompted and taken (including change in treatment management) with versus without having Floodlight data available.
A medically qualified moderator from the research team observed each simulation session and recorded outcome measures such as time spent on each case. The participants were asked to self-report on the fidelity of simulation, user experience metrics, workflow metrics, clinical consultations, patient concerns, and care decisions. This was done using postcase surveys, postcase conduct surveys, postsimulation surveys, or a postsimulation interview. Each of these was measured using various questions.
| Measure | Moderator measured | Participant neurologist reported | |||
| Survey after case preparation | Survey after case conduct | Survey after simulation | Interview after simulation | ||
| Fidelity of simulation | |||||
| Similarity of patient cases to those seen in day-to-day practice | X | ||||
| Similarity of simulation to day-to-day practice consultations | X | ||||
| Using clinical simulation to evaluate the Floodlight MS HCP Portal | X | X | |||
| User experience metrics | |||||
| User feedback on the Floodlight MS HCP Portal | X | X | |||
| Usability of the Floodlight MS HCP Portal | X | ||||
| Workflow metrics | |||||
| Time taken to prepare for teleconsultation | X | ||||
| Time taken to conduct teleconsultation | X | ||||
| Reported mental effort associated with information acquisition and application | X | X | |||
| Clinical consultations, patient concerns, and care decisions | |||||
| Reported confidence level when preparing for and making care decisions during a teleconsultation | X | X | |||
| Identification of the patient’s primary concern | X | X | X | ||
| Time taken to identify the patient’s primary concern during teleconsultation (independently verified by a second clinician using screen recordings) | X | ||||
| Recognition of “invisible symptoms of MS” and their proactive discussion with patients | X | X | |||
| Care decisions taken and reported prompts for these decisions | X | X | X | ||
aMS: multiple sclerosis.
bHCP: health care professional.
cInvisible symptoms, including fatigue, mood disorders, cognitive impairments, pain, bladder or bowel dysfunction, sexual dysfunction, and vision changes can be experienced by people living with MS. Some of these symptoms may not be externally evident or easy to report and quantify and can be difficult for the patient to report and articulate during a clinical encounter (eg, due to subjectivity bias, poor recall, or stigma) or for physicians to capture (eg, limitations in available measurement tools, ample reliance on patient recall on the visit day).
Data Analysis
For continuous, time-dependent, moderator-assessed measures, the median times were calculated and presented as box and whisker diagrams. Additionally, these time-dependent measures were collated into groups, and all categorical results were presented as absolute numbers within each category and presented as bar charts. An a priori decision was made not to perform hypothesis-testing statistics on data from this feasibility study.
Ethical Considerations
CONSONANCE patient data anonymity was retained throughout the preparation and conduct of this study. The protocol (22-PROH-102: Evaluation of Floodlight MS using clinical simulation) was reviewed by Pearl Institutional Review Board on September 20, 2022, and determined to be exempt from ethics approval according to U.S. Food and Drug Administration 21 Code of Federal Regulations (CFR) 56.104 and 45CFR46.104(b)(3): (3) Benign Behavioral Interventions - Adults. The study was conducted in accordance with the most recent version of the Declaration of Helsinki. Signed consent forms were obtained from the participant neurologists before the simulated teleconsultations were conducted. Participant neurologists were compensated for their time according to local Fair Market Value rates.
Results
Participant Neurologist Demographics
A total of 10 US-based neurologists were recruited as participants (7 from large academic centers and 3 from specialist neurology centers). A total of 6 participant neurologists had >10 years of experience in performing MS consultations, 1 had 5‐10 years of experience, and 3 had 3‐5 years of experience. The number of consultations performed by participant neurologists in their current practice in a typical week was >50 (n=2), 40‐50 (n=1), 30‐40 (n=1), 20‐30 (n=3), or 10‐20 (n=3). Further, participant neurologists performed 25%‐50% (n=3) or <25% (n=7) of all consultations as teleconsultations in their current practice.
Patient Case and Simulation Fidelity
Participant neurologists felt that the simulations were high-fidelity representations of real consultations. All reported that the patient cases and scenarios were “very” or “somewhat” similar to typical practice (). Feedback from participants on the performance of the professional patient actor was mostly very positive:
Shockingly. He was very good. I say that. I haven’t done this in a while, but very often a mock patient doesn’t get the real essence of MS. This guy did great. I even wondered, did he have MS, how he got so good at it? ... They would have had to practice and discuss things with many patients and doctors to get this good.
The actor did a great job. Intuitively knew how to pick up on what I was asking.
| Participant neurologists’ experiences | Number of participant neurologists | ||||
| Very different | Somewhat different | Neutral | Somewhat similar | Very similar | |
| How similar were these patient cases to those you see in your day-to-day practice? | 0 | 0 | 0 | 4 | 6 |
| How similar were these consultations to the consultations you complete during a typical week? | 0 | 0 | 0 | 5 | 5 |
aParticipant neurologists reported similarity of patient cases to those seen in day-to-day practice and similarity of simulation to day-to-day practice consultations via surveys and questionnaires.
User Experience Metrics
The Floodlight MS HCP Portal scored an “above average” score of 79 on the System Usability Scale (). In terms of overall experience of using the Floodlight technology, the Floodlight MS HCP Portal scored an “acceptable” Net Promoter Score of 10, based on the participant neurologists’ reports of how likely they would be to recommend it to a colleague ().
Participant neurologists agreed that data visualizations in the Floodlight MS HCP Portal were “very” or “somewhat” useful for upper extremity function (10/10), cognition (10/10), gait (9/10), and mood status (obtained via the Floodlight Daily Status Questions; 9/10; ). The “mood status” aspect of the Patient Journal was rated as the most useful of the data visualizations presented, with 8/10 ratings of “very useful”.


Workflow Metrics
In terms of the time taken for patient case preparation, cases took a median of 7 minutes and 36 seconds with Floodlight and 5 minutes and 57 seconds without Floodlight (). The median time taken to conduct the teleconsultations was 22 minutes with Floodlight and 20 minutes without Floodlight (). When using the Floodlight technology, participant neurologists also reported investing a slightly higher level of mental effort to prepare for (; very high mental effort to rather high mental effort: n=3 vs n=2) and conduct patient cases (; very, very high mental effort to rather high mental effort: n=6 vs n=5).

Clinical Consultations and Patient Concerns
About 6 out of the 10 neurologists “strongly agreed” that it was easier to identify patient concerns within the relevant functional domains with access to Floodlight data during the consultation preparation phase versus without access (where 2/10 “strongly agreed”; ). The median time taken for participant neurologists to start discussing the primary domain of concern was 2 minutes, 14 seconds when using the Floodlight technology and 2 minutes, and 17 seconds without the Floodlight technology ().

Recognition of “Invisible” Symptoms of MS
The median time taken to discuss each “invisible” symptom was between 1 and 1.5 minutes (). In Patient Case A, which was structured to show a prominent cognitive decline, cognitive issues, and mood change, discussions exceeded 2 minutes. In Patient Case B, bladder dysfunction (which was referred to in the clinical letters and without the support of Floodlight information) was also discussed for over 2 minutes. While differences can be gleaned at the individual symptom level, there was no clear pattern to the length of the discussions. Being aware of the symptom by means of the clinical letters did not necessarily seem to be a clear determinant in the participant neurologist spending more time on discussions with the patient actor. There were also no consistent differences in the median time spent discussing symptoms across domains with and without the use of the Floodlight technology.
| Symptom | Patient Case A | Patient Case B | ||
| With Floodlight (min) | Without Floodlight (min) | With Floodlight (min) | Without Floodlight (min) | |
| Vision changes | 01:32 | 01:18 | 00:30 | 00:10 |
| Sexual dysfunction | 00:05 | 01:15 | 00:17 | 00:55 |
| Mood changes | 02:18 | 01:26 | 01:32 | 01:50 |
| Physical and emotional pain | 01:55 | 01:15 | 01:54 | 01:12 |
| Bowel and bladder dysfunction | 00:30 | 01:10 | 01:14 | 02:47 |
| Cognitive changes | 02:16 | 02:06 | 00:27 | 00:53 |
| Fatigue | 00:40 | 01:01 | 01:08 | 01:39 |
| Median across all domains | 01:25 | 01:22 | 01:07 | 01:34 |
aInvisible symptoms, including fatigue, mood disorders, cognitive impairments, pain, bladder or bowel dysfunction, sexual dysfunction, and vision changes can be experienced by people living with MS. Some of these symptoms may not be externally evident or easy to report and quantify and can be difficult for the patient to report and articulate during a clinical encounter (eg, due to subjectivity bias, poor recall, or stigma) or for physicians to capture (eg, limitations in available measurement tools, ample reliance on patient recall on the visit day).
bThese are the symptoms that were previously reported, per the clinical documentation provided, for Patient Case A and Patient Case B.
Care Decisions and Change in Disease Management Decisions
Floodlight data were reported as a trigger for 20 of the 37 care and change in disease management decisions (including regarding disease-modifying therapies, additional medications, referrals, or investigations), always in combination with other triggers such as patient history (20/20) or clinical exam findings (9/20; ). Overall, a greater number of care decisions were taken when using the Floodlight technology (37) versus without (29), including making a referral to another HCP (9 vs 6), or adding, changing, or discontinuing a nonpharmacologic therapy (6 vs 5) or disease-modifying therapy (4 vs 2).

Overall Feedback
Overall, participant neurologists valued access to longitudinal data, particularly regarding hand function and cognition. Many participant neurologists found the aspects of the user interface challenging when using the Floodlight MS HCP Portal. They felt that user interface improvements were the most important changes to be made, and they noted the following could help improve the interface: “Separating different data points and putting them into separate charts,” “...an overall curve, visualizing the trend over time,” and “…being able to change the y axis to separate data points for easier review.”
Data Availability
For up-to-date details on Roche’s global policy on the sharing of clinical information and how to request access to related clinical study documents, please visit [].
Discussion
Principal Findings
The clinical simulations of MS teleconsultations described in this study demonstrate the ability of this research methodology to deliver high-fidelity patient cases. They also enabled direct comparison of teleconsultations conducted with and without the use of digital health technologies.
Clinical simulation offered several advantages, including the ability to present relevant, representative, paradigmatic MS cases and to efficiently generate insights on the utility and usability of new digital technology in clinical visits. This provided early real-world insights from neurologists who were naïve to the Floodlight technology and who received only brief introductory training on its use. Specifically, it provided an opportunity to move beyond the expected impact of the tool and, instead, explore the approach used by participants naïve to the digital technology (after a brief orientation) in actual clinical decision-making.
Using real patient data from the CONSONANCE study enabled access to extensive existing data without additional patient consent or data acquisition, allowing case selection based on in-depth clinical information. The richness of the CONSONANCE dataset minimized the need for synthetic data, which was limited but where used, derived from valid sources such as narrative radiology reports and clinical letters. While this approach ensured high-fidelity and clinically relevant patient scenarios, it introduced a limitation in the generalizability of the simulated cases as the patient population in the CONSONANCE study was composed entirely of people with progressive multiple sclerosis. This highlights the capability of clinical simulation to leverage rich, real-world datasets while also underscoring the importance of including a broader spectrum of patient profiles and validated synthetic cases to enhance external validity, scalability, and applicability across diverse MS phenotypes. Additional simulated patient cases can be easily generated from other clinical trials or by generating synthetic (but realistic) patient case data. Further simulation research should expand the number and diversity of patient cases, including data from varied trial populations and validated synthetic cases to better capture the heterogeneity of MS care. It is important to ensure that these cases are representative of typical patients seen in clinical practice and developed with appropriate ethical considerations to avoid the introduction of unwanted bias. Additionally, it is also essential to ensure that clinical trial data used for patient cases do not allow patient identities to be inferred and that blinding procedures are preserved.
This study explored complex MS cases, with multiple patient concerns and functional domains affected, to capture the complexity of real-world MS clinical visits and explore the impact of utilizing a digital tool (Floodlight technology). As with “real” clinical encounters with people living with MS, distilling prior history and functional testing into condensed clinical letters for use in the consultations was challenging, due to the heterogeneity of symptoms, scales, and other assessments used. For example, bladder symptoms reported by patients in the SymptoMScreen were consistently underscored in the clinicians’ EDSS Bowel/Bladder Functional System [-].
Research applications of simulation can include training methodologies for educational purposes [], such as in training for medical students or less experienced HCPs, and to enhance clinical competence []. Additionally, simulation can be used to improve primary care office–based readiness to respond to medical emergencies [] or to enhance cultural sensitivity, communication, and empathy among HCPs []. Simulation can also be used as an investigative methodology to answer research questions that otherwise could not be answered feasibly, safely, ethically, or in a timely fashion in clinical settings [], as pertains to our study.
The simulation methodology provided key insights into integrating a digital health tool into routine MS care. Initial findings pertained to how neurologists naïve to this technology began to access and interpret these data. Indeed, there is a known socialization process required for the adoption of new technology [], and this must be accounted for in the plans that developers have for the implementation of new tools. Neurologists scored the clinician-facing HCP Portal for accessing Floodlight MS data “above average” on the System Usability Scale, indicating a high level of usability. Feedback from postsimulation surveys and interviews provides a foundation for iterative improvements to the user interface. Among the available data on the portal, the “mood status” aspect of the Patient Journal (Daily Status Questions) was rated as the most useful of the data visualizations presented, which is potentially indicative of a medical need (eg, more systematic depression monitoring). Finally, as expected, incorporating Floodlight data into teleconsultation preparation required slightly more time and mental effort than consultation without such technology. Interestingly, the median time was only 1.7 minutes longer for patient case preparation, which seems relatively modest compared with the potential gains from having Floodlight data available. This also indicates that the brief presimulation training was effective and supports the tool’s scalability. With increased familiarity, time required to use the Floodlight technology during the visits would likely decrease. However, even a short amount of additional time requirement could nonetheless present a potential barrier to use. For example, when evaluating the initial launch of the patient codesigned, high-usability-scoring, point-of-care clinical dashboard MS NeuroShare, clinicians were less likely to use it on busy clinic days, despite the reported improvements in the quality of communication and patient understanding of management care plans. New solutions to alleviate clinician workloads must be evaluated against the time required to implement them, such as educating patients on the clinical applications [].
Digital health tools can capture objective data between clinical visits. This can avoid reliance on patient recall of symptoms, helping to provide clinicians with a more complete picture of symptoms and enabling discussions between HCPs and patients. The second set of exploratory insights, therefore, concerned the impact that ready access to Floodlight data might have on clinical discussion and management. Overall, while participant neurologists started discussing the primary domain of concern after 2 minutes both with or without the Floodlight technology, most participants said that it was easier to identify patients’ concerns when they had access to the Floodlight MS HCP Portal (A). Further, more care decisions were taken when using Floodlight data versus without, for example, making a referral to another HCP or amending a disease-modifying therapy.
Limitations
While simulations can represent and provide insights on the dynamics of MS in real-life clinical practice, the method is limited by the ability to account for operational and behavioral variables that occur in clinical practices, and it may, therefore, not be a fully accurate representation. While this methodology is inherently scalable, its primary limitation lies in the effort needed to recruit and train actors. By comparison, on-site studies involving real patients are even more constrained, as they can only include individuals residing near the clinical study site(s). Although the actors performed convincingly, simulations cannot fully substitute for real patient interactions. Real patients may experience factors such as anxiety about their clinical outcomes, fatigue from poor sleep, or depression, all of which can influence the dynamics of clinical encounters; however, it must be noted that this is not one of the aims of this study. The limited numbers of participants and patient cases in this study, important for this initial testing of the research methodology, naturally limit generalizability to real-life clinical consultations. Virtual simulations could enrich the recruitment of diverse neurologists by offering more convenient participation as well as the potential to recruit neurologists from a wider geographic area. This will be important in studies that seek to quantify differences in clinical decision-making. Furthermore, a potential source of bias relevant to all simulations involving observation of tasks is the Hawthorne effect, where participants modify their behavior when they know they are being observed [,]. Mitigation strategies to overcome this included informing participating neurologists in advance about the study objectives, emphasizing that the research focus was not primarily on diagnostic accuracy, and assuring them that their identities would remain confidential. To further minimize the Hawthorne effect, the observer’s camera and microphone were switched off during sessions; however, the observer remained present on the call to ensure study fidelity and provide technical support if needed. It should be acknowledged that despite these measures, such biases cannot be entirely eliminated in simulation-based research, although increasing the number of participants might reduce their impact.
Further, exploring cognitive impairment as a primary domain of concern highlighted some limitations of using simulated encounters because some participant neurologists described typically engaging patients’ family members or care partners when assessing the extent and impact of a patient’s cognitive challenges, and family members were not included in the simulated scenarios. An alternative approach in clinical simulations exploring more focused digital tools would be to curate simpler patient cases with only 1 domain of concern and a few specific invisible symptoms and to focus the investigation on more specific aspects.
Conclusions
Clinical simulation of MS teleconsultations allowed evidence generation on the potential utility of a digital health technology tool in MS care. The simulated patient cases and environment were found to be of high fidelity and enabled the direct comparison of teleconsultations conducted with and without the use of the Floodlight technology. A number of domains could be explored, including tool usability, influence of Floodlight data on clinical decisions, and impact on clinical workflows. Employing this methodology provided preliminary evidence that the Floodlight technology could support routine MS care, as it was found to be usable and useful, with minimal additional time and mental effort required on the part of neurologists previously naïve to the technology.
The findings of this study support expanded research into and development of clinical simulation methodologies for neurological research. This could expedite integrating innovative methods or interventions in health care settings, as well as in medical training. In the digital health space specifically, simulations could help to anticipate what socialization of technologies may require, to steer guidance on usage, explain their value to clinicians, and help health systems plan implementation.
Acknowledgments
We would like to thank all study participants and their families. We would also like to thank Active Image Roleplay for providing the professional actors, and Philipp D’Alfonso, Joan Piňol Fustero, and Harold Gonzalez Hernandez, who helped with the technical setup of the Floodlight MS HCP Portal simulated environment. Writing and editorial assistance for this manuscript was provided by Cynthia J Henley-Smith, PhD, and Terri Desmonds, PhD, of Nucleus Global, an Inizio company in the United Kingdom, and funded by F. Hoffmann-La Roche Ltd.
Funding
Funding for this study was provided by F. Hoffmann-La Roche Ltd.
Authors' Contributions
Conceptualization: AK, JH, LC, RB, SG
Data curation: SD
Formal analysis: SD, SH
Funding acquisition: LC
Investigation: JH, RB, SD
Methodology: JH, SD, SG
Project administration: JH, LC
Resources: JH, LC
Supervision: JH, LC
Visualization: SD, SH
Writing – review & editing: AK, IW, JH, LC, RB, SD, SG, SH
Conflicts of Interest
RB has received consultancy fees from Alexion, EMD Serono, Horizon, Janssen, and TG Therapeutics; and is funded by the NMSS Harry Weaver Award, NIH, DOD, and NSF, as well as Biogen, Eli Lilly, F. Hoffmann-La Roche Ltd, and Genentech. LC is an employee of F. Hoffmann-La Roche Ltd. IC was an employee of F. Hoffmann-La Roche Ltd when this study was conducted and is currently an employee of CS Genetics. SD was a subcontractor to Prova Health when this study was conducted and is currently an employee of Imperial College London. SG is an employee of and shareholder in Prova Health and an employee of Imperial College London. JH is an employee of and shareholder in Prova Health. SH is an employee of and shareholder in F. Hoffmann-La Roche Ltd. AK is an employee of and shareholder in F. Hoffmann-La Roche Ltd.
References
- Kappos L, Wolinsky JS, Giovannoni G, et al. Contribution of relapse-independent progression vs relapse-associated worsening to overall confirmed disability accumulation in typical relapsing multiple sclerosis in a pooled analysis of 2 randomized clinical trials. JAMA Neurol. Sep 1, 2020;77(9):1132-1140. [CrossRef] [Medline]
- Moccia M, Lanzillo R, Palladino R, et al. Cognitive impairment at diagnosis predicts 10-year multiple sclerosis progression. Mult Scler. Apr 2016;22(5):659-667. [CrossRef] [Medline]
- Filippi M, Bar-Or A, Piehl F, et al. Multiple sclerosis. Nat Rev Dis Primers. Nov 8, 2018;4(1):43. [CrossRef] [Medline]
- Thompson AJ, Baranzini SE, Geurts J, Hemmer B, Ciccarelli O. Multiple sclerosis. Lancet. Apr 21, 2018;391(10130):1622-1636. [CrossRef] [Medline]
- van der Walt A, Butzkueven H, Shin RK, et al. Developing a digital solution for remote assessment in multiple sclerosis: from concept to software as a medical device. Brain Sci. Sep 21, 2021;11(9):1247. [CrossRef] [Medline]
- Kurtzke JF. Rating neurologic impairment in multiple sclerosis: an expanded disability status scale (EDSS). Neurology (ECronicon). Nov 1983;33(11):1444-1452. [CrossRef] [Medline]
- Fouad AM, Abdel Naseer M, Farghaly M, Hegazy MI. New algorithmic approach for easier and faster extended disability status scale calculation. Mult Scler J Exp Transl Clin. 2023;9(1):20552173231155055. [CrossRef] [Medline]
- Hatcher-Martin JM, Busis NA, Cohen BH, et al. American Academy of Neurology telehealth position statement. Neurology (ECronicon). Aug 17, 2021;97(7):334-339. [CrossRef] [Medline]
- Bove R, Poole S, Cuneo R, et al. Remote observational research for multiple sclerosis: a natural experiment. Neurol Neuroimmunol Neuroinflamm. Mar 2023;10(2):e200070. [CrossRef] [Medline]
- McGinley M, Carlson JJ, Reihm J, et al. VIRTUAL versus usual in-office care for multiple sclerosis: the VIRTUAL-MS multi-site randomized clinical trial study protocol. Contemp Clin Trials. Jul 2024;142:107544. [CrossRef] [Medline]
- Marziniak M, Brichetto G, Feys P, Meyding-Lamadé U, Vernon K, Meuth SG. The use of digital and remote communication technologies as a tool for multiple sclerosis management: narrative review. JMIR Rehabil Assist Technol. Apr 24, 2018;5(1):e5. [CrossRef] [Medline]
- Giovannoni G, Butzkueven H, Dhib-Jalbut S, et al. Brain health: time matters in multiple sclerosis. Mult Scler Relat Disord. Sep 2016;9 Suppl 1:S5-S48. [CrossRef] [Medline]
- Giovannoni G. Disease-modifying treatments for early and advanced multiple sclerosis: a new treatment paradigm. Curr Opin Neurol. Jun 2018;31(3):233-243. [CrossRef] [Medline]
- University of California, San Francisco MS-EPIC Team, Cree BAC, Hollenbach JA, et al. Silent progression in disease activity-free relapsing multiple sclerosis. Ann Neurol. May 2019;85(5):653-666. [CrossRef] [Medline]
- Bove R, White CC, Giovannoni G, et al. Evaluating more naturalistic outcome measures: a 1-year smartphone study in multiple sclerosis. Neurol Neuroimmunol Neuroinflamm. Dec 2015;2(6):e162. [CrossRef] [Medline]
- Boukhvalova AK, Kowalczyk E, Harris T, et al. Identifying and quantifying neurological disability via smartphone. Front Neurol. 2018;9:740. [CrossRef] [Medline]
- Quinn CC, Clough SS, Minor JM, Lender D, Okafor MC, Gruber-Baldini A. WellDoc mobile diabetes management randomized controlled trial: change in clinical and behavioral outcomes and patient and physician satisfaction. Diabetes Technol Ther. Jun 2008;10(3):160-168. [CrossRef] [Medline]
- Bauer AM, Rue T, Munson SA, et al. Patient-oriented health technologies: patients’ perspectives and use. J Mob Technol Med. Aug 2017;6(2):1-10. [CrossRef] [Medline]
- Colligan E, Metzler A, Tiryaki E. Shared decision-making in multiple sclerosis. Mult Scler. Feb 2017;23(2):185-190. [CrossRef] [Medline]
- Guo C, Ashrafian H, Ghafur S, Fontana G, Gardner C, Prime M. Challenges for the evaluation of digital health solutions—a call for innovative evidence generation approaches. NPJ Digit Med. 2020;3:110. [CrossRef] [Medline]
- Al-Elq AH. Simulation-based medical teaching and learning. J Family Community Med. Jan 2010;17(1):35-40. [CrossRef] [Medline]
- Cheng A, Auerbach M, Hunt EA, et al. Designing and conducting simulation-based research. Pediatrics. Jun 2014;133(6):1091-1101. [CrossRef] [Medline]
- Bove R, Bevan C, Crabtree E, et al. Toward a low-cost, in-home, telemedicine-enabled assessment of disability in multiple sclerosis. Mult Scler. Oct 2019;25(11):1526-1534. [CrossRef] [Medline]
- Graves JS, Elantkowski M, Zhang YP, et al. Assessment of upper extremity function in multiple sclerosis: feasibility of a digital pinching test. JMIR Form Res. Oct 2, 2023;7:e46521. [CrossRef] [Medline]
- Dondelinger F, Thomann AE, Volkova-Volkmar E, Lipsmeier F, Sprengel J, McGinley M, et al. A digital remote assessment for measuring impairment in information processing speed in people with MS. Mult Scler. 2021;27(Suppl):P303. [CrossRef]
- Creagh AP, Simillion C, Scotland A, et al. Smartphone-based remote assessment of upper extremity function for multiple sclerosis using the Draw a Shape Test. Physiol Meas. Jun 19, 2020;41(5):054002. [CrossRef] [Medline]
- Cheng WY, Bourke AK, Lipsmeier F, et al. U-turn speed is a valid and reliable smartphone-based measure of multiple sclerosis-related gait and balance impairment. Gait Posture. Feb 2021;84:120-126. [CrossRef] [Medline]
- Bourke AK, Scotland A, Lipsmeier F, Gossens C, Lindemann M. Gait characteristics harvested during a smartphone-based self-administered 2-minute walk test in people with multiple sclerosis: test-retest reliability and minimum detectable change. Sensors (Basel). Oct 19, 2020;20(20):5906. [CrossRef] [Medline]
- Creagh AP, Simillion C, Bourke AK, et al. Smartphone- and smartwatch-based remote characterisation of ambulation in multiple sclerosis during the two-minute walk test. IEEE J Biomed Health Inform. Mar 2021;25(3):838-849. [CrossRef] [Medline]
- Midaglia L, Mulero P, Montalban X, et al. Adherence and satisfaction of smartphone- and smartwatch-based remote active testing and passive monitoring in people with multiple sclerosis: nonrandomized interventional feasibility study. J Med Internet Res. Aug 30, 2019;21(8):e14863. [CrossRef] [Medline]
- Lakin L, Davis BE, Binns CC, Currie KM, Rensel MR. Comprehensive approach to management of multiple sclerosis: addressing invisible symptoms—a narrative review. Neurol Ther. Jun 2021;10(1):75-98. [CrossRef] [Medline]
- Hyzy M, Bond R, Mulvenna M, et al. System usability scale benchmarking for digital health apps: meta-analysis. JMIR Mhealth Uhealth. Aug 18, 2022;10(8):e37290. [CrossRef] [Medline]
- Paas F, Ayres P, Pachman M. Assessment of cognitive load in multimedia learning: theory, methods and applications. In: Robinson DH, Schraw G, editors. Recent Innovations in Educational Technology That Facilitate Student Learning. Information Age Publishing Inc; 2008:11-36. URL: https://researchers.mq.edu.au/en/publications/assessment-of-cognitive-load-in-multimedia-learning-theory-method/ [Accessed 2025-12-02] ISBN: 9781593116521
- Luck I. What is a Good Net Promoter Score (NPS)? 4 Approaches. Customer Gauge; 2023. URL: https://customergauge.com/what-is-a-good-net-promoter-score [Accessed 2024-07-23]
- Our commitment to transparency of clinical study information. Roche. URL: https://go.roche.com/data_sharing [Accessed 2025-11-19]
- Meyer-Moock S, Feng YS, Maeurer M, Dippel FW, Kohlmann T. Systematic literature review and validity evaluation of the Expanded Disability Status Scale (EDSS) and the Multiple Sclerosis Functional Composite (MSFC) in patients with multiple sclerosis. BMC Neurol. Mar 25, 2014;14:58. [CrossRef] [Medline]
- Strober L, DeLuca J, Benedict RH, et al. Symbol Digit Modalities Test: a valid clinical trial endpoint for measuring cognition in multiple sclerosis. Mult Scler. Nov 2019;25(13):1781-1790. [CrossRef] [Medline]
- Kaplan TB, Gopal A, Block VJ, et al. Challenges to longitudinal characterization of lower urinary tract dysfunction in multiple sclerosis. Mult Scler Relat Disord. Jun 2022;62:103793. [CrossRef] [Medline]
- Monachino A, Caraher C, Ginsberg J, Bailey C, White E. Medical emergencies in the primary care setting: an evidence based practice approach using simulation to improve readiness. J Pediatr Nurs. 2019;49:72-78. [CrossRef] [Medline]
- Turkelson C, Cooper D, Creech C, Schellenberg K, Behnke L. Using simulation to enhance nurse practitioner students cultural sensitivity, communication, and empathy with vulnerable populations. Clin Simul Nurs. Jul 2021;56:108-116. [CrossRef]
- Bove R, Bruce CA, Lunders CK, et al. Electronic health record technology designed for the clinical encounter: MS NeuroShare. Neurol Clin Pract. Aug 2021;11(4):318-326. [CrossRef] [Medline]
- Gould D, Lindström H, Purssell E, Wigglesworth N. Electronic hand hygiene monitoring: accuracy, impact on the Hawthorne effect and efficiency. J Infect Prev. Jul 2020;21(4):136-143. [CrossRef] [Medline]
- Sujatha BK, Reddy MT, Pathak P. Camouflage in research—the Hawthorne effect. Int J Dev Res. 2019;9:15710. URL: https://journalijdr.com/camouflage-research-%E2%80%93-hawthorne-effect [Accessed 2025-12-02]
Abbreviations
| EDSS: Expanded Disability Status Scale |
| HCP: health care professional |
| MS: multiple sclerosis |
| PRO: patient-reported outcome |
| SDMT: Symbol Digit Modalities Test |
Edited by Qingyu Chen; submitted 22.Nov.2024; peer-reviewed by Jeroen Van Schependom, Kola Adegoke; final revised version received 05.Aug.2025; accepted 27.Oct.2025; published 11.Dec.2025.
Copyright© Riley Bove, Luca Capezzuto, Imogen West, Simon Dryden, Saira Ghafur, Jack Halligan, Stanislas Hubeaux, Agne Kazlauskaite. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 11.Dec.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.

