Published on in Vol 13 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/35987, first published .
Detection of Polyphonic Alarm Sounds From Medical Devices Using Frequency-Enhanced Deep Learning: Simulation Study

Detection of Polyphonic Alarm Sounds From Medical Devices Using Frequency-Enhanced Deep Learning: Simulation Study

Detection of Polyphonic Alarm Sounds From Medical Devices Using Frequency-Enhanced Deep Learning: Simulation Study

1Graduate School of Informatics, Kyoto University, 54 Kawara-cho, Shogoin, Sakyo-ku, Kyoto, Japan

2Graduate School of Medicine, Kyoto University, Kyoto, Japan

3Kyoto University Hospital, Kyoto, Japan

4Graduate School of Information Science, University of Hyogo, Kobe, Japan

5Department of Information Science, Kindai University, Osaka, Japan

Corresponding Author:

Kazumasa Kishimoto, PhD


Background: Although an increasing number of bedside medical devices are equipped with wireless connections for reliable notifications, many nonnetworked devices remain effective at detecting abnormal patient conditions and alerting medical staff through auditory alarms. Staff members, however, can miss these notifications, especially when in distant areas or other private rooms. In contrast, the signal-to-noise ratio of alarm systems for medical devices in the neonatal intensive care unit is 0 dB or higher. A feasible system for automatic sound identification with high accuracy is needed to prevent alarm sounds from being missed by the staff.

Objective: The purpose of this study was to design a method for classifying multiple alarm sounds collected with a monaural microphone in a noisy environment.

Methods: Features of 7 alarm sounds were extracted using a mel filter bank and incorporated into a classifier using convolutional and recurrent neural networks. To estimate its clinical usefulness, the classifier was evaluated with mixtures of up to 7 alarm sounds and hospital ward noise.

Results: The proposed convolutional recurrent neural network model was evaluated using a simulation dataset of 7 alarm sounds mixed with hospital ward noise. At a signal-to-noise ratio of 0 dB, the best-performing model (convolutional neural network 3+bidirectional gate recurrent unit) achieved an event-based F1-score of 0.967, with a precision of 0.944 and a recall of 0.991. When the venous foot pump class was excluded, the classwise recall of the classifier ranged from 0.990 to 1.000.

Conclusions: The proposed classifier was found to be highly accurate in detecting alarm sounds. Although the performance of the proposed classifier in a clinical environment can be improved, the classifier could be incorporated into an alarm sound detection system. The classifier, combined with network connectivity, could improve the notification of abnormal status detected by unconnected medical devices.

JMIR Med Inform 2025;13:e35987

doi:10.2196/35987

Keywords



Background

While an increasing number of bedside medical devices, such as syringe pumps, have wireless connections that enable reliable data transmission to hospital information systems, many nonnetworked devices are still used in general hospital wards. Although dongles connected to the external output terminal of these devices may allow wireless connections [1,2], most devices are not equipped with external output terminals. Instead, these devices use auditory alert signals (alarm sounds) to notify medical staff of abnormal conditions. Medical staff members may not hear these alarms, especially when they are in distant areas or other private rooms. Between 2010 and 2015, the Japan Council for Quality Health Care reported 173 accidents and other incidents, including 23 cases of unnoticed alarms [3]. The report included the following comments about environmental factors:

There are many blind spots due to the facility’s structure.
When I entered a patient’s room, I could not hear alarm sounds from another room.
The alarm sound did not reach the staff because the room was far from the nurse station.
The staff could not hear the alarm in the farthest private room in the ICU.
The recording room was structured so that the alarm sound could not be heard.

These findings indicate the need for reliable alarm notification to ensure patient safety. Alarm sounds emitted by medical devices are regulated by the International Organization for Standardization and the International Electrotechnical Commission (ISO/IEC) 60601-1-8. This standard specifies the melodies and lengths of alarm sounds to reduce the risk of misunderstanding, confusion, and omission of alarm sounds from various medical devices, even when these sounds overlap and reverberate. This standard prescribes that the sounds should be organized based on the priority of corresponding abnormal situations, with alarm sounds for different situations varying in melody and length. Thus, sound event detection (SED) is expected to identify every kind of abnormal situation detected by medical devices [4-8]. The standard also defines visual alarm signals, but monitoring the signals of multiple devices with cameras is not feasible without blind spots.

As SED can be implemented using a single (monaural) microphone, it was selected as the approach to detect alarm status. Clinical application of SED requires robustness against noise because environmental noise in hospital wards is generally substantial. SEDs with deep learning have been found to be sufficiently robust against noise [9,10].

This study proposes a deep learning–based method for classifying patient abnormalities detected by medical devices with polyphonic alarm sounds collected with a monaural microphone. The ability of the classifier to identify abnormal states of these devices was evaluated using simulation datasets of their alarm sounds superimposed on hospital ward noise (HWN). Therefore, the objective of this study was to design and evaluate a convolutional recurrent neural network (CRNN) for accurately detecting and classifying multiple, overlapping alarm sounds from medical devices in a simulated noisy hospital environment. We hypothesized that a hybrid CRNN model could achieve high performance suitable for clinical application by effectively capturing both the spectral and temporal features of the alarm sounds.

Related Works

Accurate transmission of alarms from unconnected medical devices requires precise recognition of visual and auditory alarms. Several studies have reported high accuracy in detecting simultaneous alarm sounds mixed with a substantial level of environmental noise [4-8]. For SED, deep learning is more robust against noise than conventional methods [9,10]. On the basis of these studies, an edge device was placed at the patient’s bedside to monitor abnormal conditions detected by multiple medical devices with individual alarm sounds. In our previous study, the classifier had F1-scores of 0.727 at signal-to-noise ratios (SNRs) of 0 dB and applied a convolutional neural network (CNN) [11]. To our knowledge, no previous study has used a deep learning–based recurrent neural network (RNN) to detect polyphonic alarm sounds emitted by medical devices. Recent advances in attention-based architectures, such as the audio spectrogram transformer and conformer, have demonstrated strong performance in general SED tasks [12,13]. These models use self-attention mechanisms to capture long-range dependencies, potentially enhancing robustness in noisy environments. Exploring such transformer-based approaches for clinical alarm sound detection remains an important direction for future research.


Overview

The first approach to SED combines a Gaussian mixture model with a hidden Markov model, using features such as the mel-frequency cepstral coefficient from traditional methods of speech recognition [14,15].

Other approaches include the separation of sound sources by matching them using a template extracted from the input sound. This can be achieved by sound source separation techniques, such as nonnegative matrix factorization. Nonnegative matrix factorization monitors a single signal to create a basis matrix and identifies the separated sounds [16-18].

Recent approaches based on neural networks have significantly improved the performance of SED. One approach consists of SED of real-life sounds with feedforward neural networks based on a multilayer perceptron trained in a spectrum of mixed sounds [6]. An RNN with the ability to remember past states can process sequential information of the acoustic signal. RNNs with bidirectional long short-term memory have achieved excellent results in complex audio detection such as speech recognition and polyphonic piano note transcription [19-21].

Furthermore, CNNs commonly used in image recognition can robustly predict sounds with its filter shifted by both time and frequency axes [22]. However, long-term prediction remains difficult due to the limited width of the time window [9]. Therefore, although alarm sounds consist of relatively simple tones, it is necessary to predict not only the frequency axis but also the time axis to inform the priority with a pattern. Application of an RNN to polyphonic SED enabled long-term prediction by integrating the information over the time window. This study combines the strengths of both CNN and RNN to benefit from both approaches. A similar approach has shown excellent performance in automatic speech recognition [23,24].

Therefore, this approach was expected to achieve sufficient performance for potential clinical application in alarm sound recognition.

Deep Alarm Sound Detection

Experiment Overview

The proposed polyphonic alarm sound detection consisted of feature extraction, classification algorithms, and model training (Figure 1). A mel filter bank (MFB) was used to extract the features of alarm sounds, and CNN and RNN were applied to the classifier. As deep learning requires substantial training data, a large amount of acoustic data were recorded in a quiet room. The recorded data were mixed with pseudo noise before being used for training. This approach aimed to maximize the generalizability of the classifier for expected use in a noisy environment. The feature extraction step extracted data to be input to the classifier from the collected alarm sounds. The classification algorithms were designed to use deep learning models for classifiers. The model training augmented the data to create a robust model.

Figure 1. Overview of the experiment. SNR: signal-to-noise ratio.
Feature Extraction

The features of polyphonic alarm sounds of bedside medical devices were extracted. Many acoustic classifiers use log mel spectrogram multiplied spectra and MFB based on the characteristics underlying human frequency perception [25]. The acoustic data were transformed to a power spectrogram with the short-time Fourier transform of the Hamming window, which had a window size of 1024 and a hop length of 512. The inner product on the power spectrogram was calculated with MFB, and its logarithms in 40 dimensions were calculated.

Classification Algorithms

The CRNN, a combination of CNN and RNN, was used for the classification model (Figure 2) [26]. This system used 256×40 pixels of the log mel spectrogram as a feature. A CNN block in the proposed model consisted of a convolutional layer with 128 filters, batch normalization, the activation function of the rectified linear unit, and a dropout of 50% [27]. Using the size reduction method, the max pooling layer was applied to the frequency axis. In image recognition, replacing the pooling layer with a CNN stride 2 has been reported to improve its performance by reducing the calculation cost [28]. In this study, the size of the stride to the frequency axis was reduced to 2, and the pooling layer was excluded because the classification target was 1 frame of the sound.

Figure 2. Architecture of the proposed convolutional recurrent neural network. The architecture consists of three main components: (1) convolutional neural network (CNN) block (convolutional layers with batch normalization, rectified linear unit [ReLU] activation, max pooling, and dropout), (2) recurrent neural network (RNN) block (bidirectional gate recurrent unit [BiGRU] or bidirectional simple recurrent neural network [BiSRNN] layers), and (3) time-distributed dense with sigmoid activation for classification.

The RNN block consisted of a bidirectional simple RNN (BiSRNN) or bidirectional gated recurrent unit (BiGRU). The input was set to 32 units, the activation function to tanh, the recurrent activation function to the hard sigmoid, and the dropout of each layer to 50%.

Finally, the activation function of the fully connected layer was set to sigmoid, the optimization function used was Adam [29], and the loss function was set to binary cross-entropy. Each of the 7 sound event classes had output values in the range of [0, 1] [5].

Table 1 shows the details of the proposed model. CNN3+BiSRNN was used as the baseline model. The environment was built with Python programming language (version 3.7.11; Python Software Foundation), using Keras 2.3.1 for the deep learning library (TensorFlow 2.0.0 for the back end) and Librosa 0.8.1 for the acoustic analysis module. The experimental code is available online[30].

Table 1. Details of the proposed model.
Model nameCNNa blockRNNc block
LayerNumber of layersStrideMax poolingbLayerNumber of layers
CNN3+BiSRNNd (baseline)Conv2De31×15, 2, 2BiSRNN2
CNN3+BiGRUfConv2D31×15, 2, 2BiGRU2
CNN4+BiGRUConv2D41×12, 2, 2, 2BiGRU2
ALL-CNN4+BiGRUgConv2D41×2hBiGRU2

aCNN: convolutional neural network.

bThe figures denote frequency axis. Time axis=1 (ie,1×5=5).

cRNN: recurrent neural network.

dBiSRNN: bidirectional simple recurrent neural network.

eConv2D: two-dimensional convolution.

fBiGRU: bidirectional gated recurrent unit.

gALL-CNN: customizing stride without max pooling.

hNot applicable.

Model Training

Data augmentation was applied to the training dataset to prevent overfitting of the classifier and to provide robust performance in simulation experiments [31]. The 7 alarm sounds were superimposed on white noise at SNRs ranging from 30 to 0 dB in 1 dB steps. In addition, SpecAugment was applied to the generator, with 1 random mask added for each frequency and time axis, followed by performance of 5 steps per minibatch [32].

These steps produced a trained model using 5-fold cross-validation over 150 training epochs and confirmed that the learning curve showed no signs of overfitting.

Evaluation

Data Collection

The medical devices selected included those frequently used for ventilator-equipped patients who require many medical devices in the general ward of the hospital. Alarm sounds to be identified included pulse sounds from a syringe pump (SP), enteral feeding pump (ENP), and venous foot pump (VFP) device as well as burst sounds from an infusion pump (IP), chest drainage (CD), patient monitor (PM), and the mechanical ventilator. The alarm sounds of each device were recorded using a monaural microphone placed at the head of a bed in a quiet private room in the hospital. The distance between the sound source and the microphone was the same as in a typical bedside setting (Figure 3). The sound pressure level was recorded simultaneously. The recording has a different number of active sound events superimposed on each frame. Therefore, the frame has various polyphony levels. The distribution of polyphony levels when recording the alarm sounds for the 7 devices is shown in Table 2. Audacity was used for labeling, extracting fundamental frequencies, and performing spectral analysis for annotation. Table 3 shows the detailed characteristics of the recorded alarm sounds.

Figure 3. Recording environment. Numbers in parentheses represent the distance between the microphone and each device.
Table 2. Level of each sound relative to the total level of polyphonic sound.
Polyphony levelData mount, n (%)
00 (0)
13655 (5.7)
215,011 (23.6)
322,547 (35.4)
415,911 (25)
55280 (8.3)
61230 (1.9)
764 (0.1)
Table 3. Detailed characteristics of the sounds of each alarm.
Source deviceModel (manufacturer)Peak frequencies (Hz)aSignal duration (seconds)Silence duration (seconds)Overall duration (seconds)
Infusion pumpOT-818G (JMS)[(856,856,856)-(856,856)] 2 times3.342.996.33
Syringe pumpTE-351 (Terumo)40010.260.200.46
Enteral feeding pumpAPPLIX Smart (Fresenius Kabi)40970.800.801.60
Venous foot pumpSCD700 (Coviden)21080.201.301.50
Chest drainageTHOPAZ (Medela)26321.078.9510.02
Patient monitorPVM-4761 (NihonKoden)[(783,994,1181)‐(1181,1569)] 2 timesb2.934.097.02
Mechanical ventilatorC1 (Hamilton)491 828 662c1.104.986.08

aFundamental frequencies analyzed by Audacity. Harmonics are excluded.

bHigh-priority alarm sound.

cMiddle-priority alarm sound.

Simulation Dataset

The robustness of the classifier was evaluated using a simulation dataset of alarm sounds added to HWN (which comprised conversations, footsteps, closet opening and closing sounds, intraoral suction, and ventilator exhalation sounds but not alarm sounds from other medical devices) at different SNRs (Figure 1). For the simulation dataset, the alarm sounds were recorded separately from those in the Data Collection section, using the same recording protocol (quiet private room, identical microphone placement, and device settings). During recording, all 7 devices were set to repeatedly emit alarms, sometimes overlapping due to differences in their alarm durations. Multiple alarms sometimes sounded simultaneously because of differences in the duration times of each alarm sound. Table 4 shows the maximum sound pressure of each sound source. In contrast to the training dataset—where alarm sounds were superimposed on white noise at SNRs ranging from 30 to 0 dB in 1 dB steps—the simulation dataset used for evaluation was created by superimposing the recorded HWN on alarm sounds at 4 SNR settings (30, 20, 10, and 0 dB) to reproduce realistic clinical environments. As shown in this figure, the VFP alarm exhibited the lowest sound pressure level among the devices, whereas other devices, such as SP and ENP, had relatively higher levels.

Table 4. Maximum sound pressure levels of each of the devices in the recording simulation dataset.
Source deviceSound level (dB)
Infusion pump71.9
Syringe pump77.7
Enteral feeding pump64.4
Venous toot pump61.2
Chest drainage73.7
Patient monitor62.1
Mechanical ventilator79.6
Hospital ward noise63.4
Performance Metrics

Model performance was evaluated using 5-fold cross-validation. No formal statistical significance tests were conducted, as the primary objective was descriptive benchmarking rather than hypothesis testing. Evaluation was performed using the sed_eval module [7,33]. The classifier outputs each predicted value of the 7 devices. The predicted values were dichotomized based on a cutoff value of 0.5. The onset and offset times input into the sed_eval module were calculated from the change points of the output results. Segment-based metrics are an index that determines whether the reference and the output match for each set time resolution (second). Therefore, the time resolution was set to 0.1 seconds, half of the shortest alarm duration time among the 7 devices. The event-based metrics index evaluated the timing of onset and offset from the set collar (seconds). The collar was set to 2.0 seconds based on the response of the notification system. Event-based metrics evaluation considers the actual operation and is more stringent than segment-based evaluation. Comparisons were evaluated using overall metrics (microaverage) and class-wise metrics.

For clinical applicability, inaccurate information is unacceptable, as it puts patients at risk. Therefore, the requirement for clinical application was set at an F1-score value of 0.900 or higher for event-based overall metrics and an F value of 0.950 or higher for class-wise recall metrics.

Finally, the predicted results were visualized using the sed_vis module, and the spectrogram and classification results were examined [34].

Ethical Considerations

This study did not involve human participants or animal experiments. The recordings contained no speech, personal identifiers, or patient-related information. Therefore, ethics approval was not required in accordance with institutional and international research ethics policies.


Table 5 summarizes the overall performance metrics at an SNR of 0 dB. The 679-second simulation dataset consisted of 63,649 frames, of which 63,488 frames were evaluated using 256-frame input windows. Among the segment-based metrics, CNN4+BiGRU achieved the highest F1-score, followed by CNN3+BiGRU. Conversely, for event-based metrics, CNN3+BiGRU outperformed all other models, with CNN4+BiGRU ranking second. Regarding recall, event-based metrics showed values of 0.950 or higher for all BiGRU-based models, whereas segment-based metrics yielded recall values below 0.950 for all models.

Table 5. Overall metrics (microaverage) at a signal-to-noise ratio of 0 dB across 5-fold cross-validation. The numbers in italics represent the optimal value for the proposed models.
Segment-based metrics, mean (SD)Event-based metrics, mean (SD)
F1-scorePrecisionRecallF1-scorePrecisionRecall
CNN3a+BiSRNNb (baseline)0.832 (0.007)0.822 (0.009)0.841 (0.011)0.921 (0.003)0.874 (0.009)0.973 (0.008)
CNN3+BiGRUc0.868 (0.008)0.845 (0.006)0.894 (0.012)0.967 (0.011)0.944 (0.014)0.991 (0.011)
CNN4+BiGRU0.873 (0.004)0.856 (0.003)0.890 (0.008)0.965 (0.008)0.942 (0.009)0.989 (0.008)
ALL-CNN4+BiGRU0.867 (0.010)0.839 (0.010)0.896 (0.011)0.948 (0.021)0.915 (0.032)0.983 (0.008)

aCNN: convolutional neural network.

bBiSRNN: bidirectional simple recurrent neural network.

cBiGRU: bidirectional gated recurrent unit.

Table 6 shows the event-based class-wise metrics at an SNR of 0 dB. Only event-based metrics are reported here, as they reflect the temporal accuracy of onset and offset detection, which is essential for clinical alarm management, whereas class-wise segment-based metrics are less indicative of operational performance. Class-wise segment-based results are provided in Multimedia Appendix 1 [30] for reference. Event-based evaluation showed that CNN3+BiGRU and CNN4+BiGRU had a recall value of 0.990 or higher for all devices but VFP. In contrast, the ENP, PM, and ventilator had F1-scores of 0.900 or less due to their low precision. The reference standard was the correctly annotated label from the event roll that visualized the classification results, while the output was the model’s identified results. In the absence of sound, the device was detected visually (Figure 4).

Table 6. Event-based class-wise metrics at a signal-to-noise ratio of 0 dB across 5-fold cross-validation. The numbers in italics represent the optimal value for each class.
Metrics and classF1-score, mean (SD)Precision, mean (SD)Recall, mean (SD)
CNN3a+BiSRNNb
IPc0.945 (0.016)0.900 (0.028)0.995 (0.001)
SPd0.994 (0.004)0.990 (0.008)0.998 (0.000)
ENPe0.903 (0.011)0.831 (0.016)0.990 (0.007)
VFPf0.829 (0.013)0.828 (0.036)0.834 (0.050)
CDg0.815 (0.023)0.688 (0.034)1.000 (0.000)
PMh0.854 (0.020)0.749 (0.031)0.993 (0.003)
Venti0.752 (0.058)0.606 (0.078)1.000 (0.000)
CNN3+BiGRUj
IP0.982 (0.008)0.967 (0.015)0.997 (0.001)
SP0.998 (0.000)1.000 (0.000)0.997 (0.001)
ENP0.969 (0.016)0.939 (0.029)1.000 (0.000)
VFP0.939 (0.037)0.931 (0.030)0.951 (0.074)
CD0.904 (0.031)0.826 (0.052)1.000 (0.000)
PM0.924 (0.025)0.862 (0.043)0.997 (0.001)
Vent0.834 (0.034)0.717 (0.050)1.000 (0.000)
CNN4+BiGRU
IP0.986 (0.005)0.977 (0.009)0.995 (0.002)
SP0.998 (0.000)1.000 (0.000)0.997 (0.001)
ENP0.939 (0.033)0.887 (0.058)1.000 (0.000)
VFP0.943 (0.022)0.947 (0.027)0.942 (0.054)
CD0.870 (0.030)0.770 (0.048)1.000 (0.000)
PM0.954 (0.014)0.916 (0.025)0.996 (0.002)
Vent0.797 (0.054)0.666 (0.080)1.000 (0.000)
ALL-CNN4+BiGRU
IP0.942 (0.030)0.896 (0.052)0.995 (0.003)
SP0.998 (0.000)1.000 (0.000)0.996 (0.001)
ENP0.908 (0.056)0.836 (0.090)1.000 (0.000)
VFP0.921 (0.047)0.943 (0.046)0.901 (0.057)
CD0.949 (0.018)0.903 (0.033)1.000 (0.000)
PM0.904 (0.020)0.829 (0.032)0.995 (0.003)
Vent0.805 (0.051)0.677 (0.074)1.000 (0.000)

aCNN: convolutional neural network.

bBiSRNN: bidirectional simple recurrent neural network.

cIP: infusion pump.

dSP: syringe pump.

eENP: enteral feeding pump.

fVFP: venous foot pump.

gCD: chest drainage.

hPM: patient monitor.

iVent: mechanical ventilator.

jBiGRU: bidirectional gated recurrent unit.

Figure 4. The reference and proposed model outputs of a convolutional neural network 4+bidirectional gated recurrent unit at a signal-to-noise ratio of 0 dB: (A) spectrogram of the simulation dataset and (B) event roll of the reference and proposed model outputs for the 7 devices. CD: chest drainage; ENP: enteral feeding pump; IP: infusion pump; PM: patient monitor; SP: syringe pump; VFP: venous foot pump; VENT: mechanical ventilator.

Principal Results

The proposed classifier was found to successfully detect the status of nearby medical devices. Although video-based alarm detection was also considered, it was not feasible to detect the monitors with cameras without the introduction of a blind spot. Therefore, SED was considered a more feasible approach than video recognition.

This study describes the construction of a classifier using a large amount of artificial noise data. This classifier was used to evaluate polyphonic alarm sounds among a simulation dataset of HWN. Because the classification target is a sine wave, we expected that the target could be achieved with simple neural networks. However, BiSRNN did not achieve the target performance, requiring the application of BiGRU. Both the frequency and time axes required advanced processing to recognize alarm sounds. In addition, a convolutional layer with 4 layers outperformed one with 3 layers. The F1-score of the classifier using BiGRU was 0.900 or higher, which was robust in the detection of polyphonic alarm sounds. Only VFP in each proposed classifier was unclear in the spectrogram due to VFP having the lowest sound pressure level of the 7 devices, thus resulting in low recall. In addition, masking of an impact sound over the entire frequency axis of the spectrogram would result in lower precision and overdetection.

Clinical Application as a Notification System

The notification system should alert the hospital information system without missing any situations, whether alarms are sounding on a single device or on multiple devices simultaneously. Therefore, recall of class-wise metrics was considered the most important in evaluating clinical applications. In particular, recall for the ventilator, the most essential support device of the 7 evaluated, was 1.000.

CNN4+BiGRU had the highest F1-score in the event-based overall metrics, and the recall values of ENP, CD, and the ventilator were 1.000 each. Only VFP showed a recall value below 0.950, primarily due to its lower sound pressure level. In some cases, detection occurred slightly earlier than the actual onset, which reduced the measured recall. However, the ventilator demonstrated low precision with frequent overdetections, suggesting that it could not be evaluated clinically because it could cause alarm fatigue [35]. In contrast, because the F1-scores of ENP, VFP, and CD without external output were 0.900 or higher, the system was likely feasible to notify staff of the alarm status of medical devices that could not be connected to the hospital information system. Therefore, the proposed system demonstrated feasibility as an alarm sound detection system and can be further refined for clinical use.

Integration Into Hospital Networks

Data Privacy Implications

The system processes only nonspeech alarm signals, ensuring that no patient-identifiable audio is stored or transmitted. When integrated into hospital networks, alarm classifications should be anonymized at the edge, transmitting only classification results in compliance with privacy regulations such as the General Data Protection Regulation and Japan’s Act on the Protection of Personal Information.

Real-Time Processing Feasibility

Our classifier was designed for lightweight deployment, with feature extraction and model inference feasible on edge devices. The processing latency is on the order of milliseconds per input window, enabling real-time alarm monitoring without delaying clinical response.

Regulatory Concerns

Integration of an alarm detection system into hospital infrastructure would require compliance with medical device standards, including ISO/IEC 60601-1-8 for alarm systems. Depending on the jurisdiction, such a system may be classified as a medical device software, requiring regulatory approval. Early consultation with regulatory authorities is recommended to ensure compliance and patient safety.

Limitations

This study had several limitations. First, the study evaluated only a limited range of medical devices from a single hospital, potentially limiting generalizability. Second, differences in hospital architecture (eg, room layouts, wall materials, and ambient noise) may affect sound propagation and detection. Third, despite ISO/IEC 60601-1-8 regulations, variations in alarm sounds across manufacturers and models were not assessed. Finally, large-scale live recording and annotation remain impractical; future work should use synthetic datasets and validate performance across diverse environments and device types.

Comparison With Prior Work

Several reports have examined the classification of alarm sounds using deep learning. Evaluation of single sounds of horns and bicycle bells found that the 5-layer deep neural network that applied an integrated judgment process had F1-scores of 0.99 or higher [4]. Because horns and similar devices do not create sine waves of digital sounds, they cannot be evaluated in the same way as alarm sounds of medical devices. A study of alarm sounds of medical devices in a neonatal intensive care unit found that most of the alarms had SNRs of 0 dB or higher [8]. That study, however, did not consider a classifier for polyphonic alarm sounds. A classifier using CRNN was found to be effective for polyphonic acoustic sounds [26,36].

Conclusions

Missed medical device alarms can lead to serious adverse events. To mitigate such risks, we developed a deep learning–based classifier for detecting polyphonic alarm sounds in hospital environments.

Alarm sounds emitted by medical devices are regulated by ISO/IEC 60601-1-8. Because this standard defines different tones and patterns for each device and priority, SED is expected to identify the device and priority successfully. Thus, we considered SED appropriate to determine alarm status. Automatic identification of alarm sounds in hospital rooms would facilitate safer medical care.

In the simulation experiment, the polyphonic alarm sound classifier showed excellent performance, with an F1-score of 0.945 at an SNR of 0 dB. The proposed classifier demonstrated feasibility for clinical alarm sound detection and can be further optimized. When combined with network connectivity, this classifier could improve the notification of abnormal patient status detected by medical devices without requiring each device to be individually connected.

Acknowledgments

This work was partially supported by the Japan Society for the Promotion of Science KAKENHI program (JP24K21125).

Conflicts of Interest

None declared.

Multimedia Appendix 1

Segment-based class-wise metrics at a signal-to-noise ratio of 0 dB across 5-fold cross-validation.

DOCX File, 20 KB

  1. Yoshioka J, Ishiyama S, Saitoh D, et al. Development and testing of a ventilator remote adapter via communication between the ventilator and receiving smart device utilizing the accompanying app. Jpn J Med Instrum. 2018;88(4):449-457. [CrossRef]
  2. Michiyoshi S. Inclusion of infusion pump in automatic control system. Jpn J Intensive Care Med. 2020;44(3):117-122. URL: https://jglobal.jst.go.jp/en/detail?JGLOBAL_ID=202002286048854088 [Accessed 2025-11-10]
  3. Ishikawa M, Saito N. Strategies to prevent recurrence of incidents and accidents related to medical device alarm systems. Iryo Kikigaku. 2017;87(3):285-291. [CrossRef]
  4. Shiraishi Y, Takeda T, Shitara A. Alarm sound classification system in smartphones for the deaf and hard-of-hearing using deep neural networks. Presented at: International Conference on Advances in Computer-Human Interactions; Nov 21-25, 2020:30-33; Valencia, Spain. URL: https://www.thinkmind.org/index.php?view=article&articleid=achi_2020_3_10_28007 [Accessed 2025-10-17]
  5. Cakir E, Heittola T, Huttunen H, Virtanen T. Multi-label vs. combined single-label sound event detection with deep neural networks. Presented at: European Signal Processing Conference; Aug 31 to Sep 4, 2015:2551-2555; Nice. [CrossRef]
  6. Cakir E, Heittola T, Huttunen H, Virtanen T. Polyphonic sound event detection using multi label deep neural networks. Presented at: International Joint Conference on Neural Networks; Jul 12-17, 2015:1-7; Killarney, Ireland. [CrossRef]
  7. Mesaros A, Heittola T, Virtanen T. Metrics for polyphonic sound event detection. Appl Sci (Basel). Jun 2016;6(6):162. [CrossRef]
  8. Raboshchuk G, Nadeu C, Jancovic P, et al. A knowledge-based approach to automatic detection of equipment alarm sounds in a neonatal intensive care unit environment. IEEE J Transl Eng Health Med. 2018;6(4400110). [CrossRef] [Medline]
  9. McLoughlin I, Zhang H, Xie Z, Song Y, Xiao W. Robust sound event classification using deep neural networks. IEEE/ACM Trans Audio Speech Lang Process. Mar 2015;23(3):540-552. [CrossRef]
  10. Phan H, Hertel L, Maass M, Mertins A. Robust audio event recognition with 1-max pooling convolutional neural networks. Presented at: Annual Conference of the International Speech Communication Association; Sep 8-12, 2016. [CrossRef]
  11. Kishimoto K, Takemura T, Sugiyama O, et al. Prediction of polyphonic alarm sound by deep neural networks. Jpn Soc Med Biol Eng. Mar 2022;60(1):8-15. [CrossRef]
  12. Gong Y, Chung YA, Glass J. AST: audio spectrogram transformer. Presented at: Annual Conference of the International Speech Communication Association; Aug 30 to Sep 3, 2021. [CrossRef]
  13. Gulati A, Qin J, Chiu CC, et al. Conformer: convolution-augmented transformer for speech recognition. Presented at: Annual Conference of the International Speech Communication Association; Oct 25-29, 2020. [CrossRef]
  14. Mesaros A, Heittola T, Eronen A, Virtanen T. Acoustic event detection in real life recordings. Presented at: European Signal Processing Conference; Aug 23-27, 2010:1267-1271; Aalborg, Denmark. URL: https://ieeexplore.ieee.org/document/7096611 [Accessed 2025-10-17]
  15. Heittola T, Mesaros A, Virtanen T, Gabbouj M. Supervised model training for overlapping sound events based on unsupervised source separation. Presented at: IEEE International Conference on Acoustics, Speech and Signal Processing; May 26-31, 2013:8677-8681; Vancouver, BC. [CrossRef]
  16. Innami S, Kasai H. NMF-based environmental sound source separation using time-variant gain features. Comput Math Appl. Sep 2012;64(5):1333-1342. [CrossRef]
  17. Dessein A, Cont A, Lemaitre G. Real-time detection of overlapping sound events with non-negative matrix factorization. In: Nielsen F, Bhatia R, editors. Matrix Information Geometry. Springer; 2013:341-371. [CrossRef] ISBN: 978364230232
  18. Mesaros A, Heittola T, Dikmen O, Virtanen T. Sound event detection in real life recordings using coupled matrix factorization of spectral representations and class activity annotations. Presented at: IEEE International Conference on Acoustics, Speech and Signal Processing; Apr 19-24, 2015:151-155; Queensland, Australia. [CrossRef]
  19. Graves A, Schmidhuber J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 2005;18(5-6):602-610. [CrossRef] [Medline]
  20. Graves A, Mohamed AR, Hinton G. Speech recognition with deep recurrent neural networks. Presented at: International Conference on Acoustics, Speech and Signal Processing; May 26-31, 2013:6645-6649; Vancouver, BC. [CrossRef]
  21. Bock S, Schedl M. Polyphonic piano note transcription with recurrent neural networks. Presented at: IEEE International Conference on Acoustics, Speech and Signal Processing; Mar 25-30, 2012:121-124; Kyoto, Japan. [CrossRef]
  22. Valenti M, Squartini S, Diment A, Parascandolo G, Virtanen T. A convolutional neural network approach for acoustic scene classification. Presented at: International Joint Conference on Neural Networks; May 14-19, 2017:1547-1554; Anchorage, AK. [CrossRef]
  23. Sainath TN, Vinyals O, Senior A, Sak H. Convolutional, long short-term memory, fully connected deep neural networks. Presented at: IEEE International Conference on Acoustics, Speech and Signal Processing; Apr 19-24, 2015:4580-4584; Queensland, Australia. [CrossRef]
  24. Cakir E, Parascandolo G, Heittola T, Huttunen H, Virtanen T. Convolutional recurrent neural networks for polyphonic sound event detection. IEEE/ACM Trans Audio Speech Lang Process. 2017;25(6):1291-1303. [CrossRef]
  25. Purwins H, Li B, Virtanen T, Schluter J, Chang SY, Sainath T. Deep learning for audio signal processing. IEEE J Sel Top Signal Process. May 2019;13(2):206-219. [CrossRef]
  26. Adavanne S, Virtanen T. A report on sound event detection with different binaural features. arXiv. Preprint posted online on Oct 9, 2017. URL: http://arxiv.org/abs/1710.02997 [Accessed 2021-09-09]
  27. LeCun Y, Bengio Y, Hinton G. Deep learning. Nature New Biol. May 28, 2015;521(7553):436-444. [CrossRef] [Medline]
  28. Springenberg JT, Dosovitskiy A, Brox T, Riedmiller M. Striving for simplicity: the all convolutional net. arXiv. Preprint posted online on Apr 13, 2015. URL: http://arxiv.org/abs/1412.6806 [Accessed 2021-09-18]
  29. Kingma DP, Ba J. Adam: a method for stochastic optimization. arXiv. Preprint posted online on Jan 29, 2017. URL: http://arxiv.org/abs/1412.6980 [Accessed 2021-06-07]
  30. Ce-kishi/FEDA-medalarms. GitHub. URL: https://github.com/ce-kishi/FEDA-MedAlarms [Accessed 2025-05-25]
  31. Salamon J, Bello JP. Deep convolutional neural networks and data augmentation for environmental sound classification. IEEE Signal Process Lett. Mar 2017;24(3):279-283. [CrossRef]
  32. Park DS, Chan W, Zhang Y, et al. SpecAugment: a simple data augmentation method for automatic speech recognition. Presented at: Annual Conference of the International Speech Communication Association; Sep 15-19, 2019:2613-2617; Graz, Austria. URL: https://www.isca-archive.org/interspeech_2019 [CrossRef]
  33. TUT-ARG/sed_eval. GitHub; 2021. URL: https://github.com/TUT-ARG/sed_eval [Accessed 2021-09-19]
  34. TUT-ARG/sed_vis. GitHub; 2021. URL: https://github.com/TUT-ARG/sed_vis [Accessed 2021-09-19]
  35. Scott JB, De Vaux L, Dills C, Strickland SL. Mechanical ventilation alarms and alarm fatigue. Respir Care. Oct 2019;64(10):1308-1313. [CrossRef] [Medline]
  36. Mesaros A, Diment A, Elizalde B, et al. Sound event detection in the DCASE 2017 challenge. IEEE/ACM Trans Audio Speech Lang Process. Jun 2019;27(6):992-1006. [CrossRef]


BiGRU: bidirectional gated recurrent unit
BiSRNN: bidirectional simple recurrent neural network
CD: chest drainage
CNN: convolutional neural network
CRNN: convolutional recurrent neural network
ENP: enteral feeding pump
HWN: hospital ward noise
IEC: International Electrotechnical Commission
IP: infusion pump
ISO: International Organization for Standardization
MFB: mel filter bank
PM: patient monitor
RNN: recurrent neural network
SED: sound event detection
SNR: signal-to-noise ratio
SP: syringe pump
VFP: venous foot pump


Edited by Arriel Benis; submitted 30.Mar.2025; peer-reviewed by Kenta Hori, Manuel J C S Reis; final revised version received 27.Aug.2025; accepted 05.Oct.2025; published 12.Nov.2025.

Copyright

© Kazumasa Kishimoto, Tadamasa Takemura, Osamu Sugiyama, Ryosuke Kojima, Masahiro Yakami, Goshiro Yamamoto, Tomohiro Kuroda. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 12.Nov.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.