<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.0" xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMI</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id>
      <journal-title>JMIR Medical Informatics</journal-title>
      <issn pub-type="epub">2291-9694</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v9i4e19055</article-id>
      <article-id pub-id-type="pmid">33830067</article-id>
      <article-id pub-id-type="doi">10.2196/19055</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Retracted: Medical Data Feature Learning Based on Probability and Depth Learning Mining: Model Development and Validation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Du</surname>
            <given-names>Zhenyu</given-names>
          </name>
        </contrib>
        <contrib contrib-type="editor">
          <name>
            <surname>Kalemaki</surname>
            <given-names>Katerina</given-names>
          </name>
        </contrib>
        <contrib contrib-type="editor">
          <name>
            <surname>Li</surname>
            <given-names>Hailong</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Marcs</surname>
            <given-names>Emest</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Wang</surname>
            <given-names>Hua</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Fu</surname>
            <given-names>Jing</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Yang</surname>
            <given-names>Yuanlin</given-names>
          </name>
          <degrees>MBA, BBA</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-5014-078X</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author" corresp="yes" equal-contrib="yes">
          <name name-style="western">
            <surname>Li</surname>
            <given-names>Dehua</given-names>
          </name>
          <degrees>MPhil, MBBS</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <xref rid="aff3" ref-type="aff">3</xref>
          <address>
            <institution>Quality Assessment Office, Nursing Department</institution>
            <institution>West China Second University Hospital</institution>
            <institution>Sichuan University</institution>
            <addr-line>No 20, Section 3, Renmin South Road</addr-line>
            <addr-line>Wuhou District</addr-line>
            <addr-line>Chengdu</addr-line>
            <country>China</country>
            <phone>86 1 808 684 1792</phone>
            <email>562372162@qq.com</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-3669-7052</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Department of Logistics Management</institution>
        <institution>West China Second University Hospital</institution>
        <institution>Sichuan University</institution>
        <addr-line>Chengdu</addr-line>
        <country>China</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Key Laboratory of Obstetric and Gynecologic and Pediatric Disease and Birth Defects of Ministry of Education</institution>
        <institution>Sichuan University</institution>
        <addr-line>Chengdu</addr-line>
        <country>China</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Quality Assessment Office, Nursing Department</institution>
        <institution>West China Second University Hospital</institution>
        <institution>Sichuan University</institution>
        <addr-line>Chengdu</addr-line>
        <country>China</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Dehua Li <email>562372162@qq.com</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>4</month>
        <year>2021</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>8</day>
        <month>4</month>
        <year>2021</year>
      </pub-date>
      <volume>9</volume>
      <issue>4</issue>
      <elocation-id>e19055</elocation-id>
      <history>
        <date date-type="received">
          <day>2</day>
          <month>4</month>
          <year>2020</year>
        </date>
        <date date-type="rev-request">
          <day>14</day>
          <month>4</month>
          <year>2020</year>
        </date>
        <date date-type="rev-recd">
          <day>8</day>
          <month>5</month>
          <year>2020</year>
        </date>
        <date date-type="accepted">
          <day>8</day>
          <month>5</month>
          <year>2020</year>
        </date>
      </history>
      <copyright-statement>©Yuanlin Yang, Dehua Li. Originally published in JMIR Medical Informatics (http://medinform.jmir.org), 08.04.2021.</copyright-statement>
      <copyright-year>2021</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on http://medinform.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://medinform.jmir.org/2021/4/e19055" xlink:type="simple"/>
      <related-article related-article-type="retraction-forward" id="v14i1e98673" ext-link-type="doi" xlink:href="10.2196/98673" vol="14" page="e98673" xlink:type="simple">https://medinform.jmir.org/2026/1/e98673</related-article>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Big data technology provides unlimited potential for efficient storage, processing, querying, and analysis of medical data. Technologies such as deep learning and machine learning simulate human thinking, assist physicians in diagnosis and treatment, provide personalized health care services, and promote the use of intelligent processes in health care applications.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>The aim of this paper was to analyze health care data and develop an intelligent application to predict the number of hospital outpatient visits for mass health impact and analyze the characteristics of health care big data. Designing a corresponding data feature learning model will help patients receive more effective treatment and will enable rational use of medical resources.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>A cascaded depth model was successfully implemented by constructing a cascaded depth learning framework and by studying and analyzing the specific feature transformation, feature selection, and classifier algorithm used in the framework. To develop a medical data feature learning model based on probabilistic and deep learning mining, we mined information from medical big data and developed an intelligent application that studies the differences in medical data for disease risk assessment and enables feature learning of the related multimodal data. Thus, we propose a cascaded data feature learning model.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>The depth model created in this paper is more suitable for forecasting daily outpatient volumes than weekly or monthly volumes. We believe that there are two reasons for this: on the one hand, the training data set in the daily outpatient volume forecast model is larger, so the training parameters of the model more closely fit the actual data relationship. On the other hand, the weekly and monthly outpatient volume is the cumulative daily outpatient volume; therefore, errors caused by the prediction will gradually accumulate, and the greater the interval, the lower the prediction accuracy.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>Several data feature learning models are proposed to extract the relationships between outpatient volume data and obtain the precise predictive value of the outpatient volume, which is very helpful for the rational allocation of medical resources and the promotion of intelligent medical treatment.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>deep learning</kwd>
        <kwd>data mining</kwd>
        <kwd>medical big data</kwd>
        <kwd>model building</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>Over the past two decades, there has been dramatic growth in the amount of data being generated in many areas worldwide, including health care data, sensor data, various types of user-generated data, internet data, and financial company data. Big data is emerging as the amount of data in every field grows; however, “big data” is an abstract concept that does not simply mean a large collection of data. Big data has some features that are different from data sets, and its characteristics differ from those of massive data and large data sets. Research studies examining concepts such as the Internet of Things and wearable technology have helped reduce the cost of real-time monitoring of human health, which has driven development in this industry [<xref ref-type="bibr" rid="ref1">1</xref>]. Big data technology provides unlimited potential for efficient storage, processing, querying, and analysis of medical data. Technologies such as deep learning and machine learning simulate human thinking, assist physicians in diagnosis and treatment, provide personalized health services, and promote intelligent processes of health care applications. The development and application of the Internet of Things, wireless networks, the internet, cloud computing technology, etc., provide guarantees for the analysis, processing, and transmission of big data. In short, in the field of health care, the rapid development of big data analysis, wearable technology, artificial intelligence, Kyrgyz computing, supercomputing technology, etc., all provide possibilities for the realization and development of smart medical applications [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref3">3</xref>].</p>
      <p>Medical health data is multimodal, complex data that continues to grow rapidly and contains a wealth of information. The challenges associated with medical health data include how to quickly and accurately collect medical health data and how to efficiently use high-speed networks to reliably and efficiently transmit medical health data [<xref ref-type="bibr" rid="ref4">4</xref>]. Other challenges include the use artificial intelligence–related machine learning and deep learning techniques to extract useful information from health medical big data and the development of intelligent applications for medical staff and ordinary people. In this paper, our aims included analyzing health care data, addressing intelligent application–related issues, predicting the number of hospital outpatient visits for mass health impacts, and analyzing the characteristics of health care big data. Designing a corresponding data feature learning model will help patients receive more effective treatment and enable rational use of medical resources.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Cascaded Deep Learning Model</title>
        <sec>
          <title>Model Framework</title>
          <p>Deep learning is a process of feature extraction and combination. Through multilayer nonlinear operation combination, the model can abstract high-order semantic information as data. In practice, a cascaded multilayer operation is performed on the preprocessed raw data. Each layer consists of three sublayers: feature extraction, nonlinear transformation, and feature selection. The input of the initial layer is the pretreated original data, the input of the second layer is the output of the upper layer, and the output of the third layer is the final abstract representation of the data. In each layer, the representation characteristics of the data are extracted by feature transformation, which is generally a process of dimension increasing [<xref ref-type="bibr" rid="ref5">5</xref>]. Compared with the input, the transformed features have certain representative characteristics; however, the number of features is much greater. Through feature selection, we reduce the dimension once, and at the same time, we choose the discriminant feature or the representation feature. Some feature selection methods are used to improve the regional adaptability of the model, such as max-pooling and mean-pooling operations in convolutional neural networks. Nonlinear transformation is usually performed in the middle of feature transformation and feature selection, and it is an important part of the framework. Nonlinear transformation can imitate the activation and inactivation of neurons. Another important application of nonlinear transformation is when linear transformation is used in the feature extraction. Multilayer linear transformation is still a linear transformation. Multilayer operation plays the same role as learning a linear transformation directly and does not play a role of layer-by-layer abstraction.</p>
          <p>In this paper, to eliminate possible differences in measurement scales between different features, the original data were normalized by the maximum and minimum method. In the feature extraction stage, to obtain the structural information of the data, a method of feature transformation with little relationship to the domain was selected, that is, the vanishing component analysis (VCA) method was used to extract the polynomial characteristics of the data. Because the VCA method itself is a nonlinear transformation, the nonlinear operation cannot be carried out in the framework [<xref ref-type="bibr" rid="ref6">6</xref>]. Because the dimension of the VCA input data cannot be too high, we used principal component analysis (PCA) to reduce the dimensions of the VCA input data (PCA reduces the dimension of the loss of information and can reduce the feature dimensions; therefore, it can also be regarded as a special feature selection method). In the final classification, we used a boosting algorithm, which is a method that can classify and select features; in the classification of the use of the features, we could use all the features learned at each level or only use the last level of features. At this point, our depth feature learning framework was formed, as shown in <xref rid="figure1" ref-type="fig">Figure 1</xref>.</p>
          <fig id="figure1" position="float">
            <label>Figure 1</label>
            <caption>
              <p>The cascaded deep learning framework. IVI: innovation value intuitive; AdaBoost: adaptive boost; PCA: principal component analysis; RSR: regularized self-representation; VCA: vanishing component analysis.</p>
            </caption>
            <graphic xlink:href="medinform_v9i4e19055_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Characteristic Learning</title>
          <p>According to the existing cascaded depth learning framework, we implemented a specific depth model. Therefore, we needed to study and explain the training process of the model. The cascade depth model proposed in this paper is a multilayer structure. Between different layers, the output of the upper layer is the input of the next layer. Within the same layer, a PCA dimensionality reduction stage, VCA feature transformation stage, and <italic>L<sub>2,p</sub> – RSR</italic> feature selection stage, where RSR is regularized self-representation, are included. Each layer of the model can learn the output features of the current layer [<xref ref-type="bibr" rid="ref7">7</xref>]. The abstract information of the different layers is different. The features of the lowest layer are closest to the original feature space. The high-level features can provide complementary information for the low-level features. We made full use of the characteristics learned from all levels and proposed an effective feature combination method. Finally, we used the boosting classifier based on the binary classification problem and extended its success to the multiclassification problem.</p>
        </sec>
        <sec>
          <title>PCA Dimension Reduction Stage</title>
          <p>VCA feature transformation requires stringent data space dimension control; therefore, it was necessary to reduce the dimension before the VCA feature transformation. There are two ways to reduce the dimension of PCA; one is to specify the reserved dimension directly, and the other is to set the ratio of an eigenvalue to the total sum of the eigenvalues. Proportion setting can theoretically control the retention percentage of data information; however, it is challenging to control the retained feature dimension using this method, especially when the feature dimension is difficult to control. In this model, the VCA transform will produce a large number of features, and its input space needs to strictly control the feature dimension. Therefore, we directly set the reservation dimension of the PCA transform [<xref ref-type="bibr" rid="ref8">8</xref>]. Because different dimensions retain different information, the performance of the model and the experimental results are affected, and this effect is not positively correlated with the retention dimension. Therefore, in our experiments, it was necessary to debug the PCA dimension.</p>
        </sec>
        <sec>
          <title>VCA Feature Transformation Stage</title>
          <p>The VCA method can map raw data into zero polynomial space, thus playing the role of feature extraction. Using this method, it is not necessary to know the domain of data usage or other prior knowledge because as long as the input space is a real matrix, we can learn its zero space polynomial transformation representation. This transformation method can not only extract the linear features in the data samples but can also extract the nonlinear features, that is, if the first-order polynomial contains the zero space of the data, the polynomial contains the linear information of the data. Other polynomials with different numbers can extract 2 or even higher order nonlinear characteristics of data. In this model, the use of VCA for feature transformation involves two specific problems: (1) polynomial number setting and initial feature dimension setting; (2) algorithm solution using singular value decomposition with the minimum setting.</p>
        </sec>
        <sec>
          <title>L<sub>2,p</sub> – RSR Feature Selection Phase</title>
          <p>VCA feature transformation will produce a large number of features, and we need to select features for many reasons. Moreover, feature selection is one of the reasons why the depth learning model is effective, as it can effectively select task-related features. The <italic>L<sub>2,p</sub> – RSR</italic> method proposed in this paper can not only effectively select features that play important roles in the linear representation of features but can also exclude the roles of singular samples due to the use of <italic>L<sub>2,1</sub></italic> norm constraint loss terms. This method is based on the self-representation property of the feature space. Any matrix space data possesses this property; therefore, it is domain-independent, which meets the requirements of the generalization ability of our model. In this model, we only needed to set the <italic>P</italic> value of <italic>L<sub>2,p</sub></italic> norm and regularization parameter λ value in the method. The input of the <italic>L<sub>2,p</sub> – RSR</italic> feature selection operation is the output space after VCA transformation, and the output is the output feature of the current layer of the depth model.</p>
        </sec>
        <sec>
          <title>Boosting Classification and Feature Selection</title>
          <p>After features are learned, it is necessary to classify them. There are many general classification methods, of which the nearest neighbor classifier is the simplest. Support vector machine (SVM) classifiers are also widely used in research and applications, and kernel-based SVM classifiers can also solve nonlinear problems. For this model, we used a classification method with a feature selection function: the Gentle Adaptive Boosting (AdaBoost) classifier based on a pile function. The Gentle AdaBoost algorithm based on a pile function can not only classify but can also select features from the feature space, that is, it can select one feature from each feature space and classify it, and it can select features by controlling the number of weak classifiers [<xref ref-type="bibr" rid="ref9">9</xref>]. This is in good agreement with the framework proposed in this paper. The Gentle AdaBoost algorithm can not only play a classification role but can also perform feature selection. It can also be used only as a classification method or for feature selection. Next, we analyzed and implemented the boosting classifier based on a pile function.</p>
        </sec>
        <sec>
          <title>Brief Summary of the Discrete AdaBoost Algorithm Based on a Pile Function</title>
          <p>For the discrete AdaBoost algorithm based on a pile function, the inputs are training sample X and tag Y, and the output is the classifier model <inline-graphic xlink:href="medinform_v9i4e19055_fig7.png" xlink:type="simple" mimetype="image"/>.</p>
          <p>Step 1: Weight matrix initialization: <italic>w<sub>i</sub></italic> = 1/<italic>m</italic>,<italic>i</italic> = 1, …, <italic>m</italic></p>
          <p>Step 2: Repeat: <italic>t</italic>= 1, 2, …, <italic>T</italic></p>
          <p>Step 3: For <italic>d</italic> = 1, …, <italic>n</italic>, <italic>do</italic>: (<italic>err<sup>d</sup></italic>, <italic>δ<sup>d</sup></italic>, <italic>a<sup>d</sup></italic>, <italic>b<sup>d</sup></italic>) = <inline-graphic xlink:href="medinform_v9i4e19055_fig8.png" xlink:type="simple" mimetype="image"/></p>
          <p>Step 4: <italic>feaId</italic> = ar g min (<italic>err<sub>d</sub></italic>), (<italic>feaId, δ</italic>, <italic>a</italic>, <italic>b</italic>) = (<italic>feaId,</italic> <italic>δ <sup>feaId</sup></italic>, <italic>a<sup>feaId</sup></italic>, <italic>b<sup>feaId</sup></italic>)</p>
          <p>Step 5: <italic>f<sub>t</sub></italic>(<italic>x</italic>) = <italic>ah</italic>(<italic>x <sup>feaId</sup></italic> &gt; <italic>δ</italic>) + <italic>b</italic></p>
          <p>Step 6: Update: <inline-graphic xlink:href="medinform_v9i4e19055_fig9.png" xlink:type="simple" mimetype="image"/>, <italic>i</italic> = 1, 2, …, <italic>m</italic>, standardization of <italic>w</italic> makes <inline-graphic xlink:href="medinform_v9i4e19055_fig10.png" xlink:type="simple" mimetype="image"/></p>
          <p>Step 7: Repeating end</p>
          <p>The Gentle AdaBoost algorithm based on the pile function follows the boosting algorithm framework and uses a simple classifier model:</p>
          <p>
            <disp-formula>
              <graphic xlink:href="medinform_v9i4e19055_fig11.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
            </disp-formula>
          </p>
          <p>in which the weak classifier <italic>f<sub>m</sub></italic> is defined as</p>
          <p>
            <disp-formula>
              <graphic xlink:href="medinform_v9i4e19055_fig12.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
            </disp-formula>
          </p>
          <p>The <italic>h</italic> function is the indicator function. represents the <italic>d</italic> dimension characteristic of the <italic>i</italic> sample. <italic>δ</italic> is the threshold (the so-called pile). <italic>a</italic> and <italic>b</italic> are parameters of the linear regression function. When learning the weak classifier, each feature of the sample learns a pile function based on the least squares, and the error value of the least squares is obtained and recorded; then, the corresponding feature is selected when the error value is the minimum. Therefore, (<italic>d</italic>, <italic>δ</italic>, <italic>a</italic>, <italic>b</italic>) can be obtained through the weighted least squares method:</p>
          <p>
            <disp-formula>
              <graphic xlink:href="medinform_v9i4e19055_fig13.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
            </disp-formula>
          </p>
          <p>After obtaining the weak classifier <italic>F<sub>m</sub></italic>, the weights of <italic>W</italic> are updated:</p>
          <p>
            <disp-formula>
              <graphic xlink:href="medinform_v9i4e19055_fig14.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
            </disp-formula>
          </p>
          <p>The F function is updated to <italic>F = F</italic> + <italic>f</italic>. The final classification result is <italic>sign</italic> (<italic>F</italic>(<italic>x</italic>)). The absolute value of the <italic>F</italic>(<italic>x</italic>) value provides the credibility of the classification.</p>
        </sec>
      </sec>
      <sec>
        <title>Feature Combination</title>
        <p>Through the research and analysis of each stage of the cascade depth model, we successfully constructed a multilevel depth learning model and learned multilevel features. The different layer features provide different information. The underlying feature is the closest to the real information of the data and contains the most data information; however, it contains less semantic information and is not sufficiently abstract. When a model has two or three layers, it abstracts the features at different levels, including certain semantic information. When the number of layers is higher, the model contains a higher level of abstract information. We believe that each layer of features is useful; the underlying features ensure that the data information distortion is not too great, and the high-level features can provide the underlying features with complementary structure information that the underlying features do not possess [<xref ref-type="bibr" rid="ref10">10</xref>]. If we can effectively combine each layer of features, we will obtain a good machine learning model. Using the combination of Gentle AdaBoost feature selection and classification functions, in this section, we propose an effective method of combining each layer of features and classifying them using the combination features [<xref ref-type="bibr" rid="ref11">11</xref>-<xref ref-type="bibr" rid="ref14">14</xref>]. The combination of the two levels of features is shown in <xref rid="figure2" ref-type="fig">Figure 2</xref>.</p>
        <p>The classifier we used in the cascaded depth model is the Gentle AdaBoost classifier with a feature selection function. Therefore, we input the features of the different layers into the classifier algorithm to classify them and achieve the purpose of combining the features of each layer. The Gentle AdaBoost classifier is in the form of <inline-graphic xlink:href="medinform_v9i4e19055_fig15.png" xlink:type="simple" mimetype="image"/>. Assuming that the feature space of the first layer is X<sub>1</sub> and the feature space of the second layer is X<sub>2</sub>, a strong classifier <inline-graphic xlink:href="medinform_v9i4e19055_fig16.png" xlink:type="simple" mimetype="image"/> is learned from the features of the first layer. Because the classification algorithm gives the current weight W of each sample after each weak classification, it can initialize the sample weight before F<sub>2</sub> training by using the weight distribution W of the sample at the end of F<sub>1</sub> after F<sub>1</sub> has been generated. Then, we learn the classifier <inline-graphic xlink:href="medinform_v9i4e19055_fig17.png" xlink:type="simple" mimetype="image"/>, and the final classification result is F<sub>1</sub> + F<sub>2</sub>. F<sub>1</sub> uses only the first layer feature and F<sub>2</sub> uses only the second layer feature; however, the sample weight W is used in the F<sub>1</sub> process.</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>The combination of two levels of features used in our model.</p>
          </caption>
          <graphic xlink:href="medinform_v9i4e19055_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Data Analysis</title>
        <p>We used the Letter, Pendigit, and USPS data sets to conduct comparative experiments [<xref ref-type="bibr" rid="ref15">15</xref>-<xref ref-type="bibr" rid="ref18">18</xref>]. Specifically, the example, feature, and class labels are 20,000, 16, and 26 for the Letter data set, 10,992, 16, and 10 for the Pendigit data set, and 9298, 256, and 10 for the USPS data set, respectively. From <xref ref-type="table" rid="table1">Table 1</xref>, we can see that the data pair classification accuracy was low; we selected the lowest number, and the corresponding sequence was (31 41 45 125 128 157 164 173 304). The bandwidth of the first layer PCA was (4810 1316), that of the second layer was (358 11 13 15 20) and that of the third layer as (358 11 13 15). The classification accuracy of the selected data under one level of PCA is shown in <xref ref-type="table" rid="table1">Table 1</xref>. As can be seen from the table, the optimal value was PCA1=13. With a fixed PCA1 of 13, the second layer under different PCA2 values of classification accuracy is shown in <xref ref-type="table" rid="table2">Table 2</xref>; the optimal value was PCA2=11. Moreover. the third layer results are shown in <xref ref-type="table" rid="table2">Table 2</xref>, and the corresponding PCA3 was 13. In actual bandwidth settings, PCA reserve values can be set for several more groups because Letter data have a smaller sample size per class. Finally, each layer used the PCA reservation dimension (13 11 13) setting to obtain the total classification accuracy of Letter data under each layer. <xref ref-type="table" rid="table1">Tables 1</xref>-<xref ref-type="table" rid="table2">2</xref> also show that the classification accuracy increases with the increase of the number of layers when the model classifies Letter data with attribute values. <xref rid="figure3" ref-type="fig">Figure 3</xref> and <xref rid="figure4" ref-type="fig">Figure 4</xref> show the classification accuracies of the Pendigit and USPS data sets at different layers, respectively.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Classification accuracy of the first layer of Letter data with different PCA1 retention values (%).</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="250"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <col width="150"/>
            <thead>
              <tr valign="top">
                <td>Data serial number</td>
                <td colspan="5">PCA1<sup>a</sup> retention value</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>4</td>
                <td>8</td>
                <td>10</td>
                <td>13</td>
                <td>16</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>31</td>
                <td>92.10</td>
                <td>98.63</td>
                <td>99.62</td>
                <td>99.62</td>
                <td>99.13</td>
              </tr>
              <tr valign="top">
                <td>41</td>
                <td>85.93</td>
                <td>94.58</td>
                <td>94.58</td>
                <td>97.59</td>
                <td>98.51</td>
              </tr>
              <tr valign="top">
                <td>45</td>
                <td>97.13</td>
                <td>98.46</td>
                <td>98.89</td>
                <td>98.89</td>
                <td>97.90</td>
              </tr>
              <tr valign="top">
                <td>125</td>
                <td>83.87</td>
                <td>94.20</td>
                <td>96.78</td>
                <td>98.52</td>
                <td>98.52</td>
              </tr>
              <tr valign="top">
                <td>128</td>
                <td>89.92</td>
                <td>95.96</td>
                <td>96.98</td>
                <td>99.65</td>
                <td>99.65</td>
              </tr>
              <tr valign="top">
                <td>157</td>
                <td>84.59</td>
                <td>96.33</td>
                <td>94.92</td>
                <td>97.87</td>
                <td>95.89</td>
              </tr>
              <tr valign="top">
                <td>164</td>
                <td>94.10</td>
                <td>95.12</td>
                <td>97.42</td>
                <td>98.80</td>
                <td>99.27</td>
              </tr>
              <tr valign="top">
                <td>173</td>
                <td>91.83</td>
                <td>95.11</td>
                <td>96.89</td>
                <td>98.58</td>
                <td>97.30</td>
              </tr>
              <tr valign="top">
                <td>304</td>
                <td>85.66</td>
                <td>96.92</td>
                <td>99.29</td>
                <td>100.00</td>
                <td>100.00</td>
              </tr>
              <tr valign="top">
                <td>Average value</td>
                <td>89.60</td>
                <td>96.18</td>
                <td>97.28</td>
                <td>98.89</td>
                <td>98.72</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>PCA1: principal component analysis 1.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Classification accuracy of Letter data under different PCA2 retention values (%).</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <colgroup>
              <col width="400"/>
            </colgroup>
            <colgroup>
              <col width="100"/>
            </colgroup>
            <colgroup>
              <col width="100"/>
              <col width="100"/>
            </colgroup>
            <colgroup>
              <col width="100"/>
            </colgroup>
            <colgroup>
              <col width="100"/>
            </colgroup>
            <colgroup>
              <col width="100"/>
            </colgroup>
            <thead>
              <tr valign="top">
                <td>Data serial number</td>
                <td colspan="6">PCA2<sup>a</sup> retention value</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>3</td>
                <td>5</td>
                <td>8</td>
                <td>11</td>
                <td>13</td>
                <td>15</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>31</td>
                <td>99.62</td>
                <td>99.62</td>
                <td>99.62</td>
                <td>99.13</td>
                <td>99.62</td>
                <td>99.62</td>
              </tr>
              <tr valign="top">
                <td>41</td>
                <td>97.59</td>
                <td>97.99</td>
                <td>97.59</td>
                <td>98.59</td>
                <td>98.59</td>
                <td>97.89</td>
              </tr>
              <tr valign="top">
                <td>45</td>
                <td>99.78</td>
                <td>99.78</td>
                <td>99.23</td>
                <td>99.78</td>
                <td>99.23</td>
                <td>99.78</td>
              </tr>
              <tr valign="top">
                <td>125</td>
                <td>99.39</td>
                <td>99.39</td>
                <td>99.39</td>
                <td>99.39</td>
                <td>99.82</td>
                <td>99.25</td>
              </tr>
              <tr valign="top">
                <td>128</td>
                <td>99.65</td>
                <td>100.00</td>
                <td>99.65</td>
                <td>99.65</td>
                <td>99.65</td>
                <td>99.65</td>
              </tr>
              <tr valign="top">
                <td>157</td>
                <td>98.94</td>
                <td>98.94</td>
                <td>98.94</td>
                <td>98.94</td>
                <td>98.49</td>
                <td>98.94</td>
              </tr>
              <tr valign="top">
                <td>164</td>
                <td>99.19</td>
                <td>99.19</td>
                <td>99.73</td>
                <td>98.73</td>
                <td>99.19</td>
                <td>99.19</td>
              </tr>
              <tr valign="top">
                <td>173</td>
                <td>98.91</td>
                <td>98.91</td>
                <td>98.91</td>
                <td>98.91</td>
                <td>98.91</td>
                <td>98.58</td>
              </tr>
              <tr valign="top">
                <td>304</td>
                <td>100.00</td>
                <td>100.00</td>
                <td>100.00</td>
                <td>100.00</td>
                <td>100.00</td>
                <td>100.00</td>
              </tr>
              <tr valign="top">
                <td>Average value</td>
                <td>99.59</td>
                <td>99.69</td>
                <td>99.58</td>
                <td>99.59</td>
                <td>99.75</td>
                <td>99.69</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>PCA2 : principal component analysis 2.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <fig id="figure3" position="float">
          <label>Figure 3</label>
          <caption>
            <p>Classification accuracy of the Pendigit data set at different layers.</p>
          </caption>
          <graphic xlink:href="medinform_v9i4e19055_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure4" position="float">
          <label>Figure 4</label>
          <caption>
            <p>Classification accuracy of the USPS data set at different levels.</p>
          </caption>
          <graphic xlink:href="medinform_v9i4e19055_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Prediction of Hospital Outpatient Volume Based on the RNN-RBM Network Model</title>
        <sec>
          <title>Prediction Model Construction</title>
          <p>Recurrent neural networks (RNNs) and restricted Boltzmann machine (RBM) networks have strong ability to predict time series. To make use of the advantages of these two network models to predict hospital outpatient volumes, the two networks were combined to form a depth structure network RNN-RBM model. This deep neural network can describe the time dependence of high-dimensional sequences.</p>
          <p>Among the various deformations of RBM, there is a deformation called conditional RBM. Conditional RBM is different from standard RBM in that it adds two types of connections: one is an autoregressive connection, and the other is a hidden layer connection between the previous time step and the current time step. A conditional RBM can also be trained by a contrastive divergence (CD) algorithm. This structure can be used to handle time series data effectively; therefore, it can be used to solve time series problems. The RNN-RBM model in this paper can also refer to conditional RBM. <inline-graphic xlink:href="medinform_v9i4e19055_fig18.png" xlink:type="simple" mimetype="image"/> and <inline-graphic xlink:href="medinform_v9i4e19055_fig19.png" xlink:type="simple" mimetype="image"/> represent the offset vectors of the visible layer and the hidden layer in the RBM model at time t, and they are updated by the hidden unit <italic>u<sup>t – 1</sup></italic> in the RNN model at time t – 1. Weight matrices <italic>W<sub>uv</sub></italic> and <italic>W<sub>uh</sub></italic> are used to connect the RNN network model and RBM network model. The bias vector can be expressed as follows:</p>
          <p>
            <disp-formula>
              <graphic xlink:href="medinform_v9i4e19055_fig20.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
            </disp-formula>
          </p>
          <p>where <italic>b<sub>v</sub></italic> and <italic>b<sub>h</sub></italic> are the initial offset values of the visible and hidden layers in the RBM network model. The RNN network model expands gradually with the time step and is used to generate the state of the hidden units in the RBM network model, which are based on the input layer <italic>v</italic><sup>(</sup><italic><sup>t</sup></italic><sup>)</sup> and the hidden layer <italic>u</italic><sup>(</sup><italic><sup>t</sup></italic><sup>)</sup> in the RNN network model. In this way, the hidden layer can only blame the activation function of the hidden unit.</p>
          <disp-formula>
            <italic>u<sup>(t)</sup> = sigmoid (b<sub>u</sub>+W<sub>uu</sub>u<sup>(t-1)</sup>+W<sub>vu</sub>v<sup>(t)</sup>)</italic>
          </disp-formula>
          <p>From this, we can see that the overall process of the algorithm is:</p>
          <list list-type="order">
            <list-item>
              <p>The hidden unit in the RNN model is activated.</p>
            </list-item>
            <list-item>
              <p><italic>u<sup>(t-1)</sup></italic> is used in Step 1 to update the bias values in the RBM network model.</p>
            </list-item>
            <list-item>
              <p>The parameters are updated in the RBM network model.</p>
            </list-item>
            <list-item>
              <p>The RBM output is used as the input of the prediction layer, and the parameters are initialized randomly.</p>
            </list-item>
            <list-item>
              <p>The backpropagation (BP) method is used to fine-tune the model from top to bottom. Error values are propagated back to the RBM and RNN network models. The weight matrices <italic>w<sub>uv</sub></italic> and <italic>w<sub>uh</sub></italic> are updated, and the RNN network models are trained to predict.</p>
            </list-item>
          </list>
        </sec>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <p>The experimental simulation data were obtained from real-world hospital outpatient volume data. In the data pretreatment stage, in this paper, we extracted the outpatient volume information of each department of the hospital and performed certain statistical processing methods. In the actual processing data, the outpatient data for hospital holidays were significantly reduced; to avoid the impact of these data, we deleted the related statistical processing methods.</p>
      <p>To better satisfy the prediction function, these data were counted according to the three time intervals of day, week, and month. To better conform to the prediction model based on the depth neural network proposed in this paper, these data were processed and expressed as a data matrix. The matrix is as follows:</p>
      <p>
        <disp-formula>
          <graphic xlink:href="medinform_v9i4e19055_fig21.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </disp-formula>
      </p>
      <p>in which <inline-graphic xlink:href="medinform_v9i4e19055_fig22.png" xlink:type="simple" mimetype="image"/> represents the outpatient volume data of department <italic>n</italic> in interval <italic>T</italic>.</p>
      <p>In this paper, RNN and RBM neural networks were combined to form a deep-seated neural network model, and the model was used to predict the outpatient volume of the hospital. In the actual simulation, we selected the outpatient volume data of 15 important outpatient clinics as the input of the depth model, that is, the input layer of the model was set to 15. The number of hidden layer neurons in the RNN was set to 20, the number of hidden layer neurons in the RBM was set to 30, and the output of the predicted layer was 15.</p>
      <p><xref rid="figure5" ref-type="fig">Figure 5</xref> and <xref rid="figure6" ref-type="fig">Figure 6</xref> show the prediction results of the RNN-RBM model. In the simulation experiment, the outpatient volume data were trained in different intervals (daily outpatient volume, weekly outpatient volume, and monthly outpatient volume). From the simulation results, it can be seen that the forecast of the daily outpatient volume is closer to the real value than that of weekly outpatient volume and monthly outpatient volume. The depth model created in this paper is more suitable for the daily outpatient volume forecast, and we analyzed the causes of this phenomenon. We believe that there are two reasons for this result. On the one hand, the amount of training data in the daily outpatient volume forecast model is larger; therefore, the training parameters of the model more closely fit the actual data relationship. On the other hand, the weekly and monthly outpatient volume is the cumulative daily outpatient volume; therefore, the errors caused by the prediction will gradually accumulate, and the greater the interval, the lower the prediction accuracy.</p>
      <p>In practical applications, medical managers often require more short-term outpatient volume forecasts. Because forecasting a shorter interval of outpatient volume can provide support for medical management, this method still has certain advantages in practical applications.</p>
      <p>We compared the outpatient volume forecasting method based on the RNN-RBM depth model with existing popular forecasting algorithms, and the results are shown in <xref ref-type="table" rid="table3">Table 3</xref>. Here, the prediction algorithm based on the auto regressive integrated moving average (ARIMA) model, the BP neural network prediction algorithm, the radial basis function (RBF) neural network prediction algorithm, and the SVM algorithm were selected.</p>
      <p>Compared with the current popular outpatient volume prediction algorithm, it can be readily concluded that the prediction algorithm based on the RNN-RBM depth model is superior to other current prediction algorithms for the daily outpatient volume, weekly outpatient volume, and monthly outpatient volume, and the prediction accuracy is relatively high.</p>
      <fig id="figure5" position="float">
        <label>Figure 5</label>
        <caption>
          <p>Daily outpatient volume forecast results.</p>
        </caption>
        <graphic xlink:href="medinform_v9i4e19055_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
      <fig id="figure6" position="float">
        <label>Figure 6</label>
        <caption>
          <p>Weekly outpatient volume forecast results.</p>
        </caption>
        <graphic xlink:href="medinform_v9i4e19055_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
      </fig>
      <table-wrap position="float" id="table3">
        <label>Table 3</label>
        <caption>
          <p>Comparison of outpatient volume prediction methods.</p>
        </caption>
        <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
          <col width="30"/>
          <col width="370"/>
          <col width="200"/>
          <col width="200"/>
          <col width="200"/>
          <thead>
            <tr valign="top">
              <td colspan="2">Method and error</td>
              <td>Daily outpatient volume</td>
              <td>Weekly outpatient volume</td>
              <td>Monthly outpatient volume</td>
            </tr>
          </thead>
          <tbody>
            <tr valign="top">
              <td colspan="5">
                <bold>RNN-RBM<sup>a</sup></bold>
              </td>
            </tr>
            <tr valign="top">
              <td rowspan="2">
                <break/>
              </td>
              <td>MAE<sup>b</sup></td>
              <td>24.83</td>
              <td>2395.1</td>
              <td>11966.6</td>
            </tr>
            <tr valign="top">
              <td>MRE<sup>c</sup> (%)</td>
              <td>3.9</td>
              <td>4.6</td>
              <td>5.9</td>
            </tr>
            <tr valign="top">
              <td colspan="5">
                <bold>ARIMA<sup>d</sup></bold>
              </td>
            </tr>
            <tr valign="top">
              <td rowspan="2">
                <break/>
              </td>
              <td>MAE</td>
              <td>37.9</td>
              <td>3562.7</td>
              <td>19525.5</td>
            </tr>
            <tr valign="top">
              <td>MRE (%)</td>
              <td>5.4</td>
              <td>6.1</td>
              <td>6.8</td>
            </tr>
            <tr valign="top">
              <td colspan="5">
                <bold>BP<sup>e</sup></bold>
              </td>
            </tr>
            <tr valign="top">
              <td rowspan="2">
                <break/>
              </td>
              <td>MAE</td>
              <td>53.48</td>
              <td>4145.5</td>
              <td>2549.5</td>
            </tr>
            <tr valign="top">
              <td>MRE (%)</td>
              <td>7.3</td>
              <td>8.9</td>
              <td>9.6</td>
            </tr>
            <tr valign="top">
              <td colspan="5">
                <bold>SVM<sup>f</sup></bold>
              </td>
            </tr>
            <tr valign="top">
              <td rowspan="2">
                <break/>
              </td>
              <td>MAE</td>
              <td>44.3</td>
              <td>3796.9</td>
              <td>2647.8</td>
            </tr>
            <tr valign="top">
              <td>MRE (%)</td>
              <td>6.4</td>
              <td>7.9</td>
              <td>9.4</td>
            </tr>
            <tr valign="top">
              <td colspan="5">
                <bold>RBF<sup>g</sup></bold>
              </td>
            </tr>
            <tr valign="top">
              <td rowspan="2">
                <break/>
              </td>
              <td>MAE</td>
              <td>38.9</td>
              <td>3586.3</td>
              <td>22978</td>
            </tr>
            <tr valign="top">
              <td>MRE (%)</td>
              <td>5.7</td>
              <td>7.1</td>
              <td>9</td>
            </tr>
          </tbody>
        </table>
        <table-wrap-foot>
          <fn id="table3fn1">
            <p><sup>a</sup>RNN-RBM: recurrent neural network-restricted Boltzmann machine.</p>
          </fn>
          <fn id="table3fn2">
            <p><sup>b</sup>MAE: mean absolute error.</p>
          </fn>
          <fn id="table3fn3">
            <p><sup>c</sup>MRE: mean relative error.</p>
          </fn>
          <fn id="table3fn4">
            <p><sup>d</sup>ARIMA: auto regressive integrated moving average.</p>
          </fn>
          <fn id="table3fn5">
            <p><sup>e</sup>BP: backpropagation.</p>
          </fn>
          <fn id="table3fn6">
            <p><sup>f</sup>SVM: support vector machine.</p>
          </fn>
          <fn id="table3fn7">
            <p><sup>g</sup>RBF: radial basis function.</p>
          </fn>
        </table-wrap-foot>
      </table-wrap>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <p>Big data in the field of health care is an integral part of the strategic layout of national big data, and the analysis and mining of valuable information is also related to the development of national health care. At present, the problems that must be solved in the analysis and application of health care data include timely and accurate collection and acquisition of health care data as well as efficient use of high-speed networks for reliable transmission of health care–related digital, image, voice, and other information. Machine learning and in-depth learning technology related to artificial intelligence can be used to mine useful information from health care–related big data and develop intelligent applications for medical staff and ordinary people. In this paper, we studied a feature learning model of medical health data based on probabilistic and in-depth learning mining, mining information from medical big data and addressing intelligent application–related problems, and we studied the differences between medical risk assessment–related data and general big data, multimodal data feature representation, and learning–related content. Several data feature learning models are proposed to extract the relationship between the outpatient volume data and to obtain precise predictive value of outpatient volume, which is very helpful to the rational allocation of medical resources and the promotion of intelligent medical treatment.</p>
    </sec>
  </body>
  <back>
    <app-group/>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AdaBoost</term>
          <def>
            <p>adaptive boosting</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">ARIMA</term>
          <def>
            <p>auto regressive integrated moving average</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">BP</term>
          <def>
            <p>backpropagation</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">CD</term>
          <def>
            <p>contrastive divergence</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">PCA</term>
          <def>
            <p>principal component analysis</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">RBF</term>
          <def>
            <p>radial basis function</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">RBM</term>
          <def>
            <p>restricted Boltzmann machine</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">RNN</term>
          <def>
            <p>recurrent neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">RSR</term>
          <def>
            <p>regularized self-representation</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">SVM</term>
          <def>
            <p>support vector machine</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">VCA</term>
          <def>
            <p>vanishing component analysis</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>The authors acknowledge the Soft Science Project of the Sichuan Science and Technology Department (Grant: 2017ZR0169).</p>
    </ack>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nie</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Akbari</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chua</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Bridging the Vocabulary Gap between Health Seekers and Healthcare Knowledge</article-title>
          <source>IEEE Trans Knowl Data Eng</source>
          <year>2015</year>
          <month>2</month>
          <day>1</day>
          <volume>27</volume>
          <issue>2</issue>
          <fpage>396</fpage>
          <lpage>409</lpage>
          <pub-id pub-id-type="doi">10.1109/tkde.2014.2330813</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lewis</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Pluye</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Rodriguez</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Grad</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Mining reflective continuing medical education data for family physician learning needs</article-title>
          <source>J Innov Health Inform</source>
          <year>2016</year>
          <month>04</month>
          <day>06</day>
          <volume>23</volume>
          <issue>1</issue>
          <fpage>834</fpage>
          <lpage>440</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://dx.doi.org/10.14236/jhi.v23i1.834"/>
          </comment>
          <pub-id pub-id-type="doi">10.14236/jhi.v23i1.834</pub-id>
          <pub-id pub-id-type="medline">27348489</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Young</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Parvin</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Machine Learning and Data Mining  in Medical Imaging</article-title>
          <source>IEEE J Biomed Health Inform</source>
          <year>2015</year>
          <month>09</month>
          <volume>19</volume>
          <issue>5</issue>
          <fpage>1587</fpage>
          <lpage>8</lpage>
          <pub-id pub-id-type="doi">10.1109/jbhi.2015.2444011</pub-id>
          <pub-id pub-id-type="medline">26574616</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wong</surname>
              <given-names>KKL</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Fong</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ng</surname>
              <given-names>EYK</given-names>
            </name>
          </person-group>
          <article-title>A Special Section on Advanced Computing Techniques for Machine Learning and Data Mining in Medical Informatics</article-title>
          <source>J Med Imaging Hlth Inform</source>
          <year>2016</year>
          <month>08</month>
          <day>01</day>
          <volume>6</volume>
          <issue>4</issue>
          <fpage>1052</fpage>
          <lpage>1055</lpage>
          <pub-id pub-id-type="doi">10.1166/jmihi.2016.1800</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Van Eaton</surname>
              <given-names>EG</given-names>
            </name>
            <name name-style="western">
              <surname>Devlin</surname>
              <given-names>AB</given-names>
            </name>
            <name name-style="western">
              <surname>Devine</surname>
              <given-names>EB</given-names>
            </name>
            <name name-style="western">
              <surname>Flum</surname>
              <given-names>DR</given-names>
            </name>
            <name name-style="western">
              <surname>Tarczy-Hornoch</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Achieving and sustaining automated health data linkages for learning systems: barriers and solutions</article-title>
          <source>EGEMS (Wash DC)</source>
          <year>2014</year>
          <volume>2</volume>
          <issue>2</issue>
          <fpage>1069</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/25848606"/>
          </comment>
          <pub-id pub-id-type="doi">10.13063/2327-9214.1069</pub-id>
          <pub-id pub-id-type="medline">25848606</pub-id>
          <pub-id pub-id-type="pii">egems1069</pub-id>
          <pub-id pub-id-type="pmcid">PMC4371442</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Cai</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Fang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Data Processing and Text Mining Technologies on Electronic Medical Records: A Review</article-title>
          <source>JHE</source>
          <year>2018</year>
          <month>04</month>
          <day>08</day>
          <volume>2018</volume>
          <fpage>1</fpage>
          <lpage>9</lpage>
          <pub-id pub-id-type="doi">10.1155/2018/4302425</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kusy</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Zajdel</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Probabilistic neural network training procedure based on Q(0)-learning algorithm in medical data classification</article-title>
          <source>Appl Intell</source>
          <year>2014</year>
          <month>8</month>
          <day>6</day>
          <volume>41</volume>
          <issue>3</issue>
          <fpage>837</fpage>
          <lpage>854</lpage>
          <pub-id pub-id-type="doi">10.1007/s10489-014-0562-9</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bashir</surname>
              <given-names>MEA</given-names>
            </name>
            <name name-style="western">
              <surname>Shon</surname>
              <given-names>HS</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>DG</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Ryu</surname>
              <given-names>KH</given-names>
            </name>
          </person-group>
          <article-title>Real-Time Automated Cardiac Health Monitoring by Combination of Active Learning and Adaptive Feature Selection</article-title>
          <source>KSII TIIS</source>
          <year>2013</year>
          <month>01</month>
          <day>29</day>
          <volume>7</volume>
          <issue>1</issue>
          <fpage>99</fpage>
          <lpage>118</lpage>
          <pub-id pub-id-type="doi">10.3837/tiis.2013.01.007</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Henriques</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Antunes</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Madeira</surname>
              <given-names>SC</given-names>
            </name>
          </person-group>
          <article-title>Generative modeling of repositories of health records for predictive tasks</article-title>
          <source>Data Min Knowl Disc</source>
          <year>2014</year>
          <month>11</month>
          <day>12</day>
          <volume>29</volume>
          <issue>4</issue>
          <fpage>999</fpage>
          <lpage>1032</lpage>
          <pub-id pub-id-type="doi">10.1007/s10618-014-0385-7</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bos</surname>
              <given-names>JW</given-names>
            </name>
            <name name-style="western">
              <surname>Lauter</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Naehrig</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Private predictive analysis on encrypted medical data</article-title>
          <source>J Biomed Inform</source>
          <year>2014</year>
          <month>08</month>
          <day>14</day>
          <volume>50</volume>
          <fpage>234</fpage>
          <lpage>43</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1532-0464(14)00088-4"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jbi.2014.04.003</pub-id>
          <pub-id pub-id-type="medline">24835616</pub-id>
          <pub-id pub-id-type="pii">S1532-0464(14)00088-4</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A Traffic Motion Object Extraction Algorithm</article-title>
          <source>Int. J. Bifurcation Chaos</source>
          <year>2016</year>
          <month>01</month>
          <day>14</day>
          <volume>25</volume>
          <issue>14</issue>
          <fpage>1540039</fpage>
          <pub-id pub-id-type="doi">10.1142/S0218127415400398</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Zou</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Research on internet information mining based on agent algorithm</article-title>
          <source>Future Generation Computer Systems</source>
          <year>2018</year>
          <month>09</month>
          <volume>86</volume>
          <fpage>598</fpage>
          <lpage>602</lpage>
          <pub-id pub-id-type="doi">10.1016/j.future.2018.04.040</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ke</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Zou</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Evaluation of Developer Efficiency Based on Improved DEA Model</article-title>
          <source>Wireless Pers Commun</source>
          <year>2018</year>
          <month>2</month>
          <day>7</day>
          <volume>102</volume>
          <issue>4</issue>
          <fpage>3843</fpage>
          <lpage>3849</lpage>
          <pub-id pub-id-type="doi">10.1007/s11277-018-5415-0</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Zou</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Sewage information monitoring system based on wireless sensor</article-title>
          <source>Desalination Water Treat</source>
          <year>2018</year>
          <month>07</month>
          <volume>121</volume>
          <fpage>73</fpage>
          <lpage>83</lpage>
          <pub-id pub-id-type="doi">10.5004/dwt.2018.22362</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cour</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Sapp</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Taskar</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Learning from Partial Labels</article-title>
          <source>The Journal of Machine Learning Research</source>
          <year>2011</year>
          <month>07</month>
          <volume>12</volume>
          <fpage>1501</fpage>
          <lpage>1536</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmlr.org/papers/volume12/cour11a/cour11a.pdf"/>
          </comment>
          <pub-id pub-id-type="doi">10.5555/1953048.2021049</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Patel</surname>
              <given-names>Vm</given-names>
            </name>
            <name name-style="western">
              <surname>Chellappa</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Phillips</surname>
              <given-names>Pj</given-names>
            </name>
          </person-group>
          <article-title>Ambiguously Labeled Learning Using Dictionaries</article-title>
          <source>IEEE Trans.Inform.Forensic Secur</source>
          <year>2014</year>
          <month>12</month>
          <volume>9</volume>
          <issue>12</issue>
          <fpage>2076</fpage>
          <lpage>2088</lpage>
          <pub-id pub-id-type="doi">10.1109/TIFS.2014.2359642</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Dietterich</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <person-group person-group-type="editor">
            <name name-style="western">
              <surname>Bartlett</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Pereira</surname>
              <given-names>FCN</given-names>
            </name>
            <name name-style="western">
              <surname>Burges</surname>
              <given-names>CJC</given-names>
            </name>
            <name name-style="western">
              <surname>Bottou</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Weinberger</surname>
              <given-names>KQ</given-names>
            </name>
          </person-group>
          <article-title>A conditional multinomial mixture model for superset label learning</article-title>
          <source>Advances in Neural Information Processing Systems 25</source>
          <year>2012</year>
          <publisher-loc>Cambridge, MA</publisher-loc>
          <publisher-name>MIT Press</publisher-name>
          <fpage>557</fpage>
          <lpage>565</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>M-L</given-names>
            </name>
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>B-B</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>X-Y</given-names>
            </name>
          </person-group>
          <article-title>Partial Label Learning via Feature-Aware Disambiguation</article-title>
          <source>KDD '16: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          <year>2016</year>
          <month>08</month>
          <conf-name>22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</conf-name>
          <conf-date>August, 2016</conf-date>
          <conf-loc>San Francisco, CA</conf-loc>
          <fpage>1335</fpage>
          <lpage>1344</lpage>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
