Maintenance Notice

Due to necessary scheduled maintenance, the JMIR Publications website will be unavailable from Monday, March 11, 2019 at 4:00 PM to 4:30 PM EST. We apologize in advance for any inconvenience this may cause you.

Who will be affected?

Advertisement

Citing this Article

Right click to copy or hit: ctrl+c (cmd+c on mac)

Published on 02.08.16 in Vol 4, No 3 (2016): Jul-Sept

Preprints (earlier versions) of this paper are available at http://preprints.jmir.org/preprint/5490, first published Feb 05, 2016.

This paper is in the following e-collection/theme issue:

    Original Paper

    A Semi-Supervised Learning Approach to Enhance Health Care Community–Based Question Answering: A Case Study in Alcoholism

    1Department of Industrial Engineering and Management Sciences, Northwestern University, Evanston, IL, United States

    2Divison of Health and Biomedical Informatics, Feinberg School of Medicine, Northwestern University, Chicago, IL, United States

    Corresponding Author:

    Papis Wongchaisuwat, MS

    Department of Industrial Engineering and Management Sciences

    Northwestern University

    2145 Sheridan Rd

    Evanston, IL, 60208

    United States

    Phone: 1 847 491 3383

    Fax:1 847 491 8005

    Email: PapisWongchaisuwat2013@u.northwestern.edu


    ABSTRACT

    Background: Community-based question answering (CQA) sites play an important role in addressing health information needs. However, a significant number of posted questions remain unanswered. Automatically answering the posted questions can provide a useful source of information for Web-based health communities.

    Objective: In this study, we developed an algorithm to automatically answer health-related questions based on past questions and answers (QA). We also aimed to understand information embedded within Web-based health content that are good features in identifying valid answers.

    Methods: Our proposed algorithm uses information retrieval techniques to identify candidate answers from resolved QA. To rank these candidates, we implemented a semi-supervised leaning algorithm that extracts the best answer to a question. We assessed this approach on a curated corpus from Yahoo! Answers and compared against a rule-based string similarity baseline.

    Results: On our dataset, the semi-supervised learning algorithm has an accuracy of 86.2%. Unified medical language system–based (health related) features used in the model enhance the algorithm’s performance by proximately 8%. A reasonably high rate of accuracy is obtained given that the data are considerably noisy. Important features distinguishing a valid answer from an invalid answer include text length, number of stop words contained in a test question, a distance between the test question and other questions in the corpus, and a number of overlapping health-related terms between questions.

    Conclusions: Overall, our automated QA system based on historical QA pairs is shown to be effective according to the dataset in this case study. It is developed for general use in the health care domain, which can also be applied to other CQA sites.

    JMIR Med Inform 2016;4(3):e24

    doi:10.2196/medinform.5490

    KEYWORDS



    Introduction

    A study by Pew Internet Project’s research reported that 87% of US adults use the Internet, and 72% of Internet users sought health information over the Internet in the past year [1]. Other studies have also analyzed the modes in which health information is shared and its impact on consumer decision making [2,3]. Although it is known that patients are seeking information that might not be obtained during the course of their regular clinical care and valuable knowledge is publicly available in the Internet, it is not trivial for users to quickly find an accurate answer to specific questions. Consequently, community-based question answering (CQA) sites such as Yahoo! Answers tend to be a potential solution to this challenge. In CQA sites, users post a question and expect the Web-based health community to promptly provide desirable answers. Despite a high volume of users’ participation, a considerable number of questions are left unanswered, and at the same time, other questions that address the same information need are answered elsewhere. This common situation drew our attention to develop an automated system for answering both unsuccessfully answered and newly posted questions.

    Substantial research exists for developing systems that address physicians’ information needs at the point of care. Info buttons and other decision support tools automatically select and retrieve information from knowledge sources at the point of care [4]. Social media platforms involve exchanges of health information among peers at any place and time [5]. The advantages and disadvantages of using a social network to address the information needs compared with a search engine are described in the study by Morris et al [6]. However, limited research has been done in addressing the information needs of patients through automated approaches that synthesize the information shared across Web-based health communities. CQA systems in the health care domain address this issue.

    QA systems are widely studied in both open and other restricted domains. One of the common approaches is to retrieve answers based on past QA, which is also fundamental to our work. Shtok et al [7] extracted an answer from resolved QA pairs obtained from Yahoo! Answers. Specifically, a statistical model was implemented to estimate the probability that the best answer from the past posts can satisfactorily answer a newly posted question. In addition to Shtok et al, Marom et al [8] implemented a predictive model involving a decision graph to generate help desk responses from historical email dialogues between users and help desk operators. Feng et al [9] constructed a system aiming to provide accurate responses to students’ discussion board questions. An important element in these QA systems is identifying the closest (the most similar) matching between a new question and other questions in a corpus. However, this is not a trivial task because both the syntactic and semantic structure of sentences should be considered to achieve an accurate matching. A syntactic tree matching approach was proposed to tackle this problem in CQA [10]. Jeon et al [11] developed a translation-based retrieval model exploiting word relationships to determine similar questions in QA archives. Various string similarity measures were also implemented to directly compute the distance between 2 different strings [12]. A topic clustering approach was introduced to find similar questions among QA pairs [13].

    An important component in QA systems is re-ranking of candidates to identify the best answer. A probabilistic answer selection framework was used to estimate the probability of an answer candidate being correct [14]. Alternatively, supervised learning-based approaches including support vector machine [15,16] and logistic regression [17] are applicable to select (rank) answers. Commonly, collecting a large number of labeled data can be very expensive or even impossible in practice. Wu et al [18] developed a novel unsupervised support vector machine classifier to overcome this problem. Other studies used different classifiers with multiple features for similar problems [19-23].

    Athenikos et al [24] conducted a thorough survey reviewing state of the art in biomedical question answering systems. Morris et al [25] presented a survey study about the behavior of users in question and answer systems. Luo et al [26] developed an algorithm, SimQ, to extract similar consumer health questions based on both syntactic and semantic analysis. Vector-based distance measures were used to compute similarity score among questions. Statistical syntactic parsing and standardized unified medical language system (UMLS) were implemented to construct syntactic and semantic features, respectively. However, to effectively use the information in CQAs, we need to not only retrieve similar questions but also provide and validate potential answers. SimQ was designed to retrieve similar questions from the NetWellness [27], a health information platform that has been maintained by clinician peer reviewers. Questions collected within NetWellness tend to be clean and well structured, whereas CQA websites tend to be noisy. Wong et al has also contributed to automatically answering health-related questions based on previously solved QA pairs [28]. They provide an interactive system where the input questions are precise and short as opposed to accepting CQA questions directly as input.

    In comparison to these systems, our work relies on implementing semi-supervised learning with expectation–maximization (EM) approach [29]. Semi-supervised learning uses both labeled and unlabeled data for training. Given labeled and unlabeled data, EM-based semi-supervised learning first trains an initial model using just the labeled set. This model is then used to estimate the label of each element in the unlabeled set. Next, the model is retrained using both labeled and unlabeled set with the estimated labels from the previous step. The new model is used to refine the estimated labels in the unlabeled set. These steps are iteratively repeated until the algorithm converges or reaches predefined number of iterations. In addition, we used dynamic time warping (DTW) [30] along with the vector-space distance [31] to measure similarity and incorporated biomedical concepts as additional features.

    In summary, our work aims to automatically answer health-related questions based on past QA. We extracted candidate questions based on similarity measure and selected possible answers by using a semi-supervised learning algorithm. Automatically retrieving answers for questions from Web-based health communities should provide the users a potential source of health information.


    Methods

    The system was built as a pipeline that involves 2 phases. The first phase implemented as a rule-based system, consists of (1) Question Extracting, which maps the Yahoo! Answers dataset to a data structure that includes question category, the short version of the question, and the 2 best answers; (2) Answer Extracting, which uses similarity measures to find answers for a question from existing QA pairs. In the second phase of Answer Re-ranking, we implemented supervised and semi-supervised learning models that refined the output of the first phase by screening out invalid answers and ranking the remaining valid answers.

    Figure 1 depicts the system architecture and flow. In training, phase I is applied for each prospective question in the training dataset (with all other questions under a consideration corresponding to all questions in the corpus being different from the current prospective question). For test, the prospective question is a test question, and all other questions are those from the training set. In this case, phase II uses the trained model to rank the candidate answer.

    We first describe the training phase. The rule-based answer extraction phase (phase I) is split into the following 2 steps:

    Figure 1. Overall architecture for training the system.
    View this figure

    Question Extracting

    For this system, we assumed that each question posted on CQA sites has a question title and its description. Once users provided possible answers to the posted question, some responses were assumed to be marked as the best answer either by the question provider or community users. The second and subsequent best answers were chosen among remaining answers based on the number of likes. The raw data collected from CQA sites are unstructured and contain unnecessary text. It is essential to retrieve short and precise questions embedded in the original question title and its description (which can include up to 4-5 question sentences). Instead of using the whole question title and description that are long and verbose, we implemented a rule-based approach to capture these possible short question sentences (subquestions). These subquestions were categorized into different groups based on the words in questions. More specifically, regular expressions based on question words were used to classify subquestions, which yielded different question classes consisting of “yes-no,” “what quantity,” “how frequent,” “when,” “why,” “how,” “where,” “who,” “whose,” “whom,” “what,” and “which,” and “others.” We considered subquestions, instead of full questions and descriptions, in the rest of this paper.

    Answer Extracting

    Given a question, it was divided into subquestions and matched with the question group using the aforementioned rule-based approach. Then, we computed the semantic distance between the prospective question and all other questions from the training sets belonging to the same group. Two distance approaches were used in our work.

    1. DTW-based approach: It is based on a sequence alignment algorithm known as DTW , which uses efficient dynamic programming to calculate a distance between 2 temporal sequences. This allows us to effectively encode the word order without adversely penalizing for missing words (such as in a relative clause). Applying it in our context, a sentence was considered as a sequence of words where the distance between each word was computed by the Levenshtein distance at a character level [32,33]. For any 2 sequences defined as

    Seq1 = < w11, w21,…, wm1> and Seq2 = < w12, w22,…, wn2> where m and n are the lengths of the sequences, Liu et al [30] defined the distance between 2 sequences (in our case, 2 sentences) as in the following Figure 2:

    Figure 2. The distance between two sequences.
    View this figure

    where f (0,0) = 0, f (i, 0) = f (0, j) = ∞, i ∈ (0, m), j (0, n)

    Here, d (wi1, wj2) is the distance between 2 words computed by the Levenshtein measure.

    2. Vector-space based approach: An alternative paradigm is to consider the sentences as a bag of words, represent them as points in a multidimensional space of individual words, and then calculate the distance between them. We implemented a unigram model with tf-idf weights based on the prospective question and other questions in the same category and computed the Euclidean distance measure.

    We further took into account the cases that share similar medical information by multiplying the distances with a given weight parameter. The best value of the weight parameter was selected based on extensive experiments. The MetaMap tool was used to recognize UMLS concepts occurring in questions [34]. If at least 1 word in the UMLS concepts of “organic chemical” and “pharmacologic substance” occurs in both the prospective question and a training question, we reduce the distance to account for the additional semantic similarity. These UMLS concepts are specifically selected as we want to provide more weight to answers that mention a treatment approach under the intuitive assumption that most CQA users aim to seek informative advice for their illness. The set of semantic types can be expanded to capture broader concepts if different domains are considered.

    The QA pairs in the training set corresponding to the smallest and the second smallest distance were extracted. Thus, we finally obtained a list of candidate answers, that is, the answers referring to smallest and second smallest questions, for each prospective question. These answers were used as the output of the baseline rule -based system. This was repeated for each question in the training set, that is, the prospective question corresponds to each question in the training set. At the end of this phase, we had triplets (Qp, Qt, At) over all questions Qp. Note that At is an answer to question Qt with Qt ≠ Qp, and each Qp yielded several such triplets.

    The machine-learning phase of answer re-ranking (phase II) is described next. The goal of this phase is to rank candidate answers from the previous step and select the best answer among them. Each triple (Qp, Qt, At) is aimed to be assigned as “valid” if At is a valid answer to Qp, or “invalid” otherwise. We describe how the model was trained in this section while detailed explanations (eg, number of labeled and unlabeled triplets) are provided in the section, “Results.” We first selected a small random subset of triplets and labeled them manually (there were too many to label all of them in this way). Both supervised and semi-supervised learning EM models were developed to predict the answerability of newly posted question and rank candidate answers. Specifically, the semi-supervised learning model was trained on labeled and unlabeled triplets. According to the semi-supervised learning model, we first trained a supervised learning algorithm including Neural Networks with the entropy objective function (NNET), Neural Networks with the L2-norm or least squares objective function (NNET_L2), support vector machine (SVM), and logistic regression based on manually labeling outputs from the aforementioned rule-based answer extraction phase. The trained model was used to classify the unlabeled part of the outputs of phase I, and then, the classifier was retrained based on the original labeled data and a randomly selected subset of unlabeled data using the estimated labels from the previous iteration. These steps were iteratively repeated to achieve a final estimated label. The supervised approach, on the other hand, only ran a classifier on the labeled subset and finished. A 10-fold cross validation was implemented in both semi-supervised and supervised approaches. Specifically, all labeled observations were partitioned into 10 parts where 1 part was set aside as a test set. The model was fitted based on the remaining 9 parts of the labeled observations (plus the entire unlabeled part for the semi-supervised learning approach). The parameters of the semi-supervised model were obtained by using the EM algorithm previously described. The fitted model was then used to predict the responses in the part that we set aside as the test set. These steps were repeated by selecting different part to set aside as the test set. All features used in the models are illustrated based on the following example as summarized in Table 1.

    Table 1. List of features used in the model.
    View this table

    Example of a Triple (Qp, Qt, At)

    Prospective Question

    Anxiety medication for drug/alcohol addiction?

    Training Question

    Is chlordiazepoxide/librium a good medication for alcohol withdrawal and the associated anxiety?

    Training Answer

    Chlordiazepoxide has been the standard drug used for rapid alcohol detox for decades and has stood the test of time. The key word is rapid the drug should really only be given for around a week. Starting at 100 mg on day 1 and reducing the dose every day to reach zero on day 8. In my experience, it deals well with both the physical and mental symptoms of withdrawal. Looking ahead, he will still need an alternative management for his anxiety to replace the alcohol. Therapy may help, possibly in a group setting

    Sets SP, ST, and SA are sets of terms corresponding to UMLS concepts occurred in Qp, Qt, and At, respectively. General features are taken from previous work [7], while we introduce UMLS-based features into the model. Features 9 and 10 are calculated by counting the number of words contained in both sets. To obtain features 12 and 13, we find the elements that are in only 1 of the 2 sets.

    Table 2 depicts examples of annotations in the corpus. The inter-rater agreement for random instances (10% of total) assigned to 2 independent reviewers is very good (95% CI of kappa from .69 to .93). The procedure to identify an answer to a newly posted question is illustrated in Figure 3 after the usual split of the corpus in train and test.

    Table 2. Corpus annotation examples.
    View this table
    Figure 3. Process flow of the testing step.
    View this figure

    The following evaluation metrics are used to test the overall performance of our algorithm.

    1. Question-based evaluation metrics

    - For this paper, we define “overall accuracy” as ratio of the number of questions with at least 1 “correct” answer divided by total number of questions in the test set. A test question is labeled as “correct” if our algorithm predicts at least 1 valid triple correctly. For the case that there is no valid answer in the question from the gold standard, we label it as “correct” if our algorithm predicts corresponding triplets as invalid.

    - The mean reciprocal rank (MRR) with test questions Q is defined as Figure 4.

    where ranki is the position of a valid instance in manually sorted probabilities from the model. If there are more than 1 valid instances in any question, minimum value of ranki is used.

    2. Triple-based evaluation metrics

    Precision, recall, and the F1-score can be used as standard measures for binary classification. We do not measure accuracy and receiver operating characteristic curves because the dataset is heavily imbalanced.

    Figure 4. The Mean Reciprocal Rank (MRR) with a set of test questions Q.
    View this figure

    Results

    To test the algorithm, we obtained a total of 4216 alcoholism-related QA threads from Yahoo! Answers. The sample outputs from our algorithm are shown in Figure 5, which indicates how our system could potentially be used by Web-based advice seekers. To extract initial candidate answers in the rule-based answer extraction, our algorithm returns 8 instances for each prospective question (obtained from 2 different similarity measures where we extract at least 2 closest questions for each measure with 2 answers for each question). An example of output reported from the rule-based answer extraction is depicted in Figure 6.

    Figure 5. System output.
    View this figure
    Figure 6. An example result returned from the algorithm to determine candidate answers.
    View this figure

    A randomly selected set of 220 threads were manually annotated and used as labeled questions. Overall, 119 of 220 questions, or 54.1%, have valid answers among those extracted in the rule-based answer extraction phase. After retrieving candidate answers, we further aim to re-rank them and select the best answer (if there is a valid answer). Note that each question corresponds to several candidate answers and thus multiple triplets (Qp, Qt, At). If at least 1 triplet is labeled as “valid,” the corresponding question is also labeled as “valid.” Specifically, the semi-supervised learning model (EM) was trained on 1553 labeled triplets (corresponding to 220 manually labeled questions) and 10,000 unlabeled triplets. In the training data of 1553 labeled triplets, 297 triplets were manually labeled as “valid” and 1256 as “invalid.” The typical 10-fold cross validation was implemented to validate the model.

    We included all features listed in Table 1 in the models. To indicate a significance of each feature, we analyzed the feature set by using information gain. The information gain is based on the entropy function, which is closely related with the objective function of the neural network NNET and logistic regression classifiers. The most influential features are the number of stop words contained in Qp, the text length, the distance of (Qp, Qt), and the number of overlapping UMLS words between Qp and Qt, that is, in SP and ST. All information gains for these significant features are listed in Table 3.

    Table 3. Information gain score of 5 significant features
    View this table

    The best model was selected by varying the cutoff probability of being valid or invalid to obtain the maximum F1-score. We selected NNET, NNET_L2, SVM, and logistic regression approaches to train the model on a subset. For the SVM classifier, the probability was obtained by fitting a logistic distribution using maximum likelihood to the decision values provided by SVM.

    The semi-supervised learning (EM) algorithm with 1 iteration trained with NNET_L2 gave the best performance for MRR and F1-score with a reasonable value of overall accuracy, whereas NNET performs best for overall accuracy, as listed in Table 4. Each value in the table is the average across 100 different runs based on different random numbers in the algorithms and the test/train splits (details provided in the following section). In Table 4, the numbers in bold represent the best value among different models and classifiers for each evaluation metric. The confusion matrices for 1 iteration of EM trained with 4 different classification models are provided in Figure 7.

    Figure 7. The confusion matrices for 1 iteration of EM trained with NNET, NNET_L2, SVM, and LOG.
    View this figure
    Table 4. Evaluation metrics.
    View this table

    We performed 2 types of statistical hypothesis tests (t-tests) at the .05 level (95% CI) to determine if 2 sets of evaluation metrics among the F1-score, overall accuracy, and MRR, obtained from different settings are significantly different from each other. First, randomness occurs within an algorithm such as the randomness in the stochastic gradient approach. Second, we consider randomness of assigning the test set, that is, the training and test sets in 10-fold cross validation are randomly assigned. We performed both types of the hypothesis tests for all possible comparisons including the model implemented (pure classification vs semi-supervised), and among the 4 different classifiers based on the numbers reported in Table 4. Overall, the semi-supervised learning model is statistically significantly better than the corresponding supervised version for all evaluation metrics. This conclusion holds for both tests. Comparing between 1 and 10 EM iterations, the evaluation metrics are not statistically different from each other. This implies that the model parameters tuned by the EM algorithm are very close to the optimal values within 1 iteration.

    We are also interested in understanding whether UMLS-based features (feature 9-13 listed in Table 1) play a role in predicting the validity of a candidate answer. Hence, we trained another model, which excludes all UMLS-based features, and compared the results (obtained from 1 iterations of EM trained with NNET_L2) with those from the original model as illustrated in Figure 8. The statistical tests at the .05 level showed significantly difference between the 2 models (with vs without UMLS-based features) for the 3 evaluation metrics. With UMLS-based features, the model gave a better performance, which is consistent across all evaluation metrics. This implies that these features played a role in distinguishing between valid and invalid answers.

    Figure 8. Performance between the original and adjusted model to test significance of UMLS-based features (health features).
    View this figure

    Discussion

    In this paper, we developed an automated QA system by using previously resolved QA pairs from a CQA site and evaluated it. Although we used Yahoo! Answers as a data source, our algorithm can be adapted and applied to other CQA sites, in particular those related to health care where UMLS applies. Among different models and classifiers experimented, EM semi-supervised learning is better than pure supervised learning, and 1 iteration of EM generally performs better than other models. Specifically, 1 iteration of EM with NNET gives the best performance in term of accuracy. NNET_L2 with 1 iteration of EM performs best in terms of the MRR and F1-scores. The NNET_L2 with 1 EM iteration is recommended to be used based on the case study data. Overall, the best model achieves an 86.2% accuracy and a 0.4 F1-score, which are significant given that the problem is challenging and the data are imperfect. Internet users typically provide responses in an ill-formed fashion. Our data also consist of a significant number of complex questions, for example, a user discusses about his or her situation in 10 to 20 sentences and then asks whether he or she is an alcoholic. Moreover, some questions are very detailed; for example, the percentage of alcohol resulting from a given combination of chemical components. There is a trade-off between precision and recall. Some of these values listed in Table 4 are small as we aim to find a good balance between the 2 values. We intentionally maximize the F1-score, which is a representative of both values. Precision and recall are reported in Table 4 for completeness. A comparison between the rule-based approach in the first phrase and the semi-supervised learning model in the second phrase reveals a significant improvement. The semi-supervised approach improves the accuracy of the model by 30% (approximately from 55% to 86%).

    Comparing with Luo et al [26] who retrieved the similar questions based on the distance measure, we relied on this idea with different approaches. To compute the similarity score between questions, we used the DTW measure instead of relying on the vector-based distance measure. Luo et al used matching questions with information in data sources that are written and reviewed by experts; we strictly use only data from Yahoo! Answers, which are very noisy. For this reason, the syntactic features proposed by Luo et al might not be useful in our model. Unfortunately, not all libraries used in Luo et al’s implementation are publicly available, and thus, direct comparison of the accuracy is not possible.

    Shtok et al [7] used resolved QA pairs to reduce the rate of unanswered questions in Yahoo! Answers. The experiment in Shtok et al was also tested with health-related questions, and the accuracy as measured by the F1-score was 0.32. Our method, which trained a semi-supervised learning model with a smaller amount of manually labeled data compared with a supervised learning model used in [7], resulted in 0.4 F1-score. A better performance might be because of several reasons. First, we categorized questions in a corpus into different groups based on question keywords. Instead of computing the distance between a test question and all other questions in the corpus, categorizing questions reduces the scope of questions an algorithm needs to search. As we categorize collected questions into different groups based on question keywords, latent topics and “wh” question matching features used in Shtok’s study are not valuable in our context. Second, our algorithm also used multiple features related to the UMLS medical topics to enhance the model’s performance when applied within the health domain where the Shtok’s system was designed for a more general usage. Although Shtok et al. relied on cosine distance, the Euclidean distance performed better in our evaluation. Among distance measures used in our work, more valid answers can be correctly identified with the DTW-based approach than the vector similarity measure, which can be observed when manually annotating the output from the rule-based answer extraction. In addition, our algorithm extracted multiple candidate answers retrieved from 2 closest QA pairs for each distance metric and the 2 best answers for each question. In each QA pair, both the best and the second best answer were extracted compared with Shtok et al where only the best answer was extracted. Finally, we implemented semi-supervised learning to gain benefits from unlabeled data, whereas Shtok et al only relied on a supervised learning model in the re-ranking phase.

    Using a semi-supervised learning model that leverages unlabeled data is reasonable against other traditional supervised learning models because obtaining labeled data is very expensive and time consuming in practice. As the features of the machine-learning algorithm are not specific to alcoholism, our system should be applicable for other related topics. On the other hand, it would be possible to increase the accuracy for “alcoholism” if we use specific features such as concepts related to alcoholism.

    In summary, the main novelty and advantages of our work against other works include the DTW-based distance approach, UMLS-based features, the semi-supervised learning algorithm, and the dataset used in the study. We introduce novel distance measures, the DTW-based approach that performs better than the typical vector-space distance method. UMLS-based features are included to enhance the model applied in the health care domain in addition to the general features in the study by the study by Shtok et al [7]. Our system is trained and tested only on the Web-based information without any additional sources. Further, obtaining the annotation from Web-based data can be very difficult and time consuming. This stresses the significance of using semi-supervised learning rather than a typical supervised learning algorithm.

    For the machine-learning component, the distance between a test question and other questions in the training dataset is important in distinguishing valid and invalid answers. The closer the distance is, the higher the chance of the corresponding answer being valid. Matching UMLS terms, which imply a closer similarity between questions, plays a role in determining the validity of the answer. Although UMLS-based features show lower information gain, the model with these features included is significantly better across all evaluation metrics. The overall accuracy is improved by 8% when these features are included.

    Information gain shows that number of stop words contained in a test question and the underling text length are the best indicators for differentiating between valid and invalid answers. We note that the number of content-rich words, represented as text length minus the number of stop words, is also taken indirectly into account by these 2 features. We fitted the model without the number of stop words feature compared with the full model. Although these 2 models are not statistically different, we include the number of stop words feature in the model as previously done by Shtok et al [7].

    Limitations and Future Work

    The main limitation of our work is the lack of assessment of the model’s generalizability. Although our algorithm is generic and does not include any features that are specific to the topic of alcoholism, we have not validated it in different domains as we do not have available data. Approximately 30% (obtained from a preliminary observation) of all questions cannot be answered based on existing answers; some of these questions also require additional resources that are more technical and reliable, such as medical textbooks, journals, and guidelines.

    Conclusions

    The question-answering system developed in this work achieves reasonably good performance in extracting and ranking answers to questions posted in CQA sites. Our work is a promising approach for automatically answering alcoholism-related questions obtained from CQA sites based only on past QA that is used as a case study. In addition, our system can potentially be applied to other health care domain questions asked by Web-based health care communities. The system and the gold standard corpus are available on GitHub [35].

    Acknowledgments

    The authors are grateful to Dr. Jina Huh from Michigan State University for providing the Yahoo! Answers dataset and Dr. Kalpana Raja from Northwestern University for helping with UMLS. This work was partly supported by funding from the National Library of Medicine grant R00LM011389.

    Conflicts of Interest

    None declared.

    References

    1. Fox S, Duggan M. Pew Research Center. 2013. Health Online 2013   URL: http://www.pewinternet.org/2013/01/15/health-online-2013/ [WebCite Cache]
    2. Lau AY, Coiera E. Impact of web searching and social feedback on consumer decision making: a prospective online experiment. J Med Internet Res 2008;10(1):e2 [FREE Full text] [CrossRef] [Medline]
    3. Nath C, Huh J, Adupa A, Jonnalagadda S. Website sharing in online health communities: A descriptive analysis. J Med Internet Res 2016;18(1):e11 [FREE Full text] [CrossRef] [Medline]
    4. Cimino J, Yu H, Del FG. Infobuttons and point of care access to knowledge. Clinical Decision Support 2007:345-372.
    5. Burton S, Tanner K, Giraud-Carrier C. Leveraging social networks for anytime-anyplace health information. Network modeling analysis in health informatics and bioinformatics 2012;1(4):173-181.
    6. Morris M, Jaime T, Panovich K. A comparison of information seeking using search engines and social networks. 2010 Jan 01 Presented at: Proceedings of 4th international AAAI conference on weblogssocial media; May 23-26, 2010; Washington, DC p. 291-294.
    7. Shtok A, Dror G, Maarek Y, Szpektor I. Learning from the past: answering new questions with past answers. 2012 Apr 16 Presented at: Proceedings of the 21st international conference on world wide web; April 16-20, 2012; Lyon, France p. 759-768. [CrossRef]
    8. Marom Y, Zukerman I. A predictive approach to help-desk response generation. 2007 Jan 06 Presented at: 20th international joint conference on artificial intelligence; January 6-12, 2007; Hyderabad, India p. 1665-1670.
    9. Feng D, Shaw E, Kim J, Hovy E. An intelligent discussion-bot for answering student queries in threaded discussions. 2006 Jan 29 Presented at: Proceedings of the 11th international conference on intelligent user interfaces; January 29 - February 01, 2006; Sydney, Australia p. 171-177.
    10. Wang K, Ming Z, Chua T. A Syntactic tree matching approach to finding similar questions in community-based QA services. 2009 Jul 19 Presented at: Proceedings 32nd annual international ACM SIGIR conference on research and development in information retrieval; July 19-23, 2009; Boston, MA p. 187-194.
    11. Jeon J, Croft W, Lee J. Finding similar questions in large question and answer archives. 2005 Oct 31 Presented at: Proceedings of the 14th ACM international conference on information and knowledge management; October 31 - November 05, 2005; Bremen, Germany p. 84-90.
    12. Bernhard D, Gurevych I. Answering learners' questions by retrieving question paraphrases from social Q&A sites. 2008 Jun 19 Presented at: Proceedings of the third workshop on innovative use of NLP for building educational applications; June 19, 2008; Columbus, OH p. 44-52.
    13. Zhang W, Liu T, Yang Y, Cao L, Zhang Y, Ji R. A topic clustering approach to finding similar questions from large question and answer archives. PLoS One 2014;9(3):e71511 [FREE Full text] [CrossRef] [Medline]
    14. Ko J, Si L, Nyberg E. A probabilistic framework for answer selection in question answering. 2007 Apr 22 Presented at: Proceedings of human language technology conference of the North American chapter of the association of computational linguistics; April 22-27, 2007; Rochester, NY p. 524-531.
    15. Moschitti A, Quarteroni S. Linguistic kernels for answer re-ranking in question answering systems. Information Processing & Management 2011;47(6):825-842. [CrossRef]
    16. Suzuki J, Sasaki Y, Maeda E. SVM answer selection for open-domain question answering. 2002 Aug 26 Presented at: Proceedings of the 19th international conference on computational linguistics; August 26-30, 2002; Taipei, Taiwan p. 1-7.
    17. Blooma M, Chua A, Goh D. Selection of the best answer in CQA services. 2010 Apr 12 Presented at: Proceedings of the 7th international conference on information technology; April 12-14, 2010; Las Vegas, NV p. 534-539.
    18. Wu Y, Zhang R, Hu X, Kashioka H. Learning unsupervised SVM classifier for answer selection in web question answering. 2007 Jun 28 Presented at: Proceedings of the joint conference on empirical methods in natural language processing and computational natural language learning; June 28-30, 2007; Prague, Czech Republic p. 33-41.
    19. Toba H, Ming Z, Adriani M, Chua T. Discovering high quality answers in community question answering archives using a hierarchy of classifiers. Inform Sciences 2014:101-115.
    20. Shah C, Pomerantz J. Evaluating and predicting answer quality in community QA. 2010 Jul 19 Presented at: SIGIR : Proceedings of the 33rd annual international ACM SIGIR conference on research development in information retrieval; July 19-23, 2010; Switzerland p. 411-418.
    21. Arai K, Handayani A. Predicting quality of answer in collaborative Q/A community. International journal of advanced research in artificial intelligence 2013;2(3):21-25.
    22. Shah C, Kitzie V, Choi E. Questioning the question - addressing the answerability of questions in community question-answering. 2014 Jan 06 Presented at: 47th Hawaii international conference on system sciences; January 6-9, 2014; Waikoloa, HI p. 1386-1395.
    23. Tian Q, Zhang P, Li B. Towards predicting the best answers in community-based question-answering services. 2013 Jul 08 Presented at: Proceedings of the 7th international conference on weblogs and social media; July 8-11, 2013; Cambridge, MA p. 725-728.
    24. Athenikos SJ, Han H. Biomedical question answering: A survey. Comput Methods Programs Biomed 2010 Jul;99(1):1-24. [CrossRef] [Medline]
    25. Morris M, Teevan J, Panovich K. What do people ask their social networks, and why?: a survey study of status message q&a behavior. 2010 Apr 10 Presented at: Proceedings of the SIGCHI conference on human factors in computing systems; April 10-15, 2010; Atlanta, GA p. 1739-1748.
    26. Luo J, Zhang G, Wentz S, Cui L, Xu R. SimQ: Real-time retrieval of similar consumer health questions. J Med Internet Res 2015;17(2):e43 [FREE Full text] [CrossRef] [Medline]
    27. Morris TA, Guard JR, Marine SA, Schick L, Haag D, Tsipis G, et al. Approaching equity in consumer health information delivery: NetWellness. J Am Med Inform Assoc 1997;4(1):6-13 [FREE Full text] [Medline]
    28. Wong W, Thangarajah J, Padgham L. Contextual question answering for the health domain. J Am Soc Inf Sci Tec Nov 2012;63(11):2313-2327.
    29. Nigam K, McCallum A, Thrun S, Mitchell T. Text classification from labeled and unlabeled documents using EM. Machine learning 2000;39(2-3):103-134.
    30. Liu XY, Zhou YM, Zheng RS. Sentence similarity based on dynamic time warping. 2007 Sep 17 Presented at: International conference on semantic computing, proceedings; September 17-19, 2007; Irvine, CA p. 250-256.
    31. Zobel J, Moffat A. Exploring the similarity space. ACM SIGIR Forum 1998;32(1):18-34.
    32. Levenshtein V. Binary codes capable of correcting deletions, insertions, and reversals. Soviet Physics Doklady 1966;10(8):707-710.
    33. Wagner R, Fischer M. The string-to-string correction problem. Journal of ACM 1974;21(1):168-173.
    34. Aronson A. Effective mapping of biomedical text to the UMLS Metathesaurus: the MetaMap program. 2001 Presented at: Proceedings / AMIA annual symposium; 2001; Washington, DC p. 17-21.
    35. Wongchaisuwat P, Klabjan D, Jonnalagadda S. GitHub. 2015. A semi-supervised learning approach to enhance community-based question answering: Code and dataset   URL: https://github.com/papisw/Health-QA [accessed 2016-07-06] [WebCite Cache]


    Abbreviations

    CQA: community-based question answering
    DTW: Dynamic time warping
    EM: expectation maximization
    MRR: mean reciprocal rank
    NNET: neural networks
    QA: questions and answers
    SVM: support vector machine
    UMLS: Unified medical language system


    Edited by G Eysenbach; submitted 05.02.16; peer-reviewed by C Giraud-Carrier, H Zhai; comments to author 22.02.16; revised version received 05.05.16; accepted 24.06.16; published 02.08.16

    ©Papis Wongchaisuwat, Diego Klabjan, Siddhartha Reddy Jonnalagadda. Originally published in JMIR Medical Informatics (http://medinform.jmir.org), 02.08.2016.

    This is an open-access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on http://medinform.jmir.org/, as well as this copyright and license information must be included.