<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Med Inform</journal-id><journal-id journal-id-type="publisher-id">medinform</journal-id><journal-id journal-id-type="index">7</journal-id><journal-title>JMIR Medical Informatics</journal-title><abbrev-journal-title>JMIR Med Inform</abbrev-journal-title><issn pub-type="epub">2291-9694</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v14i1e73725</article-id><article-id pub-id-type="doi">10.2196/73725</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Iterative Large Language Model&#x2013;Guided Sampling and Expert-Annotated Benchmark Corpus for Harmful Suicide Content Detection: Development and Validation Study</article-title></title-group><contrib-group><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Park</surname><given-names>Kyumin</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Baik</surname><given-names>Myung Jae</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Hwang</surname><given-names>YeongJun</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Shin</surname><given-names>Yen</given-names></name><degrees>BS</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Lee</surname><given-names>HoJae</given-names></name><degrees>BS</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Lee</surname><given-names>Ruda</given-names></name><degrees>MA</degrees><xref ref-type="aff" rid="aff5">5</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Lee</surname><given-names>Sang Min</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Sun</surname><given-names>Je Young Hannah</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Lee</surname><given-names>Ah Rah</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Yoon</surname><given-names>Si Yeun</given-names></name><degrees>MA</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Lee</surname><given-names>Dong-ho</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Moon</surname><given-names>Jihyung</given-names></name><degrees>MS</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Bak</surname><given-names>JinYeong</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Cho</surname><given-names>Kyunghyun</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff6">6</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Paik</surname><given-names>Jong-Woo</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Park</surname><given-names>Sungjoon</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib></contrib-group><aff id="aff1"><institution>SoftlyAI</institution><addr-line>Seoul</addr-line><country>Republic of Korea</country></aff><aff id="aff2"><institution>Department of Psychiatry, Kyung Hee University College of Medicine</institution><addr-line>Seoul</addr-line><country>Republic of Korea</country></aff><aff id="aff3"><institution>Department of Artificial Intelligence, Sungkyunkwan University</institution><addr-line>Office 27306, Engineering Building 2, 2066 Seobu-ro Jangan-gu</addr-line><addr-line>Suwon-si</addr-line><addr-line>Gyeonggi-do</addr-line><country>Republic of Korea</country></aff><aff id="aff4"><institution>KAIST</institution><addr-line>Daejeon</addr-line><country>Republic of Korea</country></aff><aff id="aff5"><institution>Department of Psychology, University of Pennsylvania</institution><addr-line>Philadelphia</addr-line><addr-line>PA</addr-line><country>United States</country></aff><aff id="aff6"><institution>Department of Computer Science, New York University</institution><addr-line>New York</addr-line><addr-line>NY</addr-line><country>United States</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Benis</surname><given-names>Arriel</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Gupta</surname><given-names>Ankit</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Wreyford</surname><given-names>Leon</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Chowdhury</surname><given-names>Shaika</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Banerjee</surname><given-names>Somnath</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to JinYeong Bak, PhD, Department of Artificial Intelligence, Sungkyunkwan University, Office 27306, Engineering Building 2, 2066 Seobu-ro Jangan-gu, Suwon-si, Gyeonggi-do, 16419, Republic of Korea, +82 31 290 7104; <email>jy.bak@skku.edu</email></corresp><fn fn-type="equal" id="equal-contrib1"><label>*</label><p>these authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2026</year></pub-date><pub-date pub-type="epub"><day>5</day><month>2</month><year>2026</year></pub-date><volume>14</volume><elocation-id>e73725</elocation-id><history><date date-type="received"><day>12</day><month>03</month><year>2025</year></date><date date-type="accepted"><day>16</day><month>11</month><year>2025</year></date></history><copyright-statement>&#x00A9; Kyumin Park, Myung Jae Baik, YeongJun Hwang, Yen Shin, HoJae Lee, Ruda Lee, Sang Min Lee, Je Young Hannah Sun, Ah Rah Lee, Si Yeun Yoon, Dong-ho Lee, Jihyung Moon, JinYeong Bak, Kyunghyun Cho, Jong-Woo Paik, Sungjoon Park. Originally published in JMIR Medical Informatics (<ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org">https://medinform.jmir.org</ext-link>), 5.2.2026. </copyright-statement><copyright-year>2026</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://medinform.jmir.org/">https://medinform.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://medinform.jmir.org/2026/1/e73725"/><abstract><sec><title>Background</title><p>Harmful suicide content on the internet poses significant risks, as it can induce suicidal thoughts and behaviors, particularly among vulnerable populations. Despite global efforts, existing moderation approaches remain insufficient, especially in high-risk regions such as South Korea, which has the highest suicide rate among Organisation for Economic Co-operation and Development countries. Previous research has primarily focused on assessing the suicide risk of the authors who wrote the content rather than the harmfulness of content itself which potentially leads the readers to self-harm or suicide, highlighting a critical gap in current approaches. Our study addresses this gap by shifting the focus from assessing the suicide risk of content authors to evaluating the harmfulness of the content itself and its potential to induce suicide risk among readers.</p></sec><sec><title>Objective</title><p>This study aimed to develop an artificial intelligence (AI)&#x2013;driven system for classifying online suicide-related content into 5 levels: illegal, harmful, potentially harmful, harmless, and non&#x2013;suicide-related. In addition, the researchers construct a multimodal benchmark dataset with expert annotations to improve content moderation and assist AI models in detecting and regulating harmful content more effectively.</p></sec><sec sec-type="methods"><title>Methods</title><p>We collected 43,244 user-generated posts from various online sources, including social media, question and answer (Q&#x0026;A) platforms, and online communities. To reduce the workload on human annotators, GPT-4 was used for preannotation, filtering, and categorizing content before manual review by medical professionals. A task description document ensured consistency in classification. Ultimately, a benchmark dataset of 452 manually labeled entries was developed, including both Korean and English versions, to support AI-based moderation. The study also evaluated zero-shot and few-shot learning to determine the best AI approach for detecting harmful content.</p></sec><sec sec-type="results"><title>Results</title><p>The multimodal benchmark dataset showed that GPT-4 achieved the highest <italic>F</italic><sub>1</sub>-scores (66.46 for illegal and 77.09 for harmful content detection). Image descriptions improved classification accuracy, while directly using raw images slightly decreased performance. Few-shot learning significantly enhanced detection, demonstrating that small but high-quality datasets could improve AI-driven moderation. However, translation challenges were observed, particularly in suicide-related slang and abbreviations, which were sometimes inaccurately conveyed in the English benchmark.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>This study provides a high-quality benchmark for AI-based suicide content detection, proving that large language models can effectively assist in content moderation while reducing the burden on human moderators. Future work will focus on enhancing real-time detection and improving the handling of subtle or disguised harmful content.</p></sec></abstract><kwd-group><kwd>artificial intelligence</kwd><kwd>dataset</kwd><kwd>suicide</kwd><kwd>suicide-related content</kwd><kwd>large language models</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p><italic>Harmful suicide content</italic> on the internet poses a significant risk because it can induce suicidal thoughts in readers, potentially leading to self-harm or suicide [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]. The harmful suicide content includes materials that encourage or glorify suicide [<xref ref-type="bibr" rid="ref3">3</xref>], making it appear as an attractive option and sharing suicide methods or instilling suicide knowledge in individuals with suicidal thoughts, thereby increasing the likelihood of actual suicide attempts [<xref ref-type="bibr" rid="ref4">4</xref>]. In some cases, exposure to such harmful suicide content has led middle school students to commit suicide [<xref ref-type="bibr" rid="ref5">5</xref>]. An analysis of adolescent suicide cases reveals that this age group, particularly female adolescents, is more vulnerable to the influence of triggering content [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref7">7</xref>]. Therefore, it is crucial to moderate such harmful suicide content before it spreads extensively.</p><p>Therefore, efforts to moderate harmful suicide content are intensifying. In the United States, initiatives focus on raising public awareness and safe content distribution, aligning with the World Health Organization guidelines [<xref ref-type="bibr" rid="ref8">8</xref>]. Meanwhile, in 2022, the United Kingdom passed a law that makes such content illegal, emphasizing its serious commitment to addressing this issue [<xref ref-type="bibr" rid="ref9">9</xref>]. In the Republic of Korea, which has the highest suicide rates among Organisation for Economic Co-operation and Development countries [<xref ref-type="bibr" rid="ref10">10</xref>], the National Assembly of the Republic of Korea amended the Suicide Prevention Act, and the government has declared the dissemination of such content as illegal since 2019 [<xref ref-type="bibr" rid="ref2">2</xref>]. Despite the increasing spread of harmful suicide content, its moderation is currently handled by only a single official and fewer than a thousand volunteers [<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref12">12</xref>]. Considering the extensive use of social media in Korea [<xref ref-type="bibr" rid="ref13">13</xref>], monitoring the large amounts of content is extremely challenging. In addition, moderating suicide content often leads to a high level of mental stress, hindering their ability to consistently and effectively monitor such content. Therefore, the need for an automatic harmful suicide content moderation system is urgent. The system can efficiently manage a growing volume of the content and ease the burden on human moderators.</p><p>However, previous research on online suicide content primarily focused on predicting the suicide risk of the authors who wrote the content. Zirikly et al [<xref ref-type="bibr" rid="ref14">14</xref>] classified the suicide risk of authors based on content posted online (Reddit) into 4 levels. Similarly, Milne et al [<xref ref-type="bibr" rid="ref15">15</xref>] and Yates et al [<xref ref-type="bibr" rid="ref16">16</xref>] conducted research to predict the suicide and self-harm risks of online content authors. Subsequently, Yang et al [<xref ref-type="bibr" rid="ref17">17</xref>] and Sawhney et al [<xref ref-type="bibr" rid="ref18">18</xref>] used weakly supervised learning to enhance detection performance or collaborated with clinicians. Furthermore, Rawat et al [<xref ref-type="bibr" rid="ref19">19</xref>] and Sawhney et al [<xref ref-type="bibr" rid="ref20">20</xref>] performed tasks to detect suicide ideation and suicide events. All these studies focused on detecting the suicide risk of the author presented in the posts; they did not consider the potential propagation of suicide risk to readers by increasing their susceptibility to self-harm or suicidal behaviors. Other studies focused on understanding the negative effects of suicide content [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref21">21</xref>], or identifying the individuals who are most affected by the content [<xref ref-type="bibr" rid="ref22">22</xref>-<xref ref-type="bibr" rid="ref25">25</xref>]. Studies among Chinese adolescents have shown significant correlations between digital media usage and suicide or self-harm [<xref ref-type="bibr" rid="ref23">23</xref>], and a meaningful relationship between suicide cases among Korean youths and searches related to suicide or self-harm [<xref ref-type="bibr" rid="ref25">25</xref>]. In addition, three-quarters of young adults who have attempted suicide have reported using the internet for suicide or self-harm&#x2013;related reasons [<xref ref-type="bibr" rid="ref26">26</xref>], highlighting the risk posed by information that can induce suicide or self-harm.</p><p>Therefore, we introduce a <italic>harmful suicide content detection</italic> task that determines the level of harmfulness of the harmful suicide content to viewers. We then develop a multimodal <italic>harmful suicide content detection benchmark</italic> and a <italic>task description document</italic>. This document contains detailed instructions for annotators on how to assess the harmfulness of suicide-related content, which could also be useful for building instructions for large language models (LLMs). The benchmark and the document are developed by medical professionals, because such content might involve harmful visual language information that requires the judgment of the professionals (eg, self-harm photographs, or name of illegal drugs that can be used for suicide). Because labeling harmful suicide content causes mental stress, we focus on creating a small yet high-quality dataset. Furthermore, we demonstrate various methods using LLMs that can be effectively performed with few-shot examples.</p><p>Our contributions are as follows: (1) We propose a harmful suicide content detection task that classifies multimodal harmful suicide content as illegal, harmful, potentially harmful, harmless, or non&#x2013;suicide-related. (2) We build a harmful suicide content detection benchmark that is multimodal and a Korean benchmark of 452 curated user-generated contents with corresponding medical expert annotations and a detailed task description document including the task details and instructions for annotators. (3) We create an English harmful suicide content detection benchmark translated from the Korean benchmark using a model (GPT-4) and analyze the quality and issues of translating suicide contents. (5) We demonstrate strategies to use LLMs to detect harmful suicide content by using the task description document and a small yet high-quality harmful suicide content detection benchmark. We test various closed and open-sourced LLMs using the machine-translated English benchmark. We observe that GPT-4 achieves <italic>F</italic><sub>1</sub>-scores of 66.46 and 77.09 in detecting illegal and harmful suicide content, respectively.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Study Design</title><p>In this section, we describe the design process for the harmful suicide content detection task and the steps to construct a harmful suicide content detection benchmark for this task. First, in the &#x201C;Task Design&#x201D; subsection, we outline the process of designing the task with emphasis on real-world implementation for detecting harmful suicide content in real time from online sources. This design phase includes detailed considerations for building the dataset.</p><p>Next, the &#x201C;Data Processing Workflow&#x201D; subsection describes the end-to-end process, from collecting harmful suicide content to developing the final harmful suicide content detection benchmark. Harmful suicide content is gathered from various online sources, and the preannotation process [<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref28">28</xref>], which uses LLMs, is used to reduce annotators&#x2019; exposure to vast quantities of harmful suicide content. Feedback generated from the annotation process by medical experts is used to update the task description document, which includes the categories and descriptions of harmful suicide content. This task description document is then used as an instruction for the preannotation process, creating an iterative process that builds and refines the harmful suicide content detection benchmark. Through this process, we ensure the inclusion of diverse, accurately labeled real-world harmful suicide content in the harmful suicide content detection benchmark while constructing a task description document that provides precise and comprehensive classification and explanations of the harmful suicide content.</p><p>As a result, from a collection of 43,244 harmful suicide content data points, we created a harmful suicide content detection benchmark consisting of 452 data points annotated by medical experts with suicide content categories, subcategories, and rationales for their classifications.</p></sec><sec id="s2-2"><title>Ethical Considerations</title><p>We obtained institutional review board (IRB) approval (approval number: KHUH 2023-04-072) from the IRB of Kyung-Hee University Hospital, South Korea, based on the principles of the Declaration of Helsinki, ensuring adherence to ethical standards. The IRB approval includes the collection of suicide-related data, the development of guidelines for data classification, and data labeling. It also specifies the acquisition of data through the collection of publicly available posts from online spaces targeting the general public, and it outlines the development of an artificial intelligence model for detecting suicide-related content using these data. All personally identifiable information (PII) (eg, user IDs, names, addresses, and phone numbers) was anonymized and replaced with standardized tags. The benchmark contains extremely disturbing text and images, including self-harming photographs, blood, tools used for suicide and self-harm, and drug information. Even among medical professionals and researchers, prolonged exposure to such images can lead to severe mental stress. Therefore, we have deliberately chosen not to aim for the creation of a large-scale dataset but rather to limit the workload to prevent further intensifying mental stress. All these processes were conducted with IRB approval obtained prior to data collection. Given the nature of harmful suicide content and the legal restrictions against its unrestricted distribution, it is challenging to share the benchmark dataset openly. We understand the legal implications of distributing data containing information that potentially induces suicide. Despite these concerns, we believe that collecting such data to build a benchmark and conducting research to prevent its spread on the internet outweigh these legal issues. Access to the benchmark will be strictly limited, allowing only those researchers with IRB approval and a commitment not to distribute the content further, ensuring responsible use for research purposes only and adherence to legal standards. We believe that our work contributes significantly to the ongoing international effort against harmful suicide content and hope to aid in preventing its spread on the internet.</p></sec><sec id="s2-3"><title>Task Design</title><p><xref ref-type="fig" rid="figure1">Figure 1</xref> illustrates the concept of using harmful suicide content detection in a real-world <italic>moderation system</italic>. The moderation system uses a model to automatically detect harmful suicide content and checks for illegal or harmful content and implements the appropriate <italic>moderation policy</italic> through a moderator&#x2019;s review. As this study introduces the harmful suicide content detection for the first time, our focus is on developing a model to automatically detect harmful suicide content rather than implementing an end-to-end moderation system. Therefore, this study focused on developing a harmful suicide content detection with this moderation system in mind, leaving the implementation of an end-to-end moderation system for future work. Specifically, we considered the inputs and outputs of harmful suicide content detection, considering the various real-world information on harmful suicide content that a moderation system might encounter as well as the moderation actions by a moderator.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Moderation system for harmful suicide content detection, categorizing online user-generated content into 5 classes by legality, harmfulness, and suicide relation. A moderator reviews content with potential illegality or harm, leading to legal reporting or content removal requests. No action is taken if no risks are found. &#x201C;O&#x201D; indicates that the corresponding value exists, and &#x201C;X&#x201D; indicates that the value does not exist. &#x201C;O/X&#x201D; denotes the potentiality of whether the corresponding value exists or not. LLM: large language model.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73725_fig01.png"/></fig><p>The model classifies the content into 5 distinct <italic>suicide categories</italic> based on illegality, harmfulness, and suicide-related aspects. For categories identified as having minimal harmfulness, a moderator validates the harmfulness through a review. Finally, the moderator moderates harmful suicide content using a moderation policy suitable for the identified harmfulness and illegality of the content, thereby minimizing human intervention and effectively moderating harmful suicide content.</p><p>To ensure that the detection system is applicable to real-world online data, the task targets various data from diverse sources to encompass a broad spectrum of user-generated content encountered on the internet. This approach ensures that the system effectively addresses the complexities and nuances of online posts (&#x201C;Task Input&#x201D; section).</p><p>The classification results produced by the harmful suicide content detection were crafted considering the functionality of a moderation system in mind. This means that the categories into which the content is sorted are specifically designed to facilitate the practical use of these results in moderating the content, ensuring that the system can serve as an effective tool for maintaining online safety and supporting mental well-being (&#x201C;Task Output&#x201D; section).</p><p>A moderator review was designed to validate the model&#x2019;s classification results and implement appropriate moderation. Specifically, for content categorized under the harmful categories, the process validates the harmfulness to ensure the reliability of the moderation system. In addition, because this process involves confirming the results of the model rather than newly identifying and classifying suicide content, it is more efficient in terms of reducing mental stress on moderators. Finally, suitable moderation policies were implemented based on the validation results. This system was developed to regulate suicide content and execute different moderation policies based on the illegality and harmfulness, thus preventing the spread of harmful suicide content online (&#x201C;Moderator Review&#x201D; section).</p><p>In summary, the entire process involves the model classifying online content into suicide categories, the moderator reviewing the results, and then implementing the corresponding moderation policy to ensure that the moderation system functions effectively. Throughout this process, multimodality information such as text and image data are used to reflect various aspects of the content in the model&#x2019;s input. Metadata from diverse sources and previous content serve as context. The model&#x2019;s output comprises 5 suicide categories, each differentiated by the presence or absence of illegality, harmfulness, and suicide-related aspects. Finally, the moderator review efficiently uses the model&#x2019;s classification results for validation, and effective moderation policies are implemented based on the content&#x2019;s illegality and harmfulness, thereby preventing the spread of harmful suicide content online.</p></sec><sec id="s2-4"><title>Task Input</title><sec id="s2-4-1"><title>Considerations</title><p>We consider the following for designing the input of the task. (1) <italic>Multimodality</italic>: Because 50% of harmful suicide content contain images or videos [<xref ref-type="bibr" rid="ref29">29</xref>], we consider text and images as inputs. (2) <italic>Source diversity</italic>: Harmful suicide content appears across various platforms, from social media to online communities [<xref ref-type="bibr" rid="ref29">29</xref>]. We collected data from diverse sources for comprehensive coverage. (3) <italic>Context information</italic>: We also incorporate previous content and metadata into our inputs. Previous content reveals the context of the target content, whereas metadata, such as user descriptions and view counts, provides additional insights, aiding in accurate harmfulness assessment.</p></sec><sec id="s2-4-2"><title>Inputs</title><p>Given the considerations, the inputs for the task are as follows: (1) <italic>User-generated content</italic>: Contents created by users. The content includes text, and possibly images and URLs. Images and URLs are converted to text manually or using machine learning models (eg, image captioning and summarization). (2) <italic>Previous content</italic>: Previous content is often required because it provides context, clarifies references, and provides background information essential for the full comprehension of user-generated content. (3) <italic>Metadata</italic>: Other contextual information about the user-generated content such as view counts, like counts, creation time, user self-description, and so on.</p></sec></sec><sec id="s2-5"><title>Task Output</title><sec id="s2-5-1"><title>Considerations</title><p>We consider the following for designing the outputs of the task. (1) <italic>Expert judgment</italic>: Harmful suicide content involves specific terminology related to suicide, such as professional drug names, slang, and abbreviations. Thus, clinical expertise is required to accurately determine the legality and harmfulness of such suicide content and to decide on an appropriate response to the content. (2) <italic>Moderation policy</italic>: If an automatic harmful suicide content detection model is developed, it should be part of a moderation system and collaborate with human moderators or domain experts [<xref ref-type="bibr" rid="ref18">18</xref>]. This implies that once the model detects harmful suicide content, it is necessary to consider appropriate actions. Therefore, the response of each output was considered when defining the output.</p></sec><sec id="s2-5-2"><title>Outputs</title><p>We develop five harmful suicide content categories. Content should be mapped to one of the following categories: (1) <italic>illegal</italic> content that encourages or assists suicidal behavior; (2) <italic>legal but harmful</italic> content that, while not illegal, significantly induces suicide; (3) <italic>potentially harmful</italic> content that could be triggering for certain individuals, whereas it may be benign for others; (4) <italic>harmless</italic> content that is either neutral or positive for suicide; and (5) <italic>nonsuicide</italic> content that is not related to suicide.</p></sec></sec><sec id="s2-6"><title>Moderator Review</title><p>Moderator review includes the process of reexamining the suicide content classified by the model using moderator (eg, a clinical expert) and implementing the appropriate moderation policy. The moderation system identifies the harmfulness and illegality of suicide content and implements a corresponding moderation policy to block the spread of such content online. Thus, through moderator review, the moderator (1) verifies the classification result of the model and (2) implements the corresponding moderation policy. The moderator review is conducted for results classified as illegal, harmful, and potentially harmful because most online information is unrelated to suicide and reviewing all the information would increase moderator fatigue. Hence, reviews are conducted only for suicide information that may cause harm. The moderator verifies the illegality and harmfulness of content within these categories and conducts the corresponding moderation policy. Therefore, the moderator requires knowledge to comprehend and understand the content and distinctions of suicide content.</p><sec id="s2-6-1"><title>Moderation Policies</title><p>The moderator reviews the model&#x2019;s classification results and implements a corresponding moderation policy. The moderation policies are as follows: (1) <italic>Report to police</italic>: This is the strongest form of moderation policy intended to subject content creators to legal regulations by reporting to legal institutions. (2) <italic>Report to online source</italic>: Reporting the content to the online source where it is posted intends to prevent the spread of harmful information by requesting the deletion of the content. (3) <italic>No report</italic>: No additional actions, such as reporting the posts, are taken, allowing it to circulate online.</p><p><italic>Report to police</italic> responds to content within the illegal suicide category containing illegal information. According to Korean law, certain types of content related to suicide are defined as illegal. Such content often includes illegal activities, such as the sale of illegal drugs; hence, reporting to legal institutions (eg, the police) imposes legal sanctions on the poster of such content.</p><p><italic>Report to online source</italic> prevents the online spread of content containing or potentially containing harmful information related to suicide, such as illegal, harmful, and potentially harmful content. Because information spreads quickly online, it is reported to the online source where it was posted, and its removal is requested to prevent dissemination. For potentially harmful information, the harmfulness of which can vary depending on the reader, the moderator assesses the degree of harmfulness and reports whether it is severe.</p><p><italic>No report</italic> is for harmless or nonsuicide content that poses no problem when posted online. Most online content is unrelated to suicide; therefore, it does not require reporting.</p><p><xref ref-type="table" rid="table1">Table 1</xref> shows the features of each category in terms of legality, harmfulness, and association with suicide. <italic>Illegal suicide content</italic> contains the most dangerous information, explicitly encouraging or facilitating suicidal behaviors. This category is critical for immediate intervention, embodying content that can actively propel individuals toward self-harm or suicide. <italic>Harmful suicide content</italic>, while not directly inciting suicide, significantly affects the audience by portraying suicide or self-harm in a manner that can trigger such actions among vulnerable individuals. The specificity of the depiction, whether through graphic imagery or detailed descriptions, amplifies its potential harm, making it a crucial target for moderating content. <italic>Potentially harmful suicide content</italic> traverses a gray area, with content that might not universally trigger harmful behaviors but could potentially do so in susceptible populations. This category underscores the complex challenge of content moderation, in which the impact of the content is not universally harmful and may vary significantly among individuals. <italic>Harmless suicide content</italic> focuses on providing support, hope, or neutral information regarding suicide without posing a risk of harm. This category plays an essential role in suicide prevention by offering resources, support, and information to reduce suicide rates. Finally, <italic>nonsuicide content</italic> serves as a catchall for material that does not pertain to suicide or self-harm, highlighting the importance of distinguishing between genuinely harmful content and content unrelated to suicide.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Illegality, harmfulness, and suicide relativity of categories and the moderation protocols<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup>.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Category</td><td align="left" valign="bottom">Illegality</td><td align="left" valign="bottom">Harmfulness</td><td align="left" valign="bottom">Suicide-<break/>related</td><td align="left" valign="bottom">Moderator<break/>review</td><td align="left" valign="bottom">Moderation<break/>policy</td></tr></thead><tbody><tr><td align="left" valign="top">Illegal</td><td align="left" valign="top">O</td><td align="left" valign="top">O</td><td align="left" valign="top">O</td><td align="left" valign="top">O</td><td align="left" valign="top">Report to police and online source</td></tr><tr><td align="left" valign="top">Harmful</td><td align="left" valign="top">X</td><td align="left" valign="top">O</td><td align="left" valign="top">O</td><td align="left" valign="top">O</td><td align="left" valign="top">Report to online source</td></tr><tr><td align="left" valign="top">Potentially harmful</td><td align="left" valign="top">X</td><td align="left" valign="top">&#x0394;</td><td align="left" valign="top">O</td><td align="left" valign="top">O</td><td align="left" valign="top">Report to online source or no report</td></tr><tr><td align="left" valign="top">Harmless</td><td align="left" valign="top">X</td><td align="left" valign="top">X</td><td align="left" valign="top">O</td><td align="left" valign="top">X</td><td align="left" valign="top">No report</td></tr><tr><td align="left" valign="top">Nonsuicide</td><td align="left" valign="top">X</td><td align="left" valign="top">N/A<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup></td><td align="left" valign="top">X</td><td align="left" valign="top">X</td><td align="left" valign="top">No report</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>The (&#x0394;) symbol represents a state that is in a gray area, indicating that the characteristic is neither fully present nor completely absent. "O" denotes the presence or applicability of a given attribute, whereas "X" denotes its absence. For instance, data categorized as "Harmful" are characterized by the absence of illegality (X), the presence of harmfulness (O), suicide-related (O), and the moderator review (O).</p></fn><fn id="table1fn2"><p><sup>b</sup>N/A: not applicable.</p></fn></table-wrap-foot></table-wrap></sec></sec><sec id="s2-7"><title>Data Processing Workflow</title><p>This section describes the data processing workflow used to create a high-quality harmful suicide content detection benchmark for detecting harmful suicide content. As illustrated in <xref ref-type="fig" rid="figure2">Figure 2</xref>, the workflow starts from data collection, detailed in &#x201C;Data Collections&#x201D; subsection, and outlines the process to develop a robust and diverse benchmark for harmful suicide content detection, described in &#x201C;Data Annotation&#x201D; subsection. The primary goal is to create a small but high-quality benchmark that can comprehensively capture real-world scenarios and task description document that can be effectively used for LLM instruction. Key challenges include the stressful nature of the annotation process, the extremely low prevalence of illegal suicide content, and the need to ensure the benchmark&#x2019;s completeness and use, given that this is a newly proposed task that must reflect various aspects of real-world data.</p><p>The methods used to address these challenges include (1) LLM sampling to reduce annotator exposure to a large volume of harmful suicide content while increasing the representation of rare, illegal suicide content; (2) an iterative annotation approach to continuously update the task description document and incorporate real-world scenarios, aiming to reflect diverse data until no further updates are required, while also considering modality (image and context) to adapt annotators effectively; and (3) filtering and validation, which involve updating the harmful suicide content detection benchmark to align with the latest task description document, categorizing and subcategorizing data accurately, and removing outliers such as incomprehensible content.</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Overview of the data processing workflow, which consists of data collection, benchmark construction, and benchmark validation. We collect suicide content and annotate it with medical experts, leveraging LLM-based preannotation to balance categories, and we iteratively refine the task description document through feedback. We then construct the benchmark corpus with human-labeled data, an English-translated version, and the updated task description, and we validate it for applicability to LLMs. AI: artificial intelligence; LLM: large language model.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73725_fig02.png"/></fig><sec id="s2-7-1"><title>Data Collections</title><p>Developing a large-scale harmful suicide content dataset is highly challenging. Harmful suicide content is infrequently encountered in real-world scenarios [<xref ref-type="bibr" rid="ref30">30</xref>], and the distressing nature of such content can cause mental strain for annotators. In addition, obtaining annotations from medical experts is expensive. Therefore, we focus on developing a high-quality curated benchmark dataset. Prior to the dataset collection, we obtained approval from the IRB of Kyung-Hee University Hospital, South Korea.</p><p>To cover the diverse source domains of the content, we collect user-generated content related to suicide from social media, Q&#x0026;A platforms, online support forums, and online communities. <xref ref-type="table" rid="table2">Table 2</xref> lists the number of raw data, benchmark data, descriptions of content, previous content, and metadata of each source.</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Number of collected harmful suicide content and harmful suicide content detection benchmark dataset for each domain<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Source</td><td align="left" valign="bottom">Raw data (no. of images)</td><td align="left" valign="bottom">Benchmark (no. of images)</td><td align="left" valign="bottom">Content</td><td align="left" valign="bottom">Previous content</td><td align="left" valign="bottom">Metadata</td></tr></thead><tbody><tr><td align="left" valign="top">Twitter</td><td align="left" valign="top">12,125 (3,671)</td><td align="left" valign="top">359 (78)</td><td align="left" valign="top">Tweets written by users</td><td align="left" valign="top">Previous tweets in the thread where the content is written</td><td align="left" valign="top">User description, view count, like count, etc</td></tr><tr><td align="left" valign="top">Online community</td><td align="left" valign="top">794 (794)</td><td align="left" valign="top">48 (48)</td><td align="left" valign="top">Title and bodies of post written by users</td><td align="left" valign="top">N/A<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup></td><td align="left" valign="top">User nickname, view count</td></tr><tr><td align="left" valign="top">Q&#x0026;A<sup><xref ref-type="table-fn" rid="table2fn3">c</xref></sup> platform</td><td align="left" valign="top">13,104 (0)</td><td align="left" valign="top">33 (0)</td><td align="left" valign="top">Question or answers written by users</td><td align="left" valign="top">Questions (if the content is an answer)</td><td align="left" valign="top">N/A</td></tr><tr><td align="left" valign="top">Online support forum</td><td align="left" valign="top">17,325 (0)</td><td align="left" valign="top">23 (0)</td><td align="left" valign="top">Counseling request posts or responses written by users or counselors</td><td align="left" valign="top">Counseling request posts (if the content is a response)</td><td align="left" valign="top">N/A</td></tr><tr><td align="left" valign="top">Total</td><td align="left" valign="top">43,244 (4,429)</td><td align="left" valign="top">452 (126)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">N/A</td><td align="left" valign="top">N/A</td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>User-generated suicide content was collected along with source-specific metadata and previous content as context.</p></fn><fn id="table2fn2"><p><sup>b</sup>N/A: not applicable.</p></fn><fn id="table2fn3"><p><sup>c</sup>Q&#x0026;A: question and answer.</p></fn></table-wrap-foot></table-wrap><sec id="s2-7-1-1"><title>Twitter</title><p>Twitter constitutes the majority of social media posts flagged for containing suicide-inducing information, with a substantial share of 74.69% [<xref ref-type="bibr" rid="ref29">29</xref>]. To collect data related to suicide from Twitter, we used the Twitter application programming interface (API) v2 to gather posts that include suicide-related keywords in their text or hashtags. These suicide-related keywords were collected from previous research [<xref ref-type="bibr" rid="ref31">31</xref>] and the guidelines of the &#x201C;Korean Suicide Inducing Information Monitoring Group&#x201D; [<xref ref-type="bibr" rid="ref32">32</xref>]. We gathered 12,021 tweets, including 3635 with images, from May to August 2023 using the Twitter API. The suicide-related keywords used in the Twitter API are summarized in <xref ref-type="table" rid="table3">Table 3</xref>.</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Suicide-related keywords<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup>.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="top">Classification of search terms</td><td align="left" valign="top">Search terms</td><td align="left" valign="top">Synonyms</td><td align="left" valign="top">Korean version of search terms</td><td align="left" valign="top">Korean version of synonyms</td></tr></thead><tbody><tr><td align="left" valign="top" rowspan="7">Suicide-related terms</td><td align="left" valign="top">Suicide</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Suicide</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xC0B4;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xC0B4;</p></list-item></list></td></tr><tr><td align="left" valign="top">Suicide method</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Suicide method</p></list-item><list-item><p>How to suicide</p></list-item><list-item><p>How to commit suicide</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xC0B4;&#xBC29;&#xBC95;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xC0B4;&#xBC29;&#xBC95;</p></list-item><list-item><p>&#xC790;&#xC0B4;&#xD558;&#xB294;&#xBC95;</p></list-item><list-item><p>&#xC790;&#xC0B4;&#xD558;&#xB294;&#xBC29;&#xBC95;</p></list-item></list></td></tr><tr><td align="left" valign="top">Dying method</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Dying method</p></list-item><list-item><p>How to die</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC8FD;&#xB294;&#xBC29;&#xBC95;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC8FD;&#xB294;&#xBC29;&#xBC95;</p></list-item><list-item><p>&#xC8FD;&#xB294;&#xBC95;</p></list-item></list></td></tr><tr><td align="left" valign="top">Suicidal idea</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Suicidal idea</p></list-item><list-item><p>Suicidal thinking</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xC0B4;&#xCDA9;&#xB3D9;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xC0B4;&#xCDA9;&#xB3D9;</p></list-item><list-item><p>&#xC790;&#xC0B4;&#xC0DD;&#xAC01;</p></list-item></list></td></tr><tr><td align="left" valign="top">Fall down</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Fall down</p></list-item><list-item><p>Suicide by jumping from a height</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xD22C;&#xC2E0;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xD22C;&#xC2E0;</p></list-item><list-item><p>&#xD22C;&#xC2E0;&#xC790;&#xC0B4;</p></list-item></list></td></tr><tr><td align="left" valign="top">Hanging</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Hanging</p></list-item><list-item><p>Hanging suicide</p></list-item><list-item><p>Neck hanging</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xBAA9;&#xB9F4;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xBAA9;&#xB9F4;</p></list-item><list-item><p>&#xBAA9;&#xB9F4;&#xC790;&#xC0B4;</p></list-item><list-item><p>&#xBAA9;&#xB9E4;&#xB2EC;&#xAE30;</p></list-item></list></td></tr><tr><td align="left" valign="top">Will</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Will</p></list-item><list-item><p>&#x00B7;How to write will</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC720;&#xC11C;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC720;&#xC11C;</p></list-item><list-item><p>&#xC720;&#xC11C;&#xC4F0;&#xB294;&#xBC95;</p></list-item></list></td></tr><tr><td align="left" valign="top" rowspan="6">Self-harm&#x2013;related terms</td><td align="left" valign="top">Self-harm</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Self-harm</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xD574;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xD574;</p></list-item></list></td></tr><tr><td align="left" valign="top">Self-harm method</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Self-harm method</p></list-item><list-item><p>How to self-harm</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xD574;&#xBC29;&#xBC95;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xD574;&#xBC29;&#xBC95;</p></list-item><list-item><p>&#xC790;&#xD574;&#xD558;&#xB294;&#xBC95;</p></list-item></list></td></tr><tr><td align="left" valign="top">Wrist cutting</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Wrist cutting</p></list-item><list-item><p>How to cut my wrist</p></list-item><list-item><p>Wrist cutting method</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC190;&#xBAA9;&#xC790;&#xD574;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC190;&#xBAA9;&#xC790;&#xD574;</p></list-item><list-item><p>&#xC190;&#xBAA9;&#xC790;&#xD574;&#xD558;&#xB294;&#xBC95;</p></list-item><list-item><p>&#xC190;&#xBAA9;&#xC790;&#xD574;&#xBC29;&#xBC95;</p></list-item></list></td></tr><tr><td align="left" valign="top">Self-harm wound</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Self-harm wound</p></list-item><list-item><p>Self-harm mark</p></list-item><list-item><p>Treatment for self-harm wound</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xD574;&#xD749;&#xD130;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xD574;&#xD749;&#xD130;</p></list-item><list-item><p>&#xC790;&#xD574;&#xC790;&#xAD6D;</p></list-item><list-item><p>&#xC790;&#xD574;&#xD749;&#xD130;&#xCE58;&#xB8CC;</p></list-item></list></td></tr><tr><td align="left" valign="top">Drug overdose</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Drug overdose</p></list-item><list-item><p>Drug lethal dose</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC57D;&#xBB3C;&#xACFC;&#xB2E4;&#xBCF5;&#xC6A9;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC57D;&#xBB3C;&#xACFC;&#xB2E4;&#xBCF5;&#xC6A9;</p></list-item><list-item><p>&#xC57D;&#xBB3C;&#xCE58;&#xC0AC;&#xB7C9;</p></list-item></list></td></tr><tr><td align="left" valign="top">Acetaminophen</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Acetaminophen overdose</p></list-item><list-item><p>Acetaminophen lethal dose</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xD0C0;&#xC774;&#xB808;&#xB180;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xD0C0;&#xC774;&#xB808;&#xB180;&#xACFC;&#xB2E4;&#xBCF5;&#xC6A9;</p></list-item><list-item><p>&#xD0C0;&#xC774;&#xB808;&#xB180;&#xCE58;&#xC0AC;&#xB7C9;</p></list-item></list></td></tr><tr><td align="left" valign="top" rowspan="7">Suicide risk factor terms</td><td align="left" valign="top">Academic score</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Academic concern</p></list-item><list-item><p>Academic score</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC131;&#xC801;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC131;&#xC801;</p></list-item><list-item><p>&#xC131;&#xC801;&#xACE0;&#xBBFC;</p></list-item></list></td></tr><tr><td align="left" valign="top">Bullying</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Bullying</p></list-item><list-item><p>Covert bullying</p></list-item><list-item><p>Outcast</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC655;&#xB530;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC655;&#xB530;</p></list-item><list-item><p>&#xC740;&#xB530;</p></list-item><list-item><p>&#xB530;&#xB3CC;&#xB9BC;</p></list-item></list></td></tr><tr><td align="left" valign="top">School violence</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>School violence</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xD559;&#xAD50;&#xD3ED;&#xB825;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xD559;&#xAD50;&#xD3ED;&#xB825;</p></list-item><list-item><p>&#xD559;&#xD3ED;</p></list-item></list></td></tr><tr><td align="left" valign="top">Family troubles</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Family troubles</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xAC00;&#xC871;&#xBB38;&#xC81C;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xAC00;&#xC871;&#xBB38;&#xC81C;</p></list-item></list></td></tr><tr><td align="left" valign="top">Domestic violence</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Domestic violence</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xAC00;&#xC815;&#xD3ED;&#xB825;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xAC00;&#xC815;&#xD3ED;&#xB825;</p></list-item></list></td></tr><tr><td align="left" valign="top">Dropout</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Dropout</p></list-item><list-item><p>How to drop out</p></list-item><list-item><p>Dropout method</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xD1F4;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xD1F4;</p></list-item><list-item><p>&#xC790;&#xD1F4;&#xD558;&#xB294;&#xBC95;</p></list-item><list-item><p>&#xC790;&#xD1F4;&#xBC29;&#xBC95;</p></list-item></list></td></tr><tr><td align="left" valign="top">Career</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Career</p></list-item><list-item><p>Career concern</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC9C4;&#xB85C;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC9C4;&#xB85C;</p></list-item><list-item><p>&#xC9C4;&#xB85C;&#xACE0;&#xBBFC;</p></list-item></list></td></tr><tr><td align="left" valign="top" rowspan="5">Suicide prevention terms</td><td align="left" valign="top">Suicide prevention</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Suicide prevention</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xC0B4;&#xC608;&#xBC29;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#x00B7;&#xC790;&#xC0B4;&#xC608;&#xBC29;</p></list-item></list></td></tr><tr><td align="left" valign="top">Call for life</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Call for life</p></list-item><list-item><p>Call for life of Korea</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC0DD;&#xBA85;&#xC758;&#xC804;&#xD654;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC0DD;&#xBA85;&#xC758;&#xC804;&#xD654;</p></list-item><list-item><p>&#xD55C;&#xAD6D;&#xC0DD;&#xBA85;&#xC758;&#xC804;&#xD654;</p></list-item></list></td></tr><tr><td align="left" valign="top">Suicide prevention center</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Suicide prevention center</p></list-item><list-item><p>1393</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xC0B4;&#xC608;&#xBC29;&#xC13C;&#xD130;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC790;&#xC0B4;&#xC608;&#xBC29;&#xC13C;&#xD130;</p></list-item><list-item><p>1393</p></list-item></list></td></tr><tr><td align="left" valign="top">Psychiatry</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#x00B7;Psychiatry</p></list-item><list-item><p>&#x00B7;Neuropsychiatry</p></list-item><list-item><p>Psychiatry department</p></list-item><list-item><p>Mental hospital</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC815;&#xC2E0;&#xACFC;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#x00B7;&#xC815;&#xC2E0;&#xACFC;</p></list-item><list-item><p>&#x00B7;&#xC2E0;&#xACBD;&#xC815;&#xC2E0;&#xACFC;</p></list-item><list-item><p>&#x00B7;&#xC815;&#xC2E0;&#xAC74;&#xAC15;&#xC758;&#xD559;&#xACFC;</p></list-item><list-item><p>&#xC815;&#xC2E0;&#xBCD1;&#xC6D0;</p></list-item></list></td></tr><tr><td align="left" valign="top">Mental health center</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>Mental health center</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC815;&#xC2E0;&#xAC74;&#xAC15;&#xBCF5;&#xC9C0;&#xC13C;&#xD130;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC815;&#xC2E0;&#xAC74;&#xAC15;&#xBCF5;&#xC9C0;&#xC13C;&#xD130;</p></list-item></list></td></tr><tr><td align="left" valign="top">Depression-related terms</td><td align="left" valign="top">Depression</td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#x00B7;Depression</p></list-item><list-item><p>&#x00B7;Depressed</p></list-item><list-item><p>&#x00B7;Depressive disorder</p></list-item><list-item><p>&#x00B7;Depressive symptom</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC6B0;&#xC6B8;&#xC99D;</p></list-item></list></td><td align="left" valign="top"><list list-type="bullet"><list-item><p>&#xC6B0;&#xC6B8;&#xC99D;</p></list-item><list-item><p>&#xC6B0;&#xC6B8;</p></list-item><list-item><p>&#x00B7;&#xC6B0;&#xC6B8;&#xC7A5;&#xC560;</p></list-item><list-item><p>&#x00B7;&#xC6B0;&#xC6B8;&#xC99D;&#xC0C1;</p></list-item></list></td></tr></tbody></table><table-wrap-foot><fn id="table3fn1"><p><sup>a</sup> Search terms and synonyms are defined in the study by Park et al [<xref ref-type="bibr" rid="ref33">33</xref>].</p></fn></table-wrap-foot></table-wrap></sec><sec id="s2-7-1-2"><title>Q&#x0026;A Platform</title><p>On Q&#x0026;A platforms, users often post questions about suicide-related issues, such as suicide methods, or respond to these queries. We collected questions and answers containing suicide-related keywords from Naver Knowledge In (a Korean Q&#x0026;A platform) [<xref ref-type="bibr" rid="ref33">33</xref>]. We collected data from March 2022 to March 2023, using the same keywords as those used for Twitter, resulting in 13,104 content items.</p></sec><sec id="s2-7-1-3"><title>Online Support Forum</title><p>In online support forums, people write about their suicide-related concerns, and counselors provide responses to support them [<xref ref-type="bibr" rid="ref31">31</xref>]. We collected posts from Lifeline Korea [<xref ref-type="bibr" rid="ref34">34</xref>] and the Companions of Life Suicide Prevention Counselling [<xref ref-type="bibr" rid="ref35">35</xref>]. We collected 17,325 pieces of content posted from March 2021 to June 2023.</p></sec><sec id="s2-7-1-4"><title>Online Community</title><p>DCinside [<xref ref-type="bibr" rid="ref36">36</xref>], a widely used online community in Korea comparable to Reddit, includes boards that function similarly to subreddits. We collected posts from 2 depression-focused boards (depression-minor and depression-mini boards) on the DCinside, known to contain suicide-related posts and where actual suicide incidents have been reported [<xref ref-type="bibr" rid="ref37">37</xref>]. We collected posts including those containing images, resulting in a total of 794 data entries. Overall, we collected 43,244 entries of harmful suicide content, in total, from 4 online sources, with each harmful suicide content entry comprising content, previous content, and metadata.</p></sec></sec></sec><sec id="s2-8"><title>Data Preprocessing</title><p>To remove private information from the benchmark and generate information that requires human intervention, such as link or image descriptions, we took the following steps. First, we removed all PII. This involves replacing URLs, names, locations, phone numbers, emails, and IDs within the text with corresponding tags. Thereafter, we provided supplementary descriptions for the contents of the external links. Given that these links may contain significant information for accurately understanding the content, we manually reviewed the links and summarized their content. Third, we added text descriptions to the images whenever they were included in the content. We used GPT-4 to generate initial descriptions, which were subsequently reviewed and refined for accuracy by the researchers. Consequently, all PII values were removed from the text of the data, and we created link descriptions that summarized the content of any URLs present in the content text, along with text descriptions for the images.</p></sec><sec id="s2-9"><title>Data Annotation</title><sec id="s2-9-1"><title>Task Description Document</title><p>The task description document was designed to explain the harmful suicide content detection and to provide guidance to the annotators. It contains vital information, including the purpose of identifying harmful suicide content and a detailed guide for annotating the content. In addition, it outlines the categories and subcategories of harmful suicide content, supplemented with real-world examples.</p><p>Our basis for understanding the definitions, categories, and examples of harmful suicide content was the &#x201C;Korean Suicide Prevention Law&#x201D; [<xref ref-type="bibr" rid="ref2">2</xref>] and documents published by the &#x201C;Korea Life Respect Hope Foundation&#x2019;s suicide/harmful information monitoring team&#x201D; [<xref ref-type="bibr" rid="ref32">32</xref>]. We found that certain category names and descriptions were unclear or overlapped, thus requiring more distinct clarifications. To address this, we involved medical professionals in the data annotation process, which led to significant revisions and refinements of the categories and their descriptions, as well as the expansion of examples for each category. Following the studies by Fiesler et al [<xref ref-type="bibr" rid="ref38">38</xref>] and Moon et al [<xref ref-type="bibr" rid="ref39">39</xref>], we used an iterative coding process such that the medical experts individually annotate the real-world content, come together to refine the task description document, and then repeat the coding process individually. This updating process was iterative and performed 3 times to ensure comprehensive refinement. Further details of the iterative process are presented in the &#x201C;Annotation Process&#x201D; section. We demonstrate each category and its description in <xref ref-type="table" rid="table4">Table 4</xref>.</p><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>Name and description of each suicide category.</p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Name</td><td align="left" valign="bottom">Description</td></tr></thead><tbody><tr><td align="left" valign="top">Illegal suicide content</td><td align="left" valign="top">Content that can actively encourage others to commit suicide or assist suicide behavior.</td></tr><tr><td align="left" valign="top">Harmful suicide content</td><td align="left" valign="top">Harmful content that is not as harmful as illegal suicide content but clearly has the effect of causing suicide or self-harm in the general public.</td></tr><tr><td align="left" valign="top">Potentially harmful suicide content</td><td align="left" valign="top">Content that may trigger suicide or self-harm in some people but may not cause it in others or may rather have a positive effect in others.</td></tr><tr><td align="left" valign="top">Harmless suicide content</td><td align="left" valign="top">Content that is not harmful, such as content that helps prevent suicide to the general public or provides neutral information related to suicide.</td></tr><tr><td align="left" valign="top">Nonsuicide content</td><td align="left" valign="top">Content unrelated to suicide.</td></tr></tbody></table></table-wrap></sec><sec id="s2-9-2"><title>Annotation Process</title><p>The annotation process was divided into 3 phases. In each phase, medical experts (a clinical expert with an MD degree and a psychiatry professor with a PhD degree) annotated real-world harmful suicide content, using the task description document as a reference. At the end of each phase, the authors and annotators reviewed and enhanced the task description document through discussions, before proceeding to the next phase.</p><p>In the first phase, medical professionals annotated harmful suicide content by referring to the initial task description document. Before starting the annotation, we preannotated and sampled the contents to be annotated from the data collections, as illustrated in <xref ref-type="fig" rid="figure3">Figure 3</xref>. Although the contents are gathered using suicide-related keywords, only a small fraction is actually harmful suicide content that can cause harm to others. Therefore, we used the task description document as instruction for the LLMs, allowing them to preliminarily categorize the content into predefined categories. This approach enhances the efficiency of the annotation process for medical experts and reduces mental strain and costs. Consequently, we used the OpenAI GPT API to preannotation and sampled 196 harmful suicide content for human annotation from the collected 2272 Twitter data, 17,325 online forum data, and 13,104 Q&#x0026;A data. Medical professionals then proceeded to annotate the sampled 196 harmful suicide content by following the annotation protocol and the initial task description document. The annotation protocol is described in the later part of this section. During the annotation process, they did not refer to the LLM labels generated by the LLM. Following the annotation, both the categories and subcategories were updated, leading to a revision of the task description document. Specifically, we refined 7 subcategories, added 2 new ones, and removed one.</p><p>In the second phase, we diversified the harmful suicide content in the benchmark and refined the task description document. Before annotation, we further preannotated and sampled 175 harmful suicide content for annotation from a pool of 8408 Twitter data points collected between May and June 2023. Similar to the first phase, we preannotated them using OpenAI GPT API with instructions written based on the task description document. Subsequently, medical professionals began the annotation of sampled harmful suicide content, strictly adhering to the annotation protocol and using the revised version of the task description document as their guide. Once the annotation process was completed, we merged the 4 subcategories into 2.</p><p>In the final phase, we added multimodal (text and image) harmful suicide content to the benchmark dataset and included online communities as an additional source domain. For the image content, we initially generated textual descriptions of harmful images using visual language LLMs (&#x201C;Data Preprocessing&#x201D; subsection). These initial descriptions were then revised to correct any inaccuracies or fill in missing details. The refined descriptions were subsequently used to preannotate the content into categories and subcategories, as defined in the task description document from the second phase. Following this process, we preannotated and sampled 95 multimodal harmful suicide content items for annotation. Medical professionals then annotated based on the annotation protocol, and the task description document was finalized by revising the previous version.</p><p>Finally, we manually verified the entire benchmark dataset. This involved identifying and eliminating any remaining PIIs from all harmful suicide content and validating the final labels. During the finalization process, 14 content items were excluded from the benchmark. These contents deal with subcultures (such as games and comics) and, therefore, are incomprehensible to all annotators and cannot be categorized into any suicide category, leading to their exclusion. In addition, the task description document was completed, providing comprehensive information on the 5 categories and 25 subcategories, including their harmful category names, descriptions, and illustrative examples.</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Illustration of the preannotation process. We construct prompts from the task description document, which includes categories, detailed descriptions, and subcategories (Table 4), and we use LLMs such as GPT-3.5 and GPT-4 to generate LLM-labeled data. We then sample the data to balance category distributions and have medical experts annotate them to build the human-labeled benchmark. AI: artificial intelligence; LLM: large language model.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73725_fig03.png"/></fig></sec><sec id="s2-9-3"><title>Annotation Protocol</title><p>In every phase, we adopted a consensus-based method for biomedical research and clinical practice [<xref ref-type="bibr" rid="ref40">40</xref>,<xref ref-type="bibr" rid="ref41">41</xref>]. For each harmful suicide content, 2 separate medical professionals (a clinical expert with an MD degree and a psychiatry professor with a PhD degree) independently labeled the category, subcategory, and rationale for their decisions regarding both the category and the subcategory. Each individual annotator assigned the label based on a comprehensive review of the user-generated content (text and image), previous content, and metadata associated with the harmful suicide content. The Inter-Annotator Agreement for category labels reached a high agreement of 0.77 (Cohen &#x03BA;) after the second phase of the annotation process. In cases where there is a discrepancy in the category label assigned by individual annotators, a consensus is established through the input of 3 annotators, which includes an additional clinical expert (a psychiatry professor with a PhD degree). During this consensus, rationales written by the 2 individual annotators are combined into a single rationale. In addition, annotators comment on any data whose association with suicide content is uncertain, as well as on instances that imply a potential need to revise the task description. These comments were used at the end of each annotation phase to refine and update the task description document.</p></sec></sec><sec id="s2-10"><title>Construction of English Benchmark</title><p>We further created an English benchmark by translating all input attributes (content text, link or image descriptions, and metadata such as user descriptions). To enhance the usability of the Korean harmful suicide content detection benchmark, it was translated into English using machine translation with GPT-4&#x2010;0613 for each attribute. As a result, we produced an English harmful suicide content detection benchmark containing the same 452 data entries as the original Korean harmful suicide content detection benchmark. The quality of the English benchmark is evaluated in detail in the &#x201C;English Benchmark Validation&#x201D; subsection.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Overview</title><p>After constructing the harmful suicide content detection benchmark through the data processing workflow, we evaluated the dataset from 3 perspectives. First, we assess whether the categories and descriptions of harmful suicide content outlined in the task description document are appropriate for providing sufficient information as instructions to LLMs to classify suicide contents (&#x201C;Leveraging Task Description&#x201D; subsection).</p><p>Next, we evaluate whether the data modalities included in the benchmark (eg, images) are suitable for classification by LLMs (&#x201C;Leveraging Multimodality&#x201D; subsection). In addition, we verify whether the LLMs can perform classification using only few-shot examples without further training through few-shot experiments (&#x201C;Leveraging Few-shot Examples&#x201D; subsection).</p><p>Finally, for the English harmful suicide content detection benchmark, we conduct experiments using various LLMs to compare performance with the original Korean benchmark to verify the use and carry out a direct analysis of data quality to verify the translation&#x2019;s quality (&#x201C;Comparison Between LLMs&#x201D; subsection and &#x201C;English Benchmark Validation&#x201D; subsection).</p><sec id="s3-1-1"><title>Setup</title><p>We used the GPT-3.5-turbo-16k API with a temperature of 0.0 and default hyperparameters, conducting 3-8 runs to calculate the average and standard error. For the few-shot experiments, we adopted an <italic>N</italic>-way <italic>K</italic>-shot approach by selecting <italic>K</italic> samples from each of the 5 classes (n=5) in the training dataset (&#x201C;Data Annotation&#x201D; subsection).</p></sec><sec id="s3-1-2"><title>Metrics</title><p>We used the following metrics:</p><list list-type="order"><list-item><p>Macro <italic>F</italic><sub>1</sub> measures the overall performance across the 5 categories.</p></list-item><list-item><p><italic>Mean Absolute Error (MAE) of Harmfulness</italic> measures the model&#x2019;s deviation in predicting harmfulness and is categorized into four levels: 3 (most harmful: Illegal Suicide Content), 2 (harmful: Harmful Suicide Content), 1 (potentially harmful: Potentially Harmful Suicide Content), and 0 (not harmful: Harmless Suicide Content and Nonsuicide Content). This metric assesses the extent of the error in terms of harmfulness.</p></list-item><list-item><p><italic>Illegal</italic> identifies illegal suicide content, and it is crucial for prompt regulation.</p></list-item><list-item><p>Harmful separates illegal or harmful suicide content from noncritical content, which is essential for moderating the content that poses harm.</p></list-item></list><p>In alignment with our objective of identifying and moderating as much harmful content as possible, our model is designed to initially detect harmful content, after which the results are carefully reviewed by a human moderator or expert. Hence, recall is prioritized over precision. This focus is particularly important for the Illegal and Harmful content categories, where missing instances of harmful suicide content pose a far greater risk than misclassifying content that is actually safe. Within our proposed framework, human moderators or clinical experts make the final judgment, ensuring that such misclassified cases can be corrected while minimizing the chance of harmful content going undetected. This design reflects how the system can function in real-world moderation scenarios, where recall is prioritized to maximize safety without disregarding precision.</p></sec></sec><sec id="s3-2"><title>Leveraging Task Description</title><p>We investigated the formulation of a task description document with diverse and extensive information into instructions because instruction construction significantly influences LLM performance [<xref ref-type="bibr" rid="ref42">42</xref>-<xref ref-type="bibr" rid="ref44">44</xref>]. The task description document for the harmful suicide content detection task contains crucial details, including the names and descriptions of 5 suicide categories as well as the names and explanations of 25 subcategories constituting up to 60% of the instruction at maximum. Category information includes detailed names and descriptions of the categories and subcategories. Thus, our experiments were designed to determine the details that most significantly impact performance by varying the granularity of the information.</p><sec id="s3-2-1"><title>Setup</title><p>We evaluate harmful suicide content detection performance by varying the detailed category information levels as follows: (1) category name; (2) category name and description; (3) category name with category description and subcategory name; and (4) category name with category description and subcategory name with subcategory description.</p><list list-type="bullet"><list-item><p>Category name and description</p></list-item><list-item><p>Category name with category description, and subcategory name</p></list-item><list-item><p>Category name with category description, subcategory name with subcategory description</p></list-item></list></sec><sec id="s3-2-2"><title>Results</title><p><xref ref-type="table" rid="table5">Table 5</xref> shows that the performance improves with more category information, with the most comprehensive level yielding the highest <italic>F</italic><sub>1</sub>-scores. Specifically, the macro <italic>F</italic><sub>1</sub>-score increases by 89% (from 18.86 to 35.75), and the illegal <italic>F</italic><sub>1</sub> and harmful <italic>F</italic><sub>1</sub>-scores increase by 200% (from 11.87 to 35.80) and 57% (from 37.63 to 59.10), respectively, as compared with when using only the category name.</p><p>The increasing trend in macro <italic>F</italic><sub>1</sub> and illegal or harmful <italic>F</italic><sub>1</sub>-scores suggests that more detailed information enhances the model&#x2019;s detection capabilities. However, adding only category descriptions decreased illegal or harmful recall (from 27.27 and 65.77 to 20.00 and 57.66, respectively).</p><table-wrap id="t5" position="float"><label>Table 5.</label><caption><p>Results from the category and subcategory information detail experiment<sup><xref ref-type="table-fn" rid="table5fn1">a</xref></sup>.</p></caption><table id="table5" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Category information</td><td align="left" valign="bottom">Subcategory information</td><td align="left" valign="bottom">Macro <italic>F</italic><sub>1</sub></td><td align="left" valign="bottom">MAE<sup><xref ref-type="table-fn" rid="table5fn2">b</xref></sup></td><td align="left" valign="bottom">Illegal <italic>F</italic><sub>1</sub></td><td align="left" valign="bottom">Illegal recall</td><td align="left" valign="bottom">Harmful <italic>F</italic><sub>1</sub></td><td align="left" valign="bottom">Harmful recall</td></tr></thead><tbody><tr><td align="left" valign="top">Name</td><td align="left" valign="top">&#x2003;N/A<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup></td><td align="left" valign="top">18.86 (0.08)</td><td align="left" valign="top">1.4809 (0.0044)</td><td align="left" valign="top">11.87 (0.04)</td><td align="left" valign="top">27.27 (0.00)</td><td align="left" valign="top">37.63 (0.11)</td><td align="left" valign="top">65.77 (0.00)</td></tr><tr><td align="left" valign="top">Name and description</td><td align="left" valign="top">&#x2003;N/A</td><td align="left" valign="top">25.13 (0.16)</td><td align="left" valign="top">1.2173 (0.0042)</td><td align="left" valign="top">13.23 (0.17)</td><td align="left" valign="top">20.00 (0.00)</td><td align="left" valign="top">39.18 (0.04)</td><td align="left" valign="top">57.66 (0.00)</td></tr><tr><td align="left" valign="top">Name and description</td><td align="left" valign="top">&#x2003;Name</td><td align="left" valign="top">32.18 (0.06)</td><td align="left" valign="top">0.9867 (0.0016)</td><td align="left" valign="top">26.36 (0.00)</td><td align="left" valign="top">30.91 (0.00)</td><td align="left" valign="top">45.26 (0.08)</td><td align="left" valign="top">66.67 (0.00)</td></tr><tr><td align="left" valign="top">Name and description</td><td align="left" valign="top">&#x2003;Name and description</td><td align="left" valign="top">35.75 (0.29<italic>)<sup><xref ref-type="table-fn" rid="table5fn1">a</xref></sup></italic></td><td align="left" valign="top">0.8549 (0.0079<bold>)</bold><sup><xref ref-type="table-fn" rid="table5fn1">a</xref></sup></td><td align="left" valign="top">35.80 (0.87<bold>)</bold><sup><xref ref-type="table-fn" rid="table5fn1">a</xref></sup></td><td align="left" valign="top">58.79 (1.21<bold>)</bold><sup><xref ref-type="table-fn" rid="table5fn1">a</xref></sup></td><td align="left" valign="top">59.10 (0.41<bold>)</bold><sup><xref ref-type="table-fn" rid="table5fn1">a</xref></sup></td><td align="left" valign="top">86.79 (0.60<bold>)</bold><sup><xref ref-type="table-fn" rid="table5fn1">a</xref></sup></td></tr></tbody></table><table-wrap-foot><fn id="table5fn1"><p><sup>a</sup>A consistent increase in macro <italic>F</italic><sub>1</sub>, illegal <italic>F</italic><sub>1</sub>, and harmful <italic>F</italic><sub>1</sub>-scores is observed as the amount of information increases. Values are mean (SE). These values indicate the best performance in each column.</p></fn><fn id="table5fn2"><p><sup>b</sup>MAE:<bold> </bold>mean absolute error.</p></fn><fn id="table5fn3"><p><sup>c</sup>N/A: not applicable.</p></fn></table-wrap-foot></table-wrap></sec></sec><sec id="s3-3"><title>Formulating LLM Inputs</title><p>We assessed performance changes by incorporating images and training examples as inputs. We focused on the impact of images as multimodal data (&#x201C;Leveraging Multimodality&#x201D; subsection) and the effect of using training data with the annotation guide as postinstruction when combined with instruction (&#x201C;Leveraging Few-shot Examples&#x201D; subsection).</p></sec><sec id="s3-4"><title>Leveraging Multimodality</title><sec id="s3-4-1"><title>Setup</title><p>The objective of this experiment was to determine the effect of image information on the classification performance of the model. We used 2 methods of conveying image information and compared their performances: the first method converts images into text descriptions, referred to as image description, whereas the second uses the images directly as inputs, referred to as vision. Three settings were tested for image descriptions: the first did not provide any image information, the second generated image descriptions using a model (gpt-4&#x2010;1106), and the third involved human modifications to the descriptions created by the model. This allowed for a comparison of the performance of the models based on the generation of text-based image descriptions. In addition, we examined the impact of images (vision) when paired with the same image descriptions to observe their influence on performance. This involved adding the original image to each image description experiment for comparison. Overall, this setup evaluates the model&#x2019;s performance in terms of the modality of suicide content through image descriptions and assesses the model&#x2019;s multimodal capabilities through vision. Notably, during the annotation process, the annotators labeled the suicide category of the content based on both the text and the original images. We used GPT-4-turbo-2024-04-09, which can use both text and image inputs, for this experiment. We conducted an experiment on 113 test data entries that included images, among which only 3 belonged to the illegal suicide category; thus, illegal metrics were excluded from the results.</p></sec><sec id="s3-4-2"><title>Results</title><p><xref ref-type="table" rid="table6">Table 6</xref> shows the impact of multimodal information on harmful suicide content detection tasks. In experiments regarding image descriptions without visual information, providing image details leads to superior performance compared with omitting them. Specifically, when using GPT-4&#x2013;generated image descriptions, macro <italic>F</italic><sub>1</sub> increased by 9.16% (from 50.46 to 55.08) and MAE decreased by 15.93% (from 0.3894 to 0.3333), indicating enhanced classification performance across all suicide categories. In addition, harmful <italic>F</italic><sub>1</sub> and recall both increased by 8.00% (from 68.50 to 73.98 and from 75.76 to 81.82), suggesting that image information significantly aids in identifying harmfulness within suicide content. Comparing GPT-4 and human-modified image descriptions, using human-modified descriptions results in a reduction of macro <italic>F</italic><sub>1</sub> by 3.55% (from 55.08 to 53.19) and increases MAE by 19.4% (from 0.3333 to 0.3982), although harmful <italic>F</italic><sub>1</sub> increases by 2.61% (from 73.98 to 75.91) and harmful recall by 16.66% (from 81.82 to 95.45), indicating that human modifications enhance clarity and detection of harmfulness in content while decreasing overall category performance.</p><table-wrap id="t6" position="float"><label>Table 6.</label><caption><p>Results of the input modality experiment on a subset of the benchmark that includes images<sup><xref ref-type="table-fn" rid="table6fn1">a</xref></sup>.</p></caption><table id="table6" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Image description (text)</td><td align="left" valign="bottom">Vision (image)</td><td align="left" valign="bottom">Macro <italic>F</italic><sub>1</sub></td><td align="left" valign="bottom">MAE<sup><xref ref-type="table-fn" rid="table6fn2">b</xref></sup></td><td align="left" valign="bottom">Illegal <italic>F</italic><sub>1</sub></td><td align="left" valign="bottom">Illegal recall</td><td align="left" valign="bottom">Harmful</td><td align="left" valign="bottom">Harmful recall</td></tr></thead><tbody><tr><td align="left" valign="top">No description</td><td align="left" valign="top">X</td><td align="left" valign="top">50.46 (0.12)</td><td align="left" valign="top">0.3864 (0.0135)</td><td align="left" valign="top">N/A<sup><xref ref-type="table-fn" rid="table6fn3">c</xref></sup></td><td align="left" valign="top">N/A</td><td align="left" valign="top">68.50 (0.21)</td><td align="left" valign="top">75.76 (0.26)</td></tr><tr><td align="left" valign="top">GPT-4 description</td><td align="left" valign="top">X</td><td align="left" valign="top">55.08 (0.09)<sup><xref ref-type="table-fn" rid="table6fn1">a</xref></sup></td><td align="left" valign="top">0.3333 (0.0102)<sup><xref ref-type="table-fn" rid="table6fn1">a</xref></sup></td><td align="left" valign="top">N/A</td><td align="left" valign="top">N/A</td><td align="left" valign="top">73.98 (0.09)</td><td align="left" valign="top">81.82 (0.00)</td></tr><tr><td align="left" valign="top">Human description</td><td align="left" valign="top">X</td><td align="left" valign="top">53.19 (0.13)</td><td align="left" valign="top">0.3982 (0.0154)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">N/A</td><td align="left" valign="top">75.91 (0.08)<sup><xref ref-type="table-fn" rid="table6fn1">a</xref></sup></td><td align="left" valign="top">95.45 (0.00)<sup><xref ref-type="table-fn" rid="table6fn1">a</xref></sup></td></tr><tr><td align="left" valign="top">No description</td><td align="left" valign="top">O</td><td align="left" valign="top">47.78 (0.05)</td><td align="left" valign="top">0.4189 (0.0102)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">N/A</td><td align="left" valign="top">67.42 (0.07)</td><td align="left" valign="top">90.91 (0.00)</td></tr><tr><td align="left" valign="top">GPT-4 description</td><td align="left" valign="top">O</td><td align="left" valign="top">50.08 (0.16)</td><td align="left" valign="top">0.3894 (0.0234)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">N/A</td><td align="left" valign="top">69.22 (0.34)</td><td align="left" valign="top">83.33 (0.26)</td></tr><tr><td align="left" valign="top">Human description</td><td align="left" valign="top">O</td><td align="left" valign="top">46.55 (0.15)</td><td align="left" valign="top">0.4631 (0.0270)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">N/A</td><td align="left" valign="top">69.37 (0.01)</td><td align="left" valign="top">90.91 (0.00)</td></tr></tbody></table><table-wrap-foot><fn id="table6fn1"><p><sup>a</sup>We compare 3 types of image descriptions (none, GPT-4-generated, and human-modified) with and without vision input. Without vision, GPT-4 descriptions achieve the best macro <italic>F</italic><sub>1</sub> and MAE, while human descriptions perform best on harmful metrics. With vision, we observe an overall decrease in performance. Values are mean (SE). These values indicate the best performance in each column. &#x201C;O&#x201D; indicates that the column is used (vision) and &#x201C;X&#x201D; indicates that  the column is not used.</p></fn><fn id="table6fn2"><p><sup>b</sup>MAE: mean absolute error.<italic> </italic></p></fn><fn id="table6fn3"><p><sup>c</sup>N/A: not applicable.</p></fn></table-wrap-foot></table-wrap><p>In experiments using image information as visual input, we found a general decrease in overall performance across all settings, with reductions in macro <italic>F</italic><sub>1</sub>, MAE, and harmful <italic>F</italic><sub>1</sub>. Even in scenarios without image descriptions, macro <italic>F</italic><sub>1</sub> decreased by 5.31% (from 50.46 to 47.78), and MAE increased by 8.41% (from 0.3864 to 0.4189). Particularly, the <italic>F</italic><sub>1</sub>-scores of potentially harmful content decreased significantly (from 54.46 to 42.77). This is owing to the model&#x2019;s sensitive reaction to certain images of potentially harmful suicide content, overestimating their harmfulness and classifying them as harmful.</p><p>However, in settings where we used images only as vision (no image description), the harmful recall score was 90.91, which was higher than when no image information was used (75.76); the score was 83.33 when using images with GPT-4 image description, which was higher than when using GPT-4 image description alone (81.82). This suggests that despite a decrease in the overall model performance owing to multimodality, using image information improves the identification of harmfulness in suicide content. In addition, vision can convey more information about harmfulness than text descriptions when human modification does not explicitly note the harmfulness of an image. Overall, the experiments with image descriptions confirmed that the information contained in an image enhances model performance in the harmful suicide content detection task, whereas adding vision information in a multimodal format decreases performance. However, the increase in harmful recall when using vision supports the potential of using vision as an effective tool for enhancing model capabilities in identifying harmful content, paving the way for future improvements in multimodal model performance.</p></sec></sec><sec id="s3-5"><title>Leveraging Few-Shot Examples</title><p>Few-shot refers to a setting where the model is provided with a small number of labeled examples as demonstrations before classifying new instances [<xref ref-type="bibr" rid="ref45">45</xref>]. The goal of our few-shot experiments is to assess whether LLMs can effectively learn from only a handful of examples in high-stakes domains, where large-scale annotation is often insufficient to train models.</p><sec id="s3-5-1"><title>Setup</title><p>We examined 1- to 5-shot configurations, corresponding to 1-5 examples per category, totaling 5-25 examples. To construct the pool of demonstrations, medical experts who participated in the benchmark annotation selected 10 representative samples per category (50 in total). These expert-selected samples were used as the source for training demonstrations, while the remaining annotated data served as the test set. From the pool, 5 examples per category were randomly drawn for each experiment. To ensure robustness, we repeated the same experiment with 3 different random seeds, each sampling distinct few-shot samples, and we report the averaged results across these runs.</p></sec><sec id="s3-5-2"><title>Results</title><p><xref ref-type="fig" rid="figure4">Figure 4</xref> shows how the performance metrics changed with the number of demonstration examples, with the standard error represented by vertical bars for each few-shot case. As the number of examples increased, macro <italic>F</italic><sub>1</sub>, MAE, and illegal metrics improved (<xref ref-type="fig" rid="figure4">Figure 4A</xref>), specifically illegal <italic>F</italic><sub>1</sub> and recall in the 5-shot (<xref ref-type="fig" rid="figure4">Figure 4B</xref>). Although the <italic>F</italic><sub>1</sub>-score for harmful effects remained relatively stable, recall increased but plateaued after a certain threshold (2-shot) (<xref ref-type="fig" rid="figure4">Figure 4</xref>C).</p><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>Results from the few-shot example experiment. Increasing examples increases illegal <italic>F</italic><sub>1</sub> and recall, with 5-shot setting achieving peak performance in the illegal metric. (A) Macro <italic>F</italic><sub>1</sub> and mean absolute error; (B) illegal <italic>F</italic><sub>1</sub> and recall; and (C) Harmful <italic>F</italic><sub>1</sub> and recall. MAE: mean absolute error.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73725_fig04.png"/></fig></sec></sec><sec id="s3-6"><title>Comparison Between LLMs</title><p>We compared the performance of various LLMs in identifying harmful suicide content. Because open-sourced LLMs have instruction-following capabilities that depend on the language they have seen in the instruction tuning phase, we conducted experiments with different models for Korean and English benchmarks to address language barriers.</p><sec id="s3-6-1"><title>Setup</title><p>We categorized the selected LLMs into closed and open-sourced models. For the Korean benchmark, we used closed models because of the lack of open-source or multilingual LLMs that can properly follow the task&#x2019;s instructions in Korean. We also included a random baseline that arbitrarily categorized content into one of the 5 categories.</p></sec></sec><sec id="s3-7"><title>Closed Models</title><p>We used OpenAI&#x2019;s GPT-3.5 (gpt-3.5-turbo-16k-0613) and GPT-4 (gpt-4&#x2010;1106-preview), which are accessed through the OpenAI API and capable of handling a context length of 1,28,000 characters. In addition, we experimented with Clova X, an LLM trained on Korean, using the Naver API [<xref ref-type="bibr" rid="ref46">46</xref>].</p></sec><sec id="s3-8"><title>Open-Sourced Models</title><p>We used the Zephyr-7B-beta model [<xref ref-type="bibr" rid="ref47">47</xref>], an enhanced version of mistral-7B, which supports a context length of up to 32,000 characters. We also use LongChat-7B-16k [<xref ref-type="bibr" rid="ref48">48</xref>] and Vicuna-7B-v1.5&#x2010;16k [<xref ref-type="bibr" rid="ref49">49</xref>], which are both fine-tuned Large Language Model Meta artificial intelligence (AI) models with a maximum context length of 16,000 characters.</p><p>To explore the model&#x2019;s adaptability of the model in few-shot learning contexts, we conducted experiments in both the zero-shot and 5-shot scenarios (&#x201C;Leveraging Few-Shot Examples&#x201D; subsection). However, for models unable to accept the context length of 12k tokens required for the 5-shot experiments, such as Clova X (4096), we limited our analysis to the zero-shot trials.</p><sec id="s3-8-1"><title>Results</title><p><xref ref-type="fig" rid="figure5">Figure 5</xref> shows the performance of the GPT models and Clova X on the Korean benchmark. GPT-4 outperformed all other models in every metric except for harmful recall (<xref ref-type="fig" rid="figure5">Figure 5A-E</xref>). GPT-3.5 follows GPT-4 in terms of performance across all metrics, except for harmful recall. Clova X showed lower performance than the GPT models but achieved the highest score in harmful recall (<xref ref-type="fig" rid="figure5">Figure 5</xref>F), indicating its high sensitivity to harmful content.</p><p><xref ref-type="fig" rid="figure6">Figure 6</xref> shows the performance of the GPTs and open-sourced LLMs on the translated English benchmark. GPT-4 exhibited the highest performance in differentiating categories in both the zero-shot and 5-shot settings across various metrics (<xref ref-type="fig" rid="figure6">Figure 6A-C and E</xref>). It leads the performance charts with macro <italic>F</italic><sub>1</sub>-scores of 46.37 in zero-shot and 52.59 in 5-shot. Notably, GPT-4 showed a significant MAE difference (0.5655 in zero-shot and 0.5755 in 5-shot), indicating that even when the category predictions were incorrect, they tended to be within similar levels of harmfulness.</p><p>GPT-3.5 ranked second to GPT-4 in category distinction performance (<xref ref-type="fig" rid="figure6">Figure 6A, B, and E</xref>) in zero-shot settings and showed comparable performance to open-sourced models in 5-shot settings. Its recall was relatively higher than GPT-4 in 5-shot settings (<xref ref-type="fig" rid="figure6">Figure 6C-F</xref>), indicating a more sensitive response to harmful information than the GPT-4.</p><p>Zephyr outperforms random in zero-shot settings with a macro <italic>F</italic><sub>1</sub> of 20.99 and MAE of 1.2711; in addition, it achieves a comparable performance to GPT-3.5 in 5-shot settings with macro <italic>F</italic><sub>1</sub> of 37.52 and MAE of 0.8217 (<xref ref-type="fig" rid="figure6">Figure 6A and B</xref> ). LongChat exhibits the largest standard error in the illegal and harmful metrics (<xref ref-type="fig" rid="figure6">Figure 6C-F</xref>), indicating that few-shot examples significantly impact performance compared with other models. It recorded the highest standard errors in illegal <italic>F</italic><sub>1</sub> and harmful <italic>F</italic><sub>1</sub> at 4.09 and 1.67, respectively (<xref ref-type="fig" rid="figure6">Figure 6C and E</xref>). LongChat also showed the lowest recall for illegal and harmful content, particularly in harmful content, suggesting that it is less sensitive to harmful information (<xref ref-type="fig" rid="figure6">Figure 6D and F</xref>). Vicuna recorded the lowest performance in category classification (<xref ref-type="fig" rid="figure6">Figure 6A</xref>) among all models but achieved high recall for illegal and harmful content (<xref ref-type="fig" rid="figure6">Figure 6D and F</xref>). Notably, it scored the highest illegal recall of 76.36 and a harmful recall of 83.78, comparable with GPT-3.5&#x2019;s 85.89.</p><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>Results from the Korean benchmark experiment. Hatched bars indicate the Korean large language model (Clova X). Although Clova X has lower overall performance than generative pretrained transformers, it excels in harmful recall. (A) Macro <italic>F</italic><sub>1</sub>; (B) mean absolute error; (C) illegal <italic>F</italic><sub>1</sub>; (D) illegal recall; (E) harmful <italic>F</italic><sub>1</sub>; and (F) harmful recall. MAE: mean absolute error.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73725_fig05.png"/></fig><fig position="float" id="figure6"><label>Figure 6.</label><caption><p>Results from the translated English benchmark experiment. Closed models (GPT-4 and GPT-3.5) show superior performance in the zero-shot setting compared with open-sourced models, whereas open-sourced models reach comparable performance to GPT-3.5 in 5-shot. (A) Macro <italic>F</italic><sub>1</sub>; (B) mean absolute error; (C) illegal <italic>F</italic><sub>1</sub>; (D) illegal recall; (E) harmful <italic>F</italic><sub>1</sub>; and (F) harmful recall. MAE: mean absolute error.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73725_fig06.png"/></fig></sec></sec><sec id="s3-9"><title>Analysis</title><sec id="s3-9-1"><title>Open-Sourced Versus Closed LLMs</title><p>GPT-4 recorded the highest performance across all accuracy metrics (macro <italic>F</italic><sub>1</sub>, MAE, illegal <italic>F</italic><sub>1</sub>, and harmful <italic>F</italic><sub>1</sub>) for all few-shot settings. In 5-shot settings, open-sourced models achieve a similar performance to GPT-3.5. However, in zero-shot settings, they struggled to understand lengthy instructions, resulting in random predictions (eg, LongChat) or biased predictions toward specific categories (eg, Vicuna), with Zephyr slightly outperforming random. In 5-shot scenarios, Zephyr matches GPT-3.5 in macro <italic>F</italic><sub>1</sub>, MAE, and harmful <italic>F</italic><sub>1</sub>, whereas LongChat and Vicuna show comparable performance in their respective metrics. Except for Vicuna, open-sourced models generally showed lower recall than closed models in terms of illegal and harmful content.</p></sec></sec><sec id="s3-10"><title>Original Korean Versus Translated English</title><p><xref ref-type="fig" rid="figure7">Figure 7</xref> shows an analysis of GPT-3.5 and GPT-4&#x2019;s performance on the Korean and translated English benchmarks. Both models performed better on the Korean benchmark across all <italic>F</italic><sub>1</sub> metrics (<xref ref-type="fig" rid="figure7">Figure 7</xref>A, C, and E). However, GPT-4 shows a decrease in macro <italic>F</italic><sub>1</sub> from the English to the Korean benchmark by 19.24% in zero-shot (from 57.42 to 46.37) and 9.51% in 5-shot (from 58.12 to 52.59), with the largest decrease in illegal <italic>F</italic><sub>1</sub> by 36.17% in zero-shot (from 64.80 to 39.85). GPT-3.5 also showed a considerable reduction in zero-shot illegal <italic>F</italic><sub>1</sub> by 62.17% (from 35.92 to 13.59). Illegal recall decreases considerably, with GPT-4 decreasing by 49.51% (from 62.43 to 31.52) and GPT-3.5 by 78.34% in illegal recall (from 58.79 to 12.73), indicating a larger decrease than in <italic>F</italic><sub>1</sub>-scores (<xref ref-type="fig" rid="figure7">Figure 7D</xref>). The decrease in harmful <italic>F</italic><sub>1</sub> is less significant, with GPT-4 decreasing by 7.00% (from 73.89 to 68.72) in zero-shot, and GPT-3.5 decreasing by 4.96% (from 59.73 to 56.77). This indicates that while translating the benchmark using GPT-4 does not significantly affect the overall quality, it may lead to issues in specific categories, notably illegal.</p><fig position="float" id="figure7"><label>Figure 7.</label><caption><p>Performance comparison of closed models (GPT-3.5 and GPT-4) on the Korean and translated English benchmarks. Closed models exhibit better classification performance on the Korean benchmark than on the English benchmark, with the most significant difference noted in classifying the illegal category. (A) Macro <italic>F</italic><sub>1</sub>; (B) mean absolute error; (C) illegal <italic>F</italic><sub>1</sub>; (D) illegal recall; (E) harmful <italic>F</italic><sub>1</sub>; and (F) harmful recall. MAE: mean absolute error.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="medinform_v14i1e73725_fig07.png"/></fig></sec><sec id="s3-11"><title>English Benchmark Validation</title><p>Considering the use of the English benchmark for experimenting with open-sourced models, it is necessary to evaluate the translation results. In particular, for content containing words related to suicide and harmful information, it is crucial to assess both the overall translation quality and how well the content has been translated. Because OpenAI&#x2019;s use policy potentially refuses to respond to harmful content, there may be instances in which proper translation has not been achieved. Therefore, we analyzed the following 2 aspects: (1) overall translation quality (quantitative analysis), and (2) translation of harmful content (qualitative analysis).</p><p>We evaluated the text content (CONTENT_TEXT) of every instance in the benchmark because every instance contains content text and it constitutes the largest proportion of text. We performed a quantitative analysis to assess translation quality using models (GPT-4&#x2010;0613) and a qualitative analysis of the translation of harmful content by the authors.</p><p><italic>Translation quality (quantitative analysis)</italic>: Translation quality assesses how similar the translated content (in English) is to the original content (in Korean). We evaluated translation quality using Scalar Quality Metric (SQM) and Direct Assessment (DA) methods [<xref ref-type="bibr" rid="ref50">50</xref>] through the GPT-4&#x2010;0613 API, which aligns closely with human evaluations. SQM evaluates the translation quality of the source text (Korean) and target text (English) on a scale of 0-100, with descriptions provided for &#x201C;no meaning preserved,&#x201D; &#x201C;some meaning preserved,&#x201D; &#x201C;most meaning preserved and few grammar mistakes,&#x201D; and &#x201C;perfect meaning and grammar.&#x201D; DA, such as SQM, rates translation quality on a scale of 0-100 but provides descriptions only for &#x201C;no meaning preserved&#x201D; and &#x201C;perfect meaning and grammar.&#x201D; On average, the translated contents scored 79.55 on SQM and 78.10 on DA, indicating that most instances of the benchmark translation results fall under &#x201C;most meaning preserved and few grammar mistakes,&#x201D; successfully retaining the original meaning.</p><p><italic>Translation of harmful content (qualitative analysis)</italic>: Illegal and harmful suicide content includes harmful words and expressions related to suicide and self-harm, encompassing abbreviations, drug names related to suicide, and expressions of methods for suicide and self-harm. Moreover, owing to OpenAI&#x2019;s use policy, there are cases in which harmful content is not translated or translation is refused. Thus, to determine how well such content was translated, we analyzed translation error cases for expressions related to suicide: (1) expressions related to suicide (abbreviations and words), and (2) OpenAI moderation.</p><p><italic>Expressions related to suicide</italic>: After analyzing 55 instances of illegal suicide content and 56 instances of harmful suicide content, we identified the following types of translation errors: (1) abbreviation translation error, (2) translation of substances used for suicide, and (3) translation of slang related to suicide and self-harm.</p><p><xref ref-type="table" rid="table7">Table 7</xref> shows examples of translation errors for each category and error type. <italic>Abbreviation translation error</italic> occurs when abbreviations related to suicide and self-harm are incorrectly interpreted. To evade online platform moderation, abbreviations related to suicide are often used. In these cases, the translation process incorrectly translates these abbreviations into entirely different words. The Korean abbreviation means &#x201C;commit double suicide and death leap&#x201D;; however, the English translation misinterprets it entirely. In this study, 12 benchmark instances were identified. <italic>Translation of Substances used for suicide</italic> refers to cases in which drugs related to suicide and self-harm were incorrectly translated. Substances used for suicide are often referred to by abbreviations to avoid online platform moderation, and the actual drug names are often translated into general names for drugs during translation. In this example, the drug &#x201C;zolpidem&#x201D; was translated as &#x201C;SleepingPill,&#x201D; which translates to the purpose of the drug (zolpidem is a type of sleeping pill) rather than the actual name of the drug. However, such translations result in the inability of the model to correctly identify the sale of specific drugs (illegal suicide category) during the category classification process. In this study, 15 instances were identified. <italic>Translation of slang related to suicide and self-harm </italic>refers to errors in the translation of clear expressions of suicide. For example, the Korean expression for &#x201C;bloodletting self-harm&#x201D; was incorrectly translated as &#x201C;blood donation&#x201D; in English, which changed the meaning of the text. Three instances were identified for this case.</p><table-wrap id="t7" position="float"><label>Table 7.</label><caption><p>Qualitative analysis of benchmark translation results. The source text is the content text from the Korean benchmark data, and the proper translation is the result translated by a human while preserving the meaning. The translation result is obtained using a model and has been applied to the English benchmark. The underlined words indicate parts where translation errors occurred in the model&#x2019;s output.</p></caption><table id="table7" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Type</td><td align="left" valign="bottom">Content text</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="2">Translation error in the abbreviation (illegal suicide content)</td></tr><tr><td align="left" valign="bottom">Source text</td><td align="left" valign="bottom"><underline>&#x3137;&#x3142;&#x3148;&#x3145; &#x314C;&#x3145;</underline> &#xD558;&#xC2E4; &#xBD84; &#xC790;&#xCC28;&#xB098; &#xBA74;&#xD5C8; &#xC788;&#xC73C;&#xC2E0; &#xBD84;&#xC73C;&#xB85C;&#xC694;</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Proper translation (human)</td><td align="left" valign="top">Looking for someone who will <underline>commit double suicide and death leap</underline>, preferably someone with a car or a driver&#x2019;s license.</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Translation result (model)</td><td align="left" valign="top">Looking for someone who can <underline>do delivery and sales</underline>, preferably someone with a car or a driver&#x2019;s license.</td></tr><tr><td align="left" valign="top" colspan="2">Translation error in substances that are used for suicide (illegal suicide content)</td></tr><tr><td align="left" valign="top">Source text</td><td align="left" valign="top">&#xC815;&#xD488;#<underline>&#xC878;&#x314D;&#x3163;&#xB380;</underline>&#xD310;&#x3141;&#x3150; &#xD558;&#xB294;&#xACF3; &#xAC00;&#xC9DC;&#xC5C6;&#xACE0; &#xC0AC;&#xAE30;&#xC5C6;&#xB294; &#xC194;&#xC9C1;&#xD55C; &#xBE44;&#xBC00;&#xC0C1;&#xB2F4; &#xAD00;&#xC2EC;&#x7A7A; #&#xC5EC;&#xC131;&#xCD5C;&#xC74C;&#x3148;&#x3154; &#x3131;&#x315C;&#xC785;&#xD558;&#xB294; &#xACF3;. &#xB77C;&#xC778;: [ID] &#xCE74;&#xD1A1;: [ID] &#xD154;&#xB808;: [ID]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Proper translation (human)</td><td align="left" valign="top">#Genuine<underline>zolpidem</underline>SellingPlace A place where there are no fakes, no scams, honest confidential consultation. Interested in #FemaleAphrodisiac Purchase place. Line: [ID] Kakaotalk: [ID] Telegram: [ID]</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Translation result (model)</td><td align="left" valign="top">#Genuine<underline>SpleepingPill</underline>SellingPlace A place where there are no fakes, no scams, honest confidential consultation. Interested in #FemaleAphrodisiac Purchase place. Line: [ID] Kakaotalk: [ID] Telegram: [ID]</td></tr><tr><td align="left" valign="top" colspan="2">Translation of slang related to suicide or self-harm (harmful suicide content)</td></tr><tr><td align="left" valign="top">Source text</td><td align="left" valign="top"><underline>&#xC0AC;&#xD608;&#xC790;&#xD574;</underline>&#xB97C; &#xD574;&#xBCF4;&#xACE0;&#xC2F6;&#xB2E4;&#xB294; &#xC0DD;&#xAC01;&#xC774; &#xB4E4;&#xC5B4;</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Proper translation (human)</td><td align="left" valign="top">I&#x2019;m thinking about wanting to try <underline>bleeding self-harm</underline>.</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Translation result (model)</td><td align="left" valign="top">I&#x2019;m thinking about wanting to try <underline>blood donation</underline>.</td></tr><tr><td align="left" valign="top" colspan="2"/></tr></tbody></table></table-wrap></sec><sec id="s3-12"><title>OpenAI Moderation</title><p>During the translation process of the benchmark data, we found a few instances where different translation errors occur from those related to expressions of suicide; some content are not translated at all. Unlike parts of the text that are translated incorrectly, the OpenAI API (gpt-4&#x2010;1106) refuses to translate and generates content (translated output) that has no relation to the original text (input text). Two instances occurred for this case; one involving content asking for help in committing suicide or self-harm and another involving direct sexual content.</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>English Benchmark Integrity</title><p>The English benchmark was generated through machine translation using GPT-4. Quantitative evaluation showed relatively high scores (SQM: 79.55, DA: 78.10), indicating that most meanings were preserved. However, qualitative analysis revealed errors in suicide-related slang, abbreviations, and substance names (eg, bleeding self-harm translated as blood donation, zolpidem simplified to sleeping pill). Such inaccuracies may affect model performance and limit the benchmark&#x2019;s validity for multilingual generalization. Therefore, results on the English benchmark should be interpreted with caution. Future work will involve human validation of high-risk categories and improved translation methods for suicide-specific terminology.</p></sec><sec id="s4-2"><title>Benchmark Size</title><p>Although the harmful suicide content benchmark is an essential step toward understanding and moderating online suicide-related content, it encompasses 452 data entries. This relatively small benchmark size is largely attributable to the fact that posts related to suicide comprise a small fraction of the total online content. In addition, filtering and deleting such content by online sources inherently limit the volume of data available for collection. Nevertheless, a carefully controlled annotation process that incorporates the knowledge of clinical experts supports the credibility of the benchmark and ensures a reliable set of labels. In addition, the task description document details 25 different subcategories of suicide content, and the benchmark comprises a wide array of suicide content, including actual data for each subcategory. Therefore, our detailed task description document and the data within our benchmark lay the groundwork for future efforts to create a large-scale suicide content dataset using the annotations described for the suicide content.</p></sec><sec id="s4-3"><title>Dataset Usage</title><p>The harmful suicide content benchmark includes a task description document containing categories and subcategories of suicide content, along with a dataset of 452 annotated entries. Users can leverage the contents of the task description document as instructions for LLMs. In addition, data labeled for training within the benchmark can be used as few-shot examples, while data labeled for testing can serve as evaluation data.</p></sec><sec id="s4-4"><title>Future Work</title><p>In clinical practice, detecting that a patient who has attempted suicide or self-harm has been exposed to suicide-related content can be highly valuable for preventing further self-harm or suicide attempts. This is because restricting harmful online content represents a relatively modifiable factor among the many risk factors associated with suicide and self-harm [<xref ref-type="bibr" rid="ref51">51</xref>,<xref ref-type="bibr" rid="ref52">52</xref>]. Building on this perspective, in future work, we plan to further validate the practical effectiveness of our suicide content detection model. Specifically, we aim to conduct a comparative study between individuals who have attempted suicide or self-harm and a control group of individuals without such experiences. By analyzing their online activity, we will investigate whether those with a history of suicidal or self-harming behaviors are more frequently exposed to harmful suicide-related content. Such a study would not only provide empirical evidence for the real-world use of our detection framework but also offer insights into the role of online harmful content in influencing vulnerable populations.</p></sec></sec></body><back><notes><sec><title>Funding</title><p>This research was partially supported by the Bio &#x0026; Medical Technology Development Program of the National Research Foun- dation (NRF) and funded by the Korean government (MSIT) (NRF-2021M3A9E4080780 and 2021M3A9E4080784) and Institute of Information &#x0026; communications Technology Planning &#x0026; Evaluation (IITP) grant funded by the Korea government (MSIT) (RS-2019-II190421, AI Graduate School Support Program(Sungkyunkwan University), and Institute of Information &#x0026; communications Technology Planning &#x0026; Evaluation (IITP) grant funded by the Korea government (MSIT) (no. RS-2024&#x2010;00509258 and no. RS-2024-00469482, Global AI Frontier Lab).</p></sec><sec><title>Data Availability</title><p>Our benchmark dataset contains sensitive and, in some cases, illegal information related to suicide. Due to the sensitivity of suicide-related content, the full benchmark dataset cannot be publicly released and is granted only through a controlled approval process [<xref ref-type="bibr" rid="ref53">53</xref>]. In addition, since the dataset also contains illegal information, it cannot be made publicly available; however, it may be shared under appropriate requests and upon review by the authors [<xref ref-type="bibr" rid="ref54">54</xref>]. We provide the website [<xref ref-type="bibr" rid="ref55">55</xref>] to obtain the benchmark dataset and the codes used in the experiments. Applicants must follow the outlined process, after which the authors will review the requests and share the dataset as appropriate. This version includes only references to the original data (eg, tweet IDs, and URLs) and annotations by medical experts, ensuring that no legal or ethical violations occur upon public release.</p></sec></notes><fn-group><fn fn-type="con"><p>KP, MJB, RL, JM, JWP, and SP contributed to conception. KP, MJB, RL, JM, KC, JWP, and SP contributed to idea and theory. KP, MJB, YS, SML, YHS, ARL, SYY, JWP, and SP contributed to task design. KP, YH, YS, HL, and SP contributed to building data collections. MJB, RL, and JWP contributed to data annotation. KP, YH, YS, and SP contributed to data processing. KP, YH, YS, SML, YHS, ARL, SYY, and SP contributed to building the benchmark. KP, YH, DL, JB, KC, and SP contributed to technical validation. All authors contributed to the final manuscript and reviewed and approved the final manuscript. JB (jy.bak@skku.edu), KC (kyunghyun.cho@nyu.edu), JWP (paikjw@khu.ac.kr), and SP (sungjoon.park@softly.ai, psj6512@gmail.com) are equal co-corresponding authors.</p></fn><fn fn-type="conflict"><p>KP, DL, JM, and SP are affiliated with SoftlyAI. MJB and ARL are affiliated with Kyung Hee University Medical Center. YH and JB are affiliated with SungKyunKwan University. YS and HL are affiliated with KAIST. RL is affiliated with the University of Pennsylvania. MJB, SML, YHS, ARL, SYY, and JWP are affiliated with Kyung Hee University. KC is affiliated with New York University.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">API</term><def><p>application programming interface</p></def></def-item><def-item><term id="abb3">DA</term><def><p>Direct Assessment</p></def></def-item><def-item><term id="abb4">IRB</term><def><p>institutional review board</p></def></def-item><def-item><term id="abb5">MAE</term><def><p>mean absolute error</p></def></def-item><def-item><term id="abb6">PII</term><def><p>Personally Identifiable Information</p></def></def-item><def-item><term id="abb7">Q&#x0026;A</term><def><p>question and answer</p></def></def-item><def-item><term id="abb8">SQM</term><def><p>Scalar Quality Metric</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="web"><article-title>Understanding self-harm and suicide content online</article-title><source>Samaritans</source><year>2020</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.samaritans.org/about-samaritans/research-policy/internet-suicide/guidelines-tech-industry/understanding-self-harm-and-suicide-content/">https://www.samaritans.org/about-samaritans/research-policy/internet-suicide/guidelines-tech-industry/understanding-self-harm-and-suicide-content/</ext-link></comment></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="web"><article-title>&#xC790;&#xC0B4;&#xC608;&#xBC29; &#xBC0F; &#xC0DD;&#xBA85;&#xC874;&#xC911;&#xBB38;&#xD654; &#xC870;&#xC131;&#xC744; &#xC704;&#xD55C; &#xBC95;&#xB960; [The act on the prevention of suicide and the creation of a culture of respect for life]</article-title><source>MOHW</source><year>2019</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://tinyurl.com/42xe7rwy">https://tinyurl.com/42xe7rwy</ext-link></comment></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zdanow</surname><given-names>C</given-names> </name><name name-style="western"><surname>Wright</surname><given-names>B</given-names> </name></person-group><article-title>The representation of self injury and suicide on emo social networking groups</article-title><source>Afr Sociol Rev Afr de Sociol</source><year>2012</year><access-date>2025-12-20</access-date><volume>16</volume><fpage>81</fpage><lpage>101</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://www.jstor.org/stable/afrisocirevi.16.2.81">https://www.jstor.org/stable/afrisocirevi.16.2.81</ext-link></comment></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Biddle</surname><given-names>L</given-names> </name><name name-style="western"><surname>Gunnell</surname><given-names>D</given-names> </name><name name-style="western"><surname>Owen-Smith</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Information sources used by the suicidal to inform choice of method</article-title><source>J Affect Disord</source><year>2012</year><month>02</month><volume>136</volume><issue>3</issue><fpage>702</fpage><lpage>709</lpage><pub-id pub-id-type="doi">10.1016/j.jad.2011.10.004</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Milmo</surname><given-names>D</given-names> </name></person-group><article-title>'The bleakest of worlds&#x2019;: how Molly Russell fell into a vortex of despair on social media</article-title><source>The Guardian</source><year>2022</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.theguardian.com/technology/2022/sep/30/how-molly-russell-fell-into-a-vortex-of-despair-on-social-media">https://www.theguardian.com/technology/2022/sep/30/how-molly-russell-fell-into-a-vortex-of-despair-on-social-media</ext-link></comment></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Balt</surname><given-names>E</given-names> </name><name name-style="western"><surname>M&#x00E9;relle</surname><given-names>S</given-names> </name><name name-style="western"><surname>Robinson</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Social media use of adolescents who died by suicide: lessons from a psychological autopsy study</article-title><source>Child Adolesc Psychiatry Ment Health</source><year>2023</year><month>04</month><day>7</day><volume>17</volume><issue>1</issue><fpage>48</fpage><pub-id pub-id-type="doi">10.1186/s13034-023-00597-9</pub-id><pub-id pub-id-type="medline">37029395</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Twenge</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Haidt</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lozano</surname><given-names>J</given-names> </name><name name-style="western"><surname>Cummins</surname><given-names>KM</given-names> </name></person-group><article-title>Specification curve analysis shows that social media use is linked to poor mental health, especially among girls</article-title><source>Acta Psychol (Amst)</source><year>2022</year><month>04</month><volume>224</volume><fpage>103512</fpage><pub-id pub-id-type="doi">10.1016/j.actpsy.2022.103512</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="report"><article-title>National suicide prevention strategies</article-title><source>World Health Organization</source><year>2018</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://iris.who.int/server/api/core/bitstreams/9b9ea82a-b798-4c63-b90e-74bf8bdbf6b1/content">https://iris.who.int/server/api/core/bitstreams/9b9ea82a-b798-4c63-b90e-74bf8bdbf6b1/content</ext-link></comment></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Donelan</surname><given-names>M</given-names> </name></person-group><article-title>Secretary of State for Science, Innovation and Technology, the Lord Parkinson of Whitley Bay &#x0026; Parliamentary Undersecretary of State for Arts and Heritage</article-title><source>Online Safety Act 2023</source><year>2023</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.legislation.gov.uk/ukpga/2023/50">https://www.legislation.gov.uk/ukpga/2023/50</ext-link></comment></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="web"><article-title>World health statistics 2023: monitoring health for the SDGS, sustainable development goals</article-title><source>World Health Organization</source><year>2023</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.who.int/publications/i/item/9789240074323">https://www.who.int/publications/i/item/9789240074323</ext-link></comment></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Jung</surname><given-names>K</given-names> </name></person-group><article-title>&#xCC98;&#xBC8C;&#xC870;&#xD56D; &#xC720;&#xBA85;&#xBB34;&#xC2E4; [Ministry of Health and Welfare requests zero investigations on suicide-inducing information on SNS, rendering punitive clauses ineffective]</article-title><year>2022</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="http://www.bosa.co.kr/news/articleView.html?idxno=2182323">http://www.bosa.co.kr/news/articleView.html?idxno=2182323</ext-link></comment></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Min</surname><given-names>S</given-names> </name></person-group><article-title>&#xC790;&#xC0B4;&#xC720;&#xBC1C;&#xC815;&#xBCF4; 4&#xB144;&#xAC04; 7&#xBC30; &#xB298;&#xC5C8;&#xB294;&#xB370; &#x00B7; &#x00B7; &#x00B7; &#xBAA8;&#xB2C8;&#xD130;&#xB9C1; &#xC804;&#xB2F4;&#xC778;&#xB825;&#xC740; 10&#xB144;&#xC9F8; 1&#xBA85;&#xBFD0; [Incidents of suicide-inducing information increase sevenfold in four years, yet monitoring staff remains solely one for a decade]</article-title><year>2023</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.khan.co.kr/article/202309241417001">https://www.khan.co.kr/article/202309241417001</ext-link></comment></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="report"><article-title>2022 internet usage survey summary report (2022&#xB144;&#xB3C4; &#xC778;&#xD130;&#xB137;&#xC774;&#xC6A9;&#xC2E4;&#xD0DC;&#xC870;&#xC0AC; &#xC694;&#xC57D;&#xBCF4;&#xACE0;&#xC11C;)</article-title><year>2022</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.msit.go.kr/bbs/view.do?sCode=user&#x0026;bbsSeqNo=79&#x0026;nttSeqNo=3173583">https://www.msit.go.kr/bbs/view.do?sCode=user&#x0026;bbsSeqNo=79&#x0026;nttSeqNo=3173583</ext-link></comment></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Zirikly</surname><given-names>A</given-names> </name><name name-style="western"><surname>Resnik</surname><given-names>P</given-names> </name><name name-style="western"><surname>Uzuner</surname><given-names>&#x00D6;</given-names> </name><name name-style="western"><surname>Hollingshead</surname><given-names>K</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Niederhoffer</surname><given-names>K</given-names> </name><name name-style="western"><surname>Hollingshead</surname><given-names>K</given-names> </name><name name-style="western"><surname>Resnik</surname><given-names>P</given-names> </name><name name-style="western"><surname>Resnik</surname><given-names>R</given-names> </name><name name-style="western"><surname>Loveys</surname><given-names>K</given-names> </name></person-group><article-title>CLPsych 2019 shared task: predicting the degree of suicide risk in reddit posts</article-title><year>2019</year><conf-name>Proceedings of the Sixth Workshop on Computational Linguistics and Clinical Psychology</conf-name><conf-date>Jun 6, 2019</conf-date><conf-loc>Minneapolis, Minnesota</conf-loc><fpage>24</fpage><lpage>33</lpage><pub-id pub-id-type="doi">10.18653/v1/W19-3003</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Milne</surname><given-names>DN</given-names> </name><name name-style="western"><surname>Pink</surname><given-names>G</given-names> </name><name name-style="western"><surname>Hachey</surname><given-names>B</given-names> </name><name name-style="western"><surname>Calvo</surname><given-names>RA</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Hollingshead</surname><given-names>K</given-names> </name><name name-style="western"><surname>Ungar</surname><given-names>L</given-names> </name></person-group><article-title>CLPsych 2016 shared task: triaging content in online peer-support forums</article-title><year>2016</year><conf-name>Proceedings of the Third Workshop on Computational Linguistics and Clinical Psychology</conf-name><conf-date>Jun 16, 2016</conf-date><conf-loc>San Diego, CA, USA</conf-loc><fpage>118</fpage><lpage>127</lpage><pub-id pub-id-type="doi">10.18653/v1/W16-0312</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Yates</surname><given-names>A</given-names> </name><name name-style="western"><surname>Cohan</surname><given-names>A</given-names> </name><name name-style="western"><surname>Goharian</surname><given-names>N</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Palmer</surname><given-names>M</given-names> </name><name name-style="western"><surname>Hwa</surname><given-names>R</given-names> </name><name name-style="western"><surname>Riedel</surname><given-names>S</given-names> </name></person-group><article-title>Depression and self-harm risk assessment in online forums</article-title><year>2017</year><conf-name>Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing</conf-name><conf-date>Sep 9-11, 2017</conf-date><conf-loc>Copenhagen, Denmark</conf-loc><fpage>2968</fpage><lpage>2978</lpage><pub-id pub-id-type="doi">10.18653/v1/D17-1322</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Yang</surname><given-names>C</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Muresan</surname><given-names>S</given-names> </name></person-group><article-title>Weakly-supervised methods for suicide risk assessment: role of related domains</article-title><conf-name>Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 22: Short Papers)</conf-name><conf-date>Aug 1-6, 2021</conf-date><pub-id pub-id-type="doi">10.18653/v1/2021.acl-short.133</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Sawhney</surname><given-names>R</given-names> </name><name name-style="western"><surname>Neerkaje</surname><given-names>A</given-names> </name><name name-style="western"><surname>Gaur</surname><given-names>M</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Muresan</surname><given-names>S</given-names> </name><name name-style="western"><surname>Nakov</surname><given-names>P</given-names> </name><name name-style="western"><surname>Villavicencio</surname><given-names>A</given-names> </name></person-group><article-title>A risk-averse mechanism for suicidality assessment on social media</article-title><year>2022</year><conf-name>Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)</conf-name><conf-date>May 22-27, 2022</conf-date><conf-loc>Dublin, Ireland</conf-loc><fpage>628</fpage><lpage>635</lpage><pub-id pub-id-type="doi">10.18653/v1/2022.acl-short.70</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Rawat</surname><given-names>BPS</given-names> </name><name name-style="western"><surname>Kovaly</surname><given-names>S</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Pigeon</surname><given-names>W</given-names> </name></person-group><article-title>ScAN: suicide attempt and ideation events dataset</article-title><conf-name>Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics</conf-name><conf-date>Jul 10-15, 2022</conf-date><conf-loc>Seattle, United States</conf-loc><fpage>1029</fpage><pub-id pub-id-type="doi">10.18653/v1/2022.naacl-main.75</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Sawhney</surname><given-names>R</given-names> </name><name name-style="western"><surname>Joshi</surname><given-names>H</given-names> </name><name name-style="western"><surname>Shah</surname><given-names>RR</given-names> </name><name name-style="western"><surname>Flek</surname><given-names>L</given-names> </name></person-group><article-title>Suicide ideation detection via social and temporal user representations using hyperbolic learning</article-title><year>2021</year><conf-name>Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics</conf-name><conf-date>Jun 6-11, 2021</conf-date><conf-loc>Online</conf-loc><fpage>2176</fpage><lpage>2190</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://aclanthology.org/2021.naacl-main">https://aclanthology.org/2021.naacl-main</ext-link></comment><pub-id pub-id-type="doi">10.18653/v1/2021.naacl-main.176</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Marchant</surname><given-names>A</given-names> </name><name name-style="western"><surname>Hawton</surname><given-names>K</given-names> </name><name name-style="western"><surname>Stewart</surname><given-names>A</given-names> </name><etal/></person-group><article-title>A systematic review of the relationship between internet use, self-harm and suicidal behaviour in young people: the good, the bad and the unknown</article-title><source>PLoS One</source><year>2017</year><volume>12</volume><issue>8</issue><fpage>e0181722</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0181722</pub-id><pub-id pub-id-type="medline">28813437</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sedgwick</surname><given-names>R</given-names> </name><name name-style="western"><surname>Epstein</surname><given-names>S</given-names> </name><name name-style="western"><surname>Dutta</surname><given-names>R</given-names> </name><name name-style="western"><surname>Ougrin</surname><given-names>D</given-names> </name></person-group><article-title>Social media, internet use and suicide attempts in adolescents</article-title><source>Curr Opin Psychiatry</source><year>2019</year><volume>32</volume><issue>6</issue><fpage>534</fpage><lpage>541</lpage><pub-id pub-id-type="doi">10.1097/YCO.0000000000000547</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>ZZ</given-names> </name><name name-style="western"><surname>Jia</surname><given-names>CX</given-names> </name></person-group><article-title>Digital media use and subsequent self-harm during a 1-year follow-up of Chinese adolescents</article-title><source>J Affect Disord</source><year>2020</year><month>12</month><volume>277</volume><fpage>279</fpage><lpage>286</lpage><pub-id pub-id-type="doi">10.1016/j.jad.2020.05.066</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Patchin</surname><given-names>JW</given-names> </name><name name-style="western"><surname>Hinduja</surname><given-names>S</given-names> </name><name name-style="western"><surname>Meldrum</surname><given-names>RC</given-names> </name></person-group><article-title>Digital self-harm and suicidality among adolescents</article-title><source>Child Adolesc Ment Health</source><year>2023</year><month>02</month><volume>28</volume><issue>1</issue><fpage>52</fpage><lpage>59</lpage><pub-id pub-id-type="doi">10.1111/camh.12574</pub-id><pub-id pub-id-type="medline">35811440</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Choi</surname><given-names>WS</given-names> </name><name name-style="western"><surname>Han</surname><given-names>J</given-names> </name><name name-style="western"><surname>Hong</surname><given-names>HJ</given-names> </name></person-group><article-title>Association Between internet searches related to suicide/self-harm and adolescent suicide death in South Korea in 2016-2020: secondary data analysis</article-title><source>J Med Internet Res</source><year>2023</year><month>04</month><day>20</day><volume>25</volume><fpage>e46254</fpage><pub-id pub-id-type="doi">10.2196/46254</pub-id><pub-id pub-id-type="medline">37079349</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mars</surname><given-names>B</given-names> </name><name name-style="western"><surname>Heron</surname><given-names>J</given-names> </name><name name-style="western"><surname>Biddle</surname><given-names>L</given-names> </name><etal/></person-group><article-title>Exposure to, and searching for, information about suicide and self-harm on the Internet: prevalence and predictors in a population based cohort of young adults</article-title><source>J Affect Disord</source><year>2015</year><month>10</month><volume>185</volume><fpage>239</fpage><lpage>245</lpage><pub-id pub-id-type="doi">10.1016/j.jad.2015.06.001</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>S</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>C</given-names> </name><name name-style="western"><surname>Zeng</surname><given-names>M</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Moens</surname><given-names>MF</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Specia</surname><given-names>L</given-names> </name><name name-style="western"><surname>Yih</surname><given-names>S t.</given-names> </name></person-group><article-title>Want to reduce labeling cost? GPT-3 can help</article-title><year>2021</year><conf-name>Findings of the Association for Computational Linguistics</conf-name><conf-date>Nov 7-11, 2021</conf-date><conf-loc>Punta Cana, Dominican Republic</conf-loc><fpage>4195</fpage><lpage>4205</lpage><pub-id pub-id-type="doi">10.18653/v1/2021.findings-emnlp.354</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>H</given-names> </name><name name-style="western"><surname>Rahman</surname><given-names>S</given-names> </name><name name-style="western"><surname>Mitra</surname><given-names>K</given-names> </name><name name-style="western"><surname>Miao</surname><given-names>Z</given-names> </name></person-group><article-title>Human-LLM collaborative annotation through effective verification of LLM labels</article-title><year>2024</year><month>05</month><day>11</day><conf-name>Proceedings of the CHI Conference on Human Factors in Computing Systems</conf-name><conf-date>May 11-16, 2024</conf-date><conf-loc>Honolulu HI USA</conf-loc><fpage>1</fpage><lpage>21</lpage><pub-id pub-id-type="doi">10.1145/3613904.3641960</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="web"><article-title>&#xC628;&#xB77C;&#xC778; &#xC790;&#xC0B4;&#xC720;&#xBC1C;&#xC815;&#xBCF4;, &#xAD6D;&#xBBFC;&#xC774; &#xC9C1;&#xC811; &#xCC3E;&#xC544;&#xB0B4;&#xACE0; &#xC0AD;&#xC81C;&#xD55C;&#xB2E4;! [Citizens directly identify and remove online suicide- inducing information!]</article-title><source>KPHN</source><year>2020</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.mohw.go.kr/board.es?mid=a10503010100&#x0026;bid=0027&#x0026;act=view&#x0026;list_no=359157">https://www.mohw.go.kr/board.es?mid=a10503010100&#x0026;bid=0027&#x0026;act=view&#x0026;list_no=359157</ext-link></comment></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Markov</surname><given-names>T</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>C</given-names> </name><name name-style="western"><surname>Agarwal</surname><given-names>S</given-names> </name><etal/></person-group><article-title>A holistic approach to undesired content detection in the real world</article-title><source>AAAI</source><year>2023</year><volume>37</volume><issue>12</issue><fpage>15009</fpage><lpage>15018</lpage><pub-id pub-id-type="doi">10.1609/aaai.v37i12.26752</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>D</given-names> </name><name name-style="western"><surname>Park</surname><given-names>S</given-names> </name><name name-style="western"><surname>Kang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Choi</surname><given-names>D</given-names> </name><name name-style="western"><surname>Han</surname><given-names>J</given-names> </name></person-group><article-title>Cross-lingual suicidal-oriented word embedding toward suicide prevention</article-title><year>2020</year><month>11</month><conf-name>Findings of the Association for Computational Linguistics</conf-name><conf-date>Nov 16-20, 2020</conf-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.aclweb.org/anthology/2020.findings-emnlp">https://www.aclweb.org/anthology/2020.findings-emnlp</ext-link></comment><pub-id pub-id-type="doi">10.18653/v1/2020.findings-emnlp.200</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="web"><article-title>&#xBBF8;&#xB514;&#xC5B4; &#xC790;&#xC0B4;&#xC815;&#xBCF4; &#xBAA8;&#xB2C8;&#xD130;&#xB9C1; &#xC2DC;&#xC2A4;&#xD15C; [<sc>M</sc>edia suicide information monitoring system]</article-title><source>KFSP</source><year>2023</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://sims.kfsp.or.kr/">https://sims.kfsp.or.kr/</ext-link></comment></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Park</surname><given-names>S</given-names> </name><name name-style="western"><surname>Park</surname><given-names>K</given-names> </name><name name-style="western"><surname>Ahn</surname><given-names>J</given-names> </name><name name-style="western"><surname>Oh</surname><given-names>A</given-names> </name></person-group><article-title>Suicidal risk detection for military personnel</article-title><conf-name>Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</conf-name><conf-date>Nov 16-20, 2020</conf-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.aclweb.org/anthology/2020.emnlp-main">https://www.aclweb.org/anthology/2020.emnlp-main</ext-link></comment><pub-id pub-id-type="doi">10.18653/v1/2020.emnlp-main.198</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="web"><article-title>&#xC0AC;&#xC774;&#xBC84;&#xC0C1;&#xB2F4; [cyber counselling]</article-title><source>Lifeline Korea</source><year>2023</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.lifeline.or.kr/business/opening_consult.php">https://www.lifeline.or.kr/business/opening_consult.php</ext-link></comment></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="web"><article-title>&#xACF5;&#xAC1C;&#xC0C1;&#xB2F4;&#xC2E4; [public counselling room]</article-title><source>KSPCC</source><year>2023</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="http://auru.counselling.or.kr/contents/sub0201.php?PHPSESSID=4f99b850a26d94de69fe936d68006e51">http://auru.counselling.or.kr/contents/sub0201.php?PHPSESSID=4f99b850a26d94de69fe936d68006e51</ext-link></comment></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="web"><article-title>&#xB514;&#xC2DC;&#xC778;&#xC0AC;&#xC774;&#xB4DC; [DCinside]</article-title><source>DCInside</source><year>2023</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.dcinside.com/">https://www.dcinside.com/</ext-link></comment></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Jo</surname><given-names>J</given-names> </name></person-group><article-title>&#xC787;&#xB2E8; &#xC0AC;&#xACE0;&#xC5D0; &#xB9D0; &#xB9CE;&#xC740; &#x2019;&#xC6B0;&#xC6B8;&#xC99D; &#xAC24;&#xB7EC;&#xB9AC;&#x2019;&#x2025;"&#xD3D0;&#xC1C4;&#xB294; &#xC548; &#xD55C;&#xB2E4;"&#xB294; &#xC815;&#xBD80; [following a series of incidents, the government states it will not shut down the controversial &#x2019;depression gallery&#x2019; despite the ongoing discussions]</article-title><source>MBC</source><year>2023</year><access-date>2026-01-29</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://imnews.imbc.com/replay/2023/nwdesk/article/6490952_36199.html">https://imnews.imbc.com/replay/2023/nwdesk/article/6490952_36199.html</ext-link></comment></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fiesler</surname><given-names>C</given-names> </name><name name-style="western"><surname>Jiang</surname><given-names>J</given-names> </name><name name-style="western"><surname>McCann</surname><given-names>J</given-names> </name><name name-style="western"><surname>Frye</surname><given-names>K</given-names> </name><name name-style="western"><surname>Brubaker</surname><given-names>JR</given-names> </name></person-group><article-title>Reddit Rules! Characterizing an Ecosystem of Governance</article-title><source>ICWSM</source><year>2018</year><volume>12</volume><issue>1</issue><pub-id pub-id-type="doi">10.1609/icwsm.v12i1.15033</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Moon</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>DH</given-names> </name><name name-style="western"><surname>Cho</surname><given-names>H</given-names> </name><etal/></person-group><article-title>Analyzing norm violations in live-stream chat</article-title><conf-name>Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</conf-name><conf-date>Dec 6-10, 2023</conf-date><conf-loc>Singapore</conf-loc><fpage>852</fpage><lpage>868</lpage><pub-id pub-id-type="doi">10.18653/v1/2023.emnlp-main.55</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gattrell</surname><given-names>WT</given-names> </name><name name-style="western"><surname>Hungin</surname><given-names>AP</given-names> </name><name name-style="western"><surname>Price</surname><given-names>A</given-names> </name><etal/></person-group><article-title>ACCORD guideline for reporting consensus-based methods in biomedical research and clinical practice: a study protocol</article-title><source>Res Integr Peer Rev</source><year>2022</year><month>06</month><day>7</day><volume>7</volume><issue>1</issue><fpage>3</fpage><pub-id pub-id-type="doi">10.1186/s41073-022-00122-0</pub-id><pub-id pub-id-type="medline">35672782</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Vakil</surname><given-names>N</given-names> </name></person-group><article-title>Consensus guidelines: method or madness?</article-title><source>Am J Gastroenterol</source><year>2011</year><month>02</month><volume>106</volume><issue>2</issue><fpage>225</fpage><lpage>227</lpage><pub-id pub-id-type="doi">10.1038/ajg.2010.504</pub-id><pub-id pub-id-type="medline">21301451</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>NF</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>K</given-names> </name><name name-style="western"><surname>Hewitt</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Lost in the middle: how language models use long contexts</article-title><source>Trans Assoc Comput Linguist</source><year>2024</year><month>02</month><day>23</day><volume>12</volume><fpage>157</fpage><lpage>173</lpage><pub-id pub-id-type="doi">10.1162/tacl_a_00638</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Iso</surname><given-names>H</given-names> </name><name name-style="western"><surname>Pezeshkpour</surname><given-names>P</given-names> </name><name name-style="western"><surname>Bhutani</surname><given-names>N</given-names> </name><name name-style="western"><surname>Hruschka</surname><given-names>E</given-names> </name></person-group><article-title>Less is more for long document summary evaluation by LLMs</article-title><year>2023</year><conf-name>Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 2</conf-name><conf-date>Mar 17-22, 2024</conf-date><conf-loc>St Julian&#x2019;s, Malta</conf-loc><fpage>330</fpage><lpage>343</lpage><pub-id pub-id-type="doi">10.18653/v1/2024.eacl-short.29</pub-id></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Zhao</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Wallace</surname><given-names>E</given-names> </name><name name-style="western"><surname>Feng</surname><given-names>S</given-names> </name><name name-style="western"><surname>Klein</surname><given-names>D</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>S</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Meila</surname><given-names>M</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>T</given-names> </name></person-group><article-title>Calibrate before use: improving few-shot performance of language models</article-title><access-date>2025-12-20</access-date><conf-name>Proceedings of the 38th International Conference on Machine Learning</conf-name><conf-date>Jul 18-24, 2021</conf-date><conf-loc>Virtual</conf-loc><fpage>12697</fpage><lpage>12706</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://proceedings.mlr.press/v139/zhao21c.html">https://proceedings.mlr.press/v139/zhao21c.html</ext-link></comment></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Brown</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Language models are few-shot learners</article-title><source>Adv Neural Inf Process Syst</source><year>2020</year><access-date>2025-12-20</access-date><volume>33</volume><fpage>1877</fpage><lpage>1901</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://papers.nips.cc/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html">https://papers.nips.cc/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html</ext-link></comment></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Kim</surname><given-names>B</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>H</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>SW</given-names> </name><etal/></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Moens</surname><given-names>MF</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Specia</surname><given-names>L</given-names> </name><name name-style="western"><surname>Yih</surname><given-names>S t.</given-names> </name></person-group><article-title>What changes can large-scale language models bring? intensive study on hyperclova: billions-scale Korean generative pretrained transformers</article-title><conf-name>Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing</conf-name><conf-date>Nov 7-11, 2021</conf-date><conf-loc>Online and Punta Cana, Dominican Republic</conf-loc><fpage>3405</fpage><lpage>3424</lpage><pub-id pub-id-type="doi">10.18653/v1/2021.emnlp-main.274</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Tunstall</surname><given-names>L</given-names> </name><etal/></person-group><article-title>Zephyr: direct distillation of LM alignment</article-title><source>arXiv</source><comment>Preprint posted online on  Oct 25, 2023</comment><pub-id pub-id-type="doi">10.48550/arXiv.2310.16944</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>D</given-names> </name><etal/></person-group><article-title>How long can open-source llms truly promise on context length</article-title><source>LMSYS Org</source><year>2023</year><access-date>2025-12-29</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://lmsys.org/blog/2023-06-29-longchat/">https://lmsys.org/blog/2023-06-29-longchat/</ext-link></comment></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Chiang</surname><given-names>WL</given-names> </name><etal/></person-group><article-title>Vicuna: an open-source chatbot impressing GPT-4 with 90%* ChatGPT quality</article-title><source>LMSYS Org</source><year>2023</year><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://lmsys.org/blog/2023-03-30-vicuna/">https://lmsys.org/blog/2023-03-30-vicuna/</ext-link></comment></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Kocmi</surname><given-names>T</given-names> </name><name name-style="western"><surname>Federmann</surname><given-names>C</given-names> </name></person-group><article-title>Large language models are state-of-the-art evaluators of translation quality</article-title><source>arXiv</source><comment>Preprint posted online on  Feb 28, 2023</comment><pub-id pub-id-type="doi">10.48550/arXiv.2302.14520</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hawton</surname><given-names>K</given-names> </name><name name-style="western"><surname>Saunders</surname><given-names>KEA</given-names> </name><name name-style="western"><surname>O&#x2019;Connor</surname><given-names>RC</given-names> </name></person-group><article-title>Self-harm and suicide in adolescents</article-title><source>Lancet</source><year>2012</year><month>06</month><day>23</day><volume>379</volume><issue>9834</issue><fpage>2373</fpage><lpage>2382</lpage><pub-id pub-id-type="doi">10.1016/S0140-6736(12)60322-5</pub-id><pub-id pub-id-type="medline">22726518</pub-id></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Skegg</surname><given-names>K</given-names> </name></person-group><article-title>Self-harm</article-title><source>Lancet</source><year>2005</year><volume>366</volume><issue>9495</issue><fpage>1471</fpage><lpage>1483</lpage><pub-id pub-id-type="doi">10.1016/S0140-6736(05)67600-3</pub-id><pub-id pub-id-type="medline">16243093</pub-id></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Shing</surname><given-names>HC</given-names> </name><name name-style="western"><surname>Nair</surname><given-names>S</given-names> </name><name name-style="western"><surname>Zirikly</surname><given-names>A</given-names> </name><name name-style="western"><surname>Friedenberg</surname><given-names>M</given-names> </name><name name-style="western"><surname>Daum&#x00E9; III</surname><given-names>H</given-names> </name><name name-style="western"><surname>Resnik</surname><given-names>P</given-names> </name></person-group><article-title>Expert, crowdsourced, and machine assessment of suicide risk via online postings</article-title><year>2018</year><conf-name>Proceedings of the Fifth Workshop on Computational Linguistics and Clinical Psychology: From Keyboard to Clinic</conf-name><conf-date>Jun 5, 2018</conf-date><conf-loc>New Orleans, LA</conf-loc><fpage>25</fpage><lpage>36</lpage><pub-id pub-id-type="doi">10.18653/v1/W18-0603</pub-id></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Choshen</surname><given-names>L</given-names> </name><name name-style="western"><surname>Eldad</surname><given-names>D</given-names> </name><name name-style="western"><surname>Hershcovich</surname><given-names>D</given-names> </name><name name-style="western"><surname>Sulem</surname><given-names>E</given-names> </name><name name-style="western"><surname>Abend</surname><given-names>O</given-names> </name></person-group><article-title>The language of legal and illegal activity on the darknet</article-title><year>2019</year><conf-name>Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</conf-name><conf-date>Jul 28 to Aug 2, 2019</conf-date><conf-loc>Florence, Italy</conf-loc><fpage>4271</fpage><lpage>4279</lpage><pub-id pub-id-type="doi">10.18653/v1/P19-1419</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Bak</surname><given-names>J</given-names> </name></person-group><article-title>Iterative large language model-guided sampling and expert-annotated benchmark corpus for harmful suicide content detection: development and validation study</article-title><source>GitHub</source><access-date>2025-12-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://github.com/Human-Language-Intelligence/Harmful-Suicide-Content-Detection">https://github.com/Human-Language-Intelligence/Harmful-Suicide-Content-Detection</ext-link></comment></nlm-citation></ref></ref-list></back></article>