There are an estimated 800,000 suicides per year globally, and approximately 16,000,000 suicide attempts. Mobile apps may help address the unmet needs of people at risk. We assessed adherence of suicide prevention advice in depression management and suicide prevention apps to six evidence-based clinical guideline recommendations: mood and suicidal thought tracking, safety plan development, recommendation of activities to deter suicidal thoughts, information and education, access to support networks, and access to emergency counseling.
A systematic assessment of depression and suicide prevention apps available in Google Play and Apple’s App Store was conducted. Apps were identified by searching 42matters in January 2019 for apps launched or updated since January 2017 using the terms “depression,” “depressed,” “depress,” “mood disorders,” “suicide,” and “self-harm.” General characteristics of apps, adherence with six suicide prevention strategies identified in evidence-based clinical guidelines using a 50-question checklist developed by the study team, and trustworthiness of the app based on HONcode principles were appraised and reported as a narrative review, using descriptive statistics.
The initial search yielded 2690 potentially relevant apps. Sixty-nine apps met inclusion criteria and were systematically assessed. There were 20 depression management apps (29%), 3 (4%) depression management and suicide prevention apps, and 46 (67%) suicide prevention apps. Eight (12%) depression management apps were chatbots. Only 5/69 apps (7%) incorporated all six suicide prevention strategies. Six apps (6/69, 9%), including two apps available in both app stores and downloaded more than one million times each, provided an erroneous crisis helpline number. Most apps included emergency contact information (65/69 apps, 94%) and direct access to a crisis helpline through the app (46/69 apps, 67%).
Non-existent or inaccurate suicide crisis helpline phone numbers were provided by mental health apps downloaded more than 2 million times. Only five out of 69 depression and suicide prevention apps offered all six evidence-based suicide prevention strategies. This demonstrates a failure of Apple and Google app stores, and the health app industry in self-governance, and quality and safety assurance. Governance levels should be stratified by the risks and benefits to users of the app, such as when suicide prevention advice is provided.
In 2016, there were an estimated 800,000 suicides globally, and approximately 16,000,000 suicide attempts . Despite reports of almost 33% decrease in the global age-standardized mortality rate from suicide between 1990 and 2016 [2, 3], suicide remains one of the leading causes of preventable deaths in both developing and developed countries. Sixty percent of individuals with suicidal ideation transitioned to a first attempt within a year of onset , a significant figure considering that half to two thirds of suicide deaths occur in the first attempt [5, 6]. The risk of suicide increases with access to means of suicide, personal or family history of mental health disorders and suicide attempts, and psychiatric comorbidity. Over 90% of people who died from suicide were affected by depression, alcohol abuse, or both [5, 7]. Suicide prevention programs targeting one or more of these factors successfully decrease the number of suicides [8, 9]. An important and widespread component of suicide prevention strategies are crisis helplines, which provide timely and anonymous advice to callers at current risk of suicide and are effective in deterring active suicidal thoughts [10,11,12].
Timely identification of persons at risk of suicide is critical to ensure adequate provision of care. Family physicians (FPs) play an important role as most individuals who died by suicide visited their FP in the month preceding death , and about 90% consulted their FP several times the prior year . Nevertheless, efforts by healthcare providers to identify patients at risk face significant hurdles, particularly the unwillingness of affected individuals to disclose suicidality fearing loss of autonomy, overreaction, and stigma [14, 15].
Forty percent of people with suicidal thoughts or behaviors do not seek medical care , or may not have access to healthcare, particularly in developing countries. Digital interventions delivered online or through mobile devices may increase access to help and mental health care. Patients feel more at ease discussing mental health conditions online than in a face-to-face encounter , and consider the Internet accessible, affordable, and convenient .
Over the last decade, the health app market has grown to include about 318,000 apps , of which more than 10,000 are mental health apps , making selection of an appropriate app cumbersome, particularly for lay users . Digital mental health interventions seem to offer a promising alternative to face-to-face visits [22, 23]. However, very few apps available in app stores have been evaluated in clinical trials [24,25,26,27,28] or by regulatory bodies like the FDA .
Previous research on the use of digital health for suicide prevention focused on highlighting features of an ideal intervention , systematically reviewing the effectiveness of online interventions and mobile apps [31, 32], app store descriptions of apps , or assessing suicide prevention strategies offered by apps . However, none of these studies evaluated suicide prevention advice offered by apps. Given this and the high turnover of apps , we conducted a comprehensive assessment of suicide prevention apps available on Google Play and Apple’s App Store worldwide, as well as assessing the suicide prevention advice offered by depression management apps.
The aims of this study were:
To systematically assess depression and suicide prevention apps’ adherence to evidence-based clinical guidelines on:
Strategies for suicide prevention;
Type and quality of advice given when the user is at risk of attempting suicide; and
To analyze the response of chatbot apps to a user who appears to be at risk of attempting suicide using simulated patient scenarios.
Systematic review methodology was adapted for the app search, selection, assessment, and data analysis.
A systematic search on Apple’s App Store and Google Play using the 42matters (https://42matters.com/) was performed in January 2019 using the terms “depression,” “depressive,” “depress,” “mood disorders,” “suicide,” and “self-harm.” The search was limited to medical, lifestyle, health and fitness, and education categories, with no country restrictions. The search engine retrieved the name, category, developer, app store description, date of first release and current version, ratings and number of raters (for iOS only), link to website, and market URL for each app.
App targets people suffering from depression; or
Assesses suicide risk; or
Provides advice to prevent users from attempting suicide; or
Follows a “call to action” model. We defined “call to action” as a message delivered by the app using active language and addressed to the user inviting her/him to take action to prevent the urge of hurting her/himself, for example “If you feel suicidal, please call the following number…”; or
Provides a link for the user to activate a phone call to a crisis helpline directly through the app
App has been uploaded or updated from 1st January 2017 onwards
App is free or requires payment to download/use and is available in Apple’s App Store or Google Play
App is in English
App is aimed at healthcare providers (physicians, psychologists, counselors, others), or the support network of the person at risk of suicide or community gatekeepers
Offers teleconsultation services with physicians, psychologists, counselors, or other healthcare providers
App content is not interactive (e.g., books, music playlists, wallpapers, others), does not ask the user to act, or does not provide a direct link to a crisis helpline through the app
App refers to self-harm with non-suicidal intent
App consists of a standalone depression screening questionnaire
App was removed from the app stores at the time of download, required a sign-up code provided by an institution, or could not be used after two attempts due to technical problems
The app selection process is presented as a flowchart  (Fig. 1).
Development of the assessment criteria
The assessment criteria were developed by the research team and comprised three main components (Additional file 2: Table S2):
General attributes of the app, including cost and ratings, target user groups, data security measures adopted to ensure user’s privacy, app crashes or malfunction, and who developed the app.
Strategies offered by the app to prevent or manage suicidality in a person at risk, based on evidence-based clinical guidelines (as a prerequisite for their potential of effectiveness) from the UK [37, 38], USA [11, 39], and WHO . The criteria comprised 50 questions organized in six domains:
Tracking of mood and suicidal thoughts, to assess acute risk of suicide, including users’ mood, triggers for suicidal thoughts, suicide plans and protective factors (reasons for living, plans for the future, coping or problem-solving skills)
Development of a safety plan, defined as a structured, six-step, standardized list of strategies and contact details of members of his/her support network that a person at risk of suicide can use during a crisis .
Recommendation of activities to deter suicidal thoughts, and follow-up on outcomes.
Information and education, educational articles on signs of suicidality, risk factors and triggers of suicide, and safety planning. Information included lists of crisis helpline numbers or emergency contact information
Access to support networks, including saving the contact information of people from the user’s support network (family, friends, and primary healthcare provider) and ability to share information with them
Access to emergency counseling provided by a healthcare professional or a crisis helpline the user can contact directly through the app, or through a chatbot, e.g.,
Trustworthiness of information provided by the app was adapted from the Health on the Net Foundation Code of Conduct (HONcode)  that evaluates the reliability of information based on citations, justification of claims, and authority of information, as well as for adherence to ethical standards of transparency, privacy, and advertising policies.
We followed a systematic, two-step process to select apps for inclusion. First, two investigators (LM and MK) screened app store descriptions of all retrieved apps in parallel. Included apps were then downloaded and screened again according to predefined inclusion and exclusion criteria. Any uncertainty regarding the inclusion or exclusion of apps was resolved by discussion between the assessors. About 20% of apps were assessed by both researchers (LM and MK) to ensure consistent application of assessment criteria, after which the remaining apps were assessed by either one of the researchers. Interrater reliability for apps assessed in parallel by both assessors was calculated using Cohen’s kappa (κ). The assessment was considered reliable if the interrater agreement was equal to or higher than 0.6 (substantial or almost perfect agreement) .
Apps were assessed using an iPhone 5c (iOS 10.3.3) or iPhone 7 (iOS 11.4.1) and a Sony XPERIA XZs (Android 8.0.0). For apps available on both platforms, both versions were assessed to account for potential differences in app functionalities, and counted as an individual app. We assessed paid and free apps without add-ons available as in-app purchases.
To further ensure consistency in the assessment process, we created a user persona which included demographics, medical diagnosis, potential answers to self-reported questionnaires, and opening statements to converse with a chatbot-based app.
Descriptive statistics were used to analyze the data. To compare the functionalities and trustworthiness of depression management and suicide prevention apps, a significance test for categorical variables was used: chi-square test if each category contained more than ten variables and two-tailed Fisher’s exact test if any of the categories in the contingency table were below ten. Statistical significance was set at p < 0.05.
The search strategy retrieved 2591 apps (1606 Android and 985 iOS) after duplicates were removed. Screening yielded 102 apps, of which 69 apps met inclusion criteria and were systematically assessed (9 Android apps, 10 iOS apps, and 25 apps available on both platforms). Twenty-three percent (23/102) of apps were assessed by both assessors with substantial agreement (κ = 0.730 (95% CI, .700 to .759), p < .0005). Figure 1 summarizes the app search and selection process. Additional file 1: Table S1 lists all the assessed apps and the suicide prevention strategies offered by each app.
General attributes of apps
The analysis included 20/69 (29%) depression management apps, 46/69 (67%) suicide prevention apps, and 3/69 (4%) apps offering depression management and suicide prevention. Table 1 provides a summary of app attributes. A total of 7/20 (35%) Android depression management apps, including three chatbots, were downloaded > 50,000 times, of which three apps (Moodpath , Wysa , and Youper ) were downloaded > 500,000 times and two apps (7 Cups  and Pacifica ) were downloaded > 1,000,000 times. An iOS version was available for these apps. Conversely, the number of downloads for suicide prevention apps ranged from 5 to > 10,000.
Most apps were free to download and use. Ten (50%) depression management apps and two (67%) depression management and suicide prevention apps offered in-app purchases for cognitive behavioral therapy-based programs, access to a health provider, or a workplace wellness program. One suicide prevention app included in-app payment for voluntary donations. Forty-five percent of all apps were created or commissioned by healthcare providers or non-profit organizations. In addition, 19/46 (41%) suicide prevention apps were created or commissioned by a government organization or university.
Fifty-one percent of all apps were categorized as “Health and Fitness” in their app store descriptions, while less than 20% of apps were categorized as “Medical.” Educational apps accounted for 18/46 (39%) suicide prevention apps.
Strategies to manage a person at risk of suicide
Most apps included at least three suicide prevention strategies (see Tables 1 and 2), more commonly emergency contact information (65/69 apps, 94%), direct access to a crisis helpline (46/69 apps, 67%), and suicide-related education (35/69 apps, 51%). A total of 5/69 apps (7%) offered all six strategies. Table 3 presents examples of apps complying with all suicide prevention strategies. Additional file 1: Table S1 provides a detailed description of the strategies used by each app. A description of findings for each strategy is provided below.
Tracking of mood and suicidal thoughts
Seventeen depression management apps (17/20, 85%), 1/3 (33%) depression management and suicide prevention app, and 10/46 (22%) suicide prevention apps tracked users’ mood or suicidal behavior. Depression management apps assessed users’ mood using self-developed questions or a validated questionnaire (Patient Health Questionnaire-9 (PHQ-9) ), while eight chatbot apps also assessed users’ suicidal behavior. Conversely, five suicide prevention apps assessed users’ mood and seven assessed users’ suicidal thoughts or behaviors. None of the apps enquired about risk factors, triggers, or protective factors, and only one checked past history of suicide.
Safety plan development
Only 2/20 (10%) depression management apps and 24/46 (52%) suicide prevention apps offered users a template to develop a safety plan, and all but one included guidance to complete the safety plan. Only 11 apps included all safety plan steps as developed by Stanley and Brown . The most common missing steps were a list of activities to deter suicidal thoughts and access to users’ support network. In seven apps, the safety plan was one component in a more comprehensive suicide prevention strategy that included educational articles, mood and suicidality assessment, and access to support network and crisis helplines. Only four apps allowed the user to share the safety plan with a member of his/her support network.
Recommendation of activities to deter suicidal thoughts
Fifteen depression management apps (15/20, 75%), 1/3 (33%) depression management and suicide prevention app, and 17/46 (37%) suicide prevention apps offered activities aimed to enhance wellbeing, improve mood, or discourage suicidal thoughts, including mindfulness, or another meditation technique, hobbies or outdoor activities, exercise, and healthy lifestyle advice.
Information and education
Two depression management apps (2/20, 10%), 3/3 (100%) depression management and suicide prevention apps, and 30/46 (65%) suicide prevention apps provided information on suicide signs, triggers, risk factors and prevention strategies, and how to complete a safety plan. Furthermore, all except two suicide prevention apps providing access to users’ support network and available on both platforms (n = 4, 9%) provided emergency contact information including crisis helpline telephone numbers, messaging service numbers, or links to relevant websites. The information in 49 apps was specific for one or several countries, limiting its global usability.
Access to support networks
Two depression management apps (2/20, 10%), 1/3 (33%) depression management and suicide prevention apps, and 25/46 (54%) suicide prevention apps allowed users to store members of their support network’s contact details. In 15 apps, including two depression management apps and 13 suicide prevention apps, this functionality was included in a safety plan.
Fourteen suicide prevention apps (14/46, 30%) allowed users to contact members of their support network directly from the app. A subset of these apps (8/46 apps, 17%) used a simple interface aimed exclusively at facilitating immediate, often simultaneous communication with one or several support network members via a telephone call or text message.
Access to emergency counseling
Nine suicide prevention apps (9/46, 20%) provided emergency access to trained counselors directly through the app. All but one app were developed by public institutions or non-governmental organizations (NGOs). Three apps specifically targeted veterans and three apps, university students.
In total, 11/20 (24%) depression management apps, 1/3 (33%) depression management and suicide prevention app, and 36/46 (78%) suicide prevention apps offered direct contact to a crisis helpline through the app. The accuracy and functionality of crisis helpline numbers provided by the apps was verified by performing an online search and found to be faulty in six (9%), four depression management, and two suicide prevention apps (Table 4).
Depression management apps were significantly more compliant than suicide prevention apps in indicating the qualifications of people involved in app development, and backing up effectiveness claims with evidence published in peer-reviewed journals or claimed to be in the process of analyzing research data. Few apps (10/69, 14%), across all categories, cited the sources of information offered in the app (Table 5).
Eight apps (8/69, 12%) included artificial intelligence-powered chatbots. Three additional apps (two iOS and one Android app) offered fixed, predetermined advice using a chatbot-style format and were not included in our analysis. In two of these apps, the chatbot was one of the features offered by the app, while in the other six apps the chatbot was the main component.
Chatbots offered advice and self-improvement strategies to users suffering from depression and other mental health disorders and they were able to tailor their advice to users’ responses. All chatbots initiated a conversation reminding the user they should not use the app if they were feeling suicidal. None of the chatbots identified “I am very sad and hopeless” as a worrying statement that may require follow-up questions. All chatbots responded to “I just feel like dying now” by seeking confirmation from users that they were having suicidal thoughts and providing access to crisis helplines.
A systematic assessment of 69 depression management and suicide prevention apps revealed that only five apps offered all six evidence-based strategies for suicide prevention, with comprehensive and holistic support. Most apps offered users up to three preventive strategies, particularly crisis helplines contact information and/or a direct connection through the app. Other evidence-based strategies differed: depression management apps assessed users’ mood and listed activities to improve mood when feeling distressed, and suicide prevention apps provided safety plan templates and multimedia educational material.
Several studies appraising the quality of health apps consistently indicated that most apps do not provide evidence-based information or decision-support strategies and may not be safe to use [51,52,53]. Appraisals of mental health and suicide prevention apps showed similar results [34, 54]. Larsen et al.  in their assessment of 49 suicide prevention apps available in Australian app stores, reported a small number of potentially harmful apps, while all apps offered at least one evidence-based intervention, an outcome aligned with our findings. Only 6/49 apps were also included in our assessment, demonstrating high turnover of apps and distinctive availability in different countries’ app stores. Similarly, De la Torre et al.  reported a systematic literature review and appraisal of app store descriptions of suicide prevention apps in Spanish app stores, retrieving 20 apps, six of which were also included in our assessment.
Most apps targeted only one aspect of suicide prevention, based on strategies recommended by evidence-based clinical guidelines [11, 37,38,39, 55], and hence, may be inadequate and potentially dangerous if used as a standalone intervention. Managing persons at risk of suicide is complex and requires collaborative partnership between the affected person and her/his support network, and a multidisciplinary healthcare team . Mobile apps could offer tools for real-time monitoring of at-risk persons and access to support whenever it is needed; however, apps should be seen as an addition to an ongoing patient-provider relationship and never as a replacement.
Six apps contained erroneous crisis helpline numbers, posing a potentially serious risk for users. Although the impact of apps on decreasing suicide deaths is difficult to assess, crisis helplines are an important component of suicide prevention strategies [11, 12] and play a role in decreasing callers’ immediate risk of suicide [10, 56]. Our findings show information may not be corroborated and clearly demonstrate the lack of self-regulation and self-monitoring of the industry. Crisis helplines are readily available in a variety of platforms and can easily be verified by developers and app stores before apps are launched. That apps containing non-existent/inaccurate crisis helplines are on the market shows that the review mechanisms that should be in place to detect errors are either inadequate or lacking.
Half of the apps belonged to the “Health and Fitness” category while apps categorized as “Medical” accounted for only 20%. The current review and approval processes established by the app stores prior to the launch of a new app do not prevent poor-quality apps from being released [57, 58]. Furthermore, app developers seem to select an app category according to business models and marketing strategies, with no transparency or real oversight on such decisions. While this app development model may work best for less sensitive categories, health apps require appropriate evaluation of content alongside the technical aspects of the app.
There are currently no consequences for releasing health apps containing inaccurate or non-evidence-based information. Systematic app assessments consistently report serious flaws that may affect users’ health and wellbeing [51, 52]. At the same time, there are increasing calls to improve health app oversight, from independent expert assessments and app libraries [59, 60], to higher standard of app development and quality assurance mechanisms, such as (voluntary or compulsory) certification or regulation prior to app release to the public [61, 62]. App libraries, such as Psyberguide  or the new NHS Apps Library , provide a curated, although very limited collection of apps for users to choose from, while official regulatory bodies (FDA and European CE marking directives) have to date approved only eight mental health apps . On the other hand, app assessment tools, such as the newly developed APA framework , place the onus of assessing app quality and efficacy on the app users or their healthcare providers. Although these are important steps toward improved app quality, they are post-launch assessments that do not prevent low-quality apps from reaching end users.
This study has several strengths. We followed rigorous systematic review methodology for app search and selection, using a specialized search engine to retrieve the maximum number of apps without country restrictions, increasing the generalizability of our findings. The search strategy retrieved apps available worldwide as well as apps restricted to specific countries. We assessed the apps using a comprehensive set of criteria backed by evidence-based clinical guidelines, and trustworthiness of information by adapting HONcode principles.
There were some limitations. By using stringent inclusion criteria, we might have missed apps targeting other mental health disorders providing suicide prevention strategies. The search strategy was limited to four app store categories therefore we may have missed relevant apps available in other categories. Although we aimed to download all eligible apps, we were unable to do so for two Android apps. We did not assess the in-app paid additions offered by depression and mental health management apps as they appeared to be not relevant to suicide prevention and may have missed important pay walled features. Our methods did not include a systematic literature review to identify apps. Therefore, we may have missed some apps developed and tested by research groups that have either not been published in app stores or were no longer available on app stores at the time of our study.
There is a growing number of apps offering suicide prevention strategies to persons at risk, although few provide a comprehensive approach including all six strategies recommended by guidelines. These apps should complement an ongoing patient-provider therapeutic relationship and not replace professional advice. Users should exercise caution when accessing crisis helplines using a suicide prevention app. An effort involving government regulatory agencies, the app development industry, healthcare providers, and the public is urgently needed to create an improved and more transparent model for development and publication of health apps.
Availability of data and materials
All data generated or analyzed during this study are included in this published article and its supplementary information files.
Bostwick JM, Pabbati C, Geske JR, McKean AJ. Suicide attempt as a risk factor for completed suicide: even more lethal than we knew. Am J Psychiatry. 2016;173(11):1094–100 https://doi.org/10.1176/appi.ajp.2016.15070854.
Richards JE, Whiteside U, Ludman EJ, et al. Understanding why patients may not report suicidal ideation at a health care visit prior to a suicide attempt: a qualitative study. Psychiatr Serv. 2019;70(1):40–5. https://doi.org/10.1176/appi.ps.201800342.
Kummervold PE, Gammon D, Bergvik S, Johnsen J-AK, Hasvold T, Rosenvinge JH. Social support in a wired world: Use of online mental health forums in Norway. Nord J Psychiatry. 2002;56(1):59–65. https://doi.org/10.1080/08039480252803945.
Shen N, Levitan M-J, Johnson A, et al. Finding a depression app: a review and content analysis of the depression app marketplace. JMIR Mhealth Uhealth. 2015;3(1):e16. https://doi.org/10.2196/mhealth.3713.
Königbauer J, Letsch J, Doebler P, Ebert D, Baumeister H. Internet- and mobile-based depression interventions for people with diagnosed depression: a systematic review and meta-analysis. J Affect Disord. 2017;223:28–40. https://doi.org/10.1016/j.jad.2017.07.021.
Stiles-Shields C, Montague E, Kwasny MJ, Mohr DC. Behavioral and cognitive intervention strategies delivered via coached apps for depression: Pilot trial. Psychol Serv. 2019;16(2):233–8. https://doi.org/10.1037/ser0000261.
Bush NE, Smolenski DJ, Denneson LM, Williams HB, Thomas EK, Dobscha SK. A virtual hope box: randomized controlled trial of a smartphone app for emotional regulation and coping with distress. Psychiatr Serv. 2017;68(4):330–6. https://doi.org/10.1176/appi.ps.201600283.
Schlosser DA, Campellone TR, Truong B, et al. Efficacy of PRIME, a mobile app intervention designed to improve motivation in young people with schizophrenia. Schizophr Bull. 2018;44(5):1010–20. https://doi.org/10.1093/schbul/sby078.
Champion L, Economides M, Chandler C. The efficacy of a brief app-based mindfulness intervention on psychosocial outcomes in healthy adults: a pilot randomised controlled trial. PLoS One. 2018;13(12):e0209482. https://doi.org/10.1371/journal.pone.0209482.
de Beurs D, Kirtley O, Kerkhof A, Portzky G, O'Connor RC. The role of mobile phone technology in understanding and preventing suicidal behavior. Crisis. 2015;36(2):79–82. https://doi.org/10.1027/0227-5910/a000316.
Witt K, Spittal MJ, Carter G, et al. Effectiveness of online and mobile telephone applications ('apps') for the self-management of suicidal ideation and self-harm: a systematic review and meta-analysis. BMC Psychiatry. 2017;17(1):297. https://doi.org/10.1186/s12888-017-1458-0.
de la Torre I, Castillo G, Arambarri J, López-Coronado M, Franco MA. Mobile apps for suicide prevention: review of virtual stores and literature. JMIR Mhealth Uhealth. 2017;5(10):e130. https://doi.org/10.2196/mhealth.8036.
Van Ameringen M, Turna J, Khalesi Z, Pullia K, Patterson B. There is an app for that! The current state of mobile applications (apps) for DSM-5 obsessive-compulsive disorder, posttraumatic stress disorder, anxiety and mood disorders. Depress Anxiety. 2017;34(6):526–39. https://doi.org/10.1002/da.22657.
Assing Hvidt E, Ploug T, Holm S. The impact of telephone crisis services on suicidal users: a systematic review of the past 45 years. Mental Health Review Journal. 2016;21(2):141–60. https://doi.org/10.1108/MHRJ-07-2015-0019.
Torous JB, Chan SR, Gipson SYT, et al. A hierarchical framework for evaluation and informed decision making regarding smartphone apps for clinical care. Psychiatr Serv. 2018;69(5):498–500. https://doi.org/10.1176/appi.ps.201700423.
Parker L, Karliychuk T, Gillies D, Mintzes B, Raven M, Grundy Q. A health app developer's guide to law and policy: a multi-sector policy analysis. BMC Med Inform Decis Mak. 2017;17(1):141. https://doi.org/10.1186/s12911-017-0535-0.
Wykes T, Schueller S. Why reviewing apps is not enough: transparency for trust (T4T) principles of responsible health app marketplaces. J Med Internet Res. 2019;21(5):e12390. https://doi.org/10.2196/12390.
This study was funded by Nanyang Technological University, Singapore, through the Centre for Population Health Sciences (CePHaS), Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore. Laura Martinengo is funded by a NTU Research Scholarship (LKCMedicine) during her PhD candidacy.
Authors and Affiliations
Centre for Population Health Sciences, Lee Kong Chian School of Medicine, Nanyang Technological University, 11 Mandalay Road, Level 18, Singapore, 308232, Singapore
Laura Martinengo, Louise Van Galen, Elaine Lum & Josip Car
Section of Acute Medicine, Department of Internal Medicine, Amsterdam UMC Location VUmc, Amsterdam, Netherlands
Louise Van Galen
School of Clinical Sciences, Faculty of Health, Queensland University of Technology, Brisbane, Australia
University of Split, School of Medicine, Split, Croatia
Research Division, Institute of Mental Health, Singapore, Singapore
Neuroscience & Mental Health Research Programme, Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore, Singapore
Global Digital Health Unit, Department of Primary Care and Public Health, School of Public Health, Imperial College London, London, UK
JC, LVG, EL, and LM conceptualized the study and developed the app assessment criteria. LM and MK screened and assessed all apps and analyzed and interpreted the data. LM, MK, and EL drafted the manuscript. MS and JC provided critical input into the study. JC, EL, and LVG provided supervision. JC obtained the funding for the study. All authors critically reviewed the manuscript and provided valuable inputs to improve it. All authors read and approved the final manuscript.
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
Martinengo, L., Van Galen, L., Lum, E. et al. Suicide prevention and depression apps’ suicide risk assessment and management: a systematic assessment of adherence to clinical guidelines.
BMC Med17, 231 (2019). https://doi.org/10.1186/s12916-019-1461-z