AI-Enabled Clinical Decision Support Tools for Mental Healthcare: A Product Review

Author

Anne-Kathrin Kleine, Eesha Kokje, Pia Hummelsberger, & Susanne Gaube

Abstract

Introduction

While AI has transformed several fields of medicine, its application in mental healthcare is still in the early stages due to challenges such as the subjective nature of psychiatric diagnoses and the complexity of mental health disorders. Nevertheless, advances are being made, with AI tools beginning to assist in tasks like detection of mental health conditions through speech and text analysis, and personalized treatment recommendation. The implementation of AI into mental healthcare holds the promise of a more accurate, efficient and personalized approach to the treatment of mental disorders (Shen et al., 2021).

One of the key challenges in designing AI systems for human-AI interaction is the uncertainty surrounding AI’s capabilities and the complexity of generating its output (Yang et al., 2020). This poses unique design challenges for AI systems, particularly in the context of healthcare, where accurate and reliable diagnoses and treatment recommendations are crucial (Shen et al., 2021). Regulatory agencies, such as the US Food and Drug Administration (FDA), have undertaken efforts to address the challenges associated with regulating AI software (Carolan et al., 2022; Gerke et al., 2020). However, the implementation of these regulatory reforms is still pending and no tangible solutions to address the complexities involved in regulating AI software is currently available, leading to a significant shortage of regulated AI-CDSS for healthcare (Mashar et al., 2023; US Food and Drug Administration, 2019, 2021).

Given the complex nature of mental health conditions, the lack of standardized, high-quality data, rigorous regulatory processes, and the ethical implications in data use, the development and regulation of AI-CDSS for mental healthcare present considerable challenges. These hurdles often prove too resource-intensive for small firms and start-ups to overcome. By mapping out the regulatory status and scientific efficacy of each product, our review provides insights into the current landscape of regulated AI-CDSS in mental healthcare, thereby identifying potential avenues for future innovation, application, and regulation.

Materials and methods

Artificial intelligence (AI)-enabled clinical decision support products

Artificial intelligence refers to “the capability of an engineered system to acquire, process and apply knowledge and skills” (International Organization for Standardization, 2020). Medical devices are intended for diagnostic, treatment, or prevention of mental disorders (European Commission, 2022; United States Code, 1964; World Health Organization, 2017). AI software that fulfills the purpose of a medical device may be considered a medical device itself (European Commission, 2019; US Food and Drug Administration, 2013). In this context, we focus on regulated diagnostic and treatment recommendation products that may be used for clinical decision support.

Search strategies and inclusion criteria

We followed four main approaches to generate an overview of AI software products for mental health. Patenting is a common practice to protect intellectual property rights and encourage innovation. As AI technologies continue to advance, patents are expected to increase, particularly for AI-enabled medical devices (ResearchandMarkets, 2020). Accordingly, first, we used information collected as part of a patent review on AI-enabled devices for mental healthcare (Kleine et al., 2023). We checked the assignee and inventor details of the retrieved patents to identify relevant companies. Detailed information about the search strategy and selection criteria is described in Kleine et al. (2023).

Second, exhibitor lists from the American Psychological Association Congress, the American Psychiatric Association Congress, the Association for Behavioral and Cognitive Therapies, the Anxiety and Depression Association of America, and the German Psychotherapy Congress were searched for relevant companies. Third, we searched the FDA Artificial Intelligence and Machine Learning (AI/ML)-Enabled Medical Devices database (Drug Administration, 2022), the Medical AI Evaluation Database (Wu et al., 2021), the Medical Futurist AI-based algorithms database (The Medical Futurist, 2023), and the HealthSkouts FDA/CE Certified Health App database (HealthSkouts, 2023).

We applied four inclusion criteria. First, the product should be a medical device. Thus, we excluded products targeting diseases unrelated to mental health (e.g., Medtronic), wellness apps (e.g., the Sonde Mental Fitness App), products used for organizing and managing patient information (e.g., Telemynd), and educational and research products or services (e.g., ABA Technologies).

Second, the product should include AI software in its core functionality. This criterion led to the exclusion of products that did not clearly employ AI algorithms (e.g., Autism Analytica).

Third, we included products specifically aimed at diagnosing or generating treatment recommendations for mental health practitioners. This led to the exclusion of diagnosis or treatment apps to be used by patients (e.g., Endeavorrx by Akili Interactive), and psychopharmaceutical and technical products (e.g., Neumora, Otsuka, and the SAINT Neuromodulation System by Magnus Medical).

The U.S. Food and Drug Administration (FDA) and the European Union (EU), via its CE marking, are the leading medical device regulatory bodies globally. The FDA and the EU’s CE marking assure high-quality standards in safety and efficacy, guaranteeing that the device has been thoroughly reviewed and tested (Muehlematter et al., 2021). Accordingly, fourth, we included products that were CE-marked or cleared by the FDA. Information about CE certificates and FDA clearance or approval was confirmed using the EUDAMED and the FDA databases. In addition, all vendors were contacted to verify and supplement the collected information.

We coded information about the targeted disorder, data input, data output, the method of deployment, integration possibilities, and date to market (Leeuwen et al., 2021). We maintain an overview of products on XYZ.com. Discrepancies between the products considered in the review and those displayed on XYZ.com may be caused by updates of the products’ regulatory status, the inclusion of new products, or refusal of companies to appear on the website.

Scientific evidence

We searched for scientific evidence proving the efficacy of the included products. First, the database PubMed was searched by vendor and product name for peer-reviewed articles published between Jan 1, 2015, and July, X 2023 (Leeuwen et al., 2021). Queries are provided in Table XYZ. Second, a manual search was performed by inspecting the vendor’s websites for listings of papers and requesting vendors to provide peer-reviewed papers. No date restriction was applied for the manual search.

Included articles were original, peer-reviewed, and in English, and aimed to demonstrate the efficacy of the AI software. Papers were included when the product name (including known former names) and/or company name were mentioned, the tool was applied on in vivo human data, and efficacy of the product was reported on an independent dataset (data on which the algorithm was not trained). Letters, commentaries, reviews, study protocols, white papers, and case reports were excluded. Papers were assessed by two of the authors who independently screened the title, abstracts, and full paper for inclusion criteria. Cases of disagreement were resolved by the reviewers in a consensus meeting.

Evaluation of scientific evidence

We used the CONSORT-AI (Consolidated Standards of Reporting Trials Artificial Intelligence) checklist to evaluate the scientific evidence (Liu et al., 2020). The CONSORT-AI extension of the original CONSORT checklist (Moher et al., 2012) includes 14 new items requiring a clear description of the AI intervention, including instructions and skills required for use, the setting in which the AI intervention is integrated, the handling of inputs and outputs of the AI intervention, and provision of an analysis of error cases (Liu et al., 2020; Tornero-Costa et al., 2023; zhou_etal21is?). Information regarding the scientific evidence was coded independently by two of the authors. Any discrepancies were documented and resolved in consensus meetings.

Data analysis

The data were analyzed using R (Version 4.2.2, rcoreteam22b?). First, we display information on the inclusion and exclusion of products found through our search strategies. Second, we provide information on the targeted disorder(s), data input, output, and the time passed between company foundation, product certification, and the publication of scientific evidence addressing product efficiency. In addition, we provide information on CE marks and FDA clearance. Finally, we evaluate the scientific evidence addressing product efficiency. We obtained all relevant information from vendors directly, through vendor websites, online press releases, and online articles.

Results

Inclusion and exclusion of products based on inclusion criteria

[1] 240

The PRISMA diagram is shown in Figure Figure 1.

Exclude reason
NO MH MEDICAL DEVICE      NOT AI SOFTWARE             NOT_CDSS 
                  44                    2                   15 

Figure 1: PRISMA Statement

Product overview

The product search resulted in three products that were cleared by the FDA, five that were CE marked, and one that was UKCA marked. One company offered two products. Most products were cloud-based stand-alone third party applications. For two tools, information about integration possibilities was missing and one tool was compatible with leading polysomnography devices.

Figure 2 displays the disorders targeted by the products.

Figure 2: Targeted disorders

Figure 3 shows an overview of the data sources used to derive recommendations.

Figure 3: Data sources

Figure 4 shows an overview of the outputs generated by the products.

Figure 4: Outputs

Figure 5 displays the type of output generated by the mental disorder targeted.

Figure 5: Output type by mental disorder

Figure 6 displays the time passed between company foundation, product certification, and the publication of scientific evidence addressing product efficiency. Please note that we excluded the foundation date of Medibio (MEBSLEEP), which was founded in 1987.

Figure 6: Outputs

Figure 7 displays the certification the products received. De Novo is a regulatory pathway for marketing approval of novel medical devices in the US that do not have a predicate device for comparison. It is used for low to moderate risk devices that are not substantially equivalent to any existing device, enabling them to enter the market without requiring the traditional pre-market approval process. 510(k) is a pre-market submission to the FDA demonstrating that a medical device to be marketed is at least as safe and effective, or “substantially equivalent,” to a legally marketed device (predicate device) that is not subject to pre-market approval. Predicate devices serve as a benchmark for comparative purpose allowing more rapid approval of new devices. They help provide assurance for safety and effectiveness of the new device through comparison. The predicate device for the EarliPoint System is the Cognoa ASD Diagnosis Aid and the predicate device of EnsoSleep is a similar previous version of EnsoSleep.

Figure 7: Outputs

Evidence

Discussion

Summary of Key Findings

The aim of this study was to increase transparency in the field of commercial AI-CDSS software in mental healthcare and its scientific evidence. Our results demonstrate that the market is still in its very infancy. We could identify only nine products that obtained FDA clearance or CE marking. The fact that all of the products received regulatory approval in the past three years (November 2020 - January 2023) may be an explanation for this. That is, despite the large number of solutions having been developed to aid clinicians in deriving diagnoses and selecting appropriate treatment approaches (Aafjes-van Doorn et al., 2021; Chekroud et al., 2021; Cho et al., 2019; Graham et al., 2019; Shatte et al., 2019), for most products we still await to discover the impact they may have in clinical practice.

  • discussion of scientific evidence
  • discussion of device type by mental disorder
  • discussion of time passed between company founding, product certification, publication of scientific evidence

Implications for Current AI Landscape in Mental Healthcare

  • What does the presence of regulated products mean for start-ups and small firms?
  • Do the regulated products shed light on effective strategies for overcoming regulatory and data-related challenges?
  • What could other firms glean from those successful products?

Intersection of Regulatory Practices and AI-CDSS Development

  • Reflect on the role of FDA clearance, CE marking, de novo regulations, and 510(k) benchmarking in the development of AI-CDSS
  • How do these regulatory practices impact the AI development process?
  • How do they shape the adoption and implementation of AI applications in mental healthcare?
  • Discussion of AI-CDSS for mental healthcare that are unregulated - reasons for lack of regulation (documented for each single product)
  • Discussion of options to get around regulation (Wellness app, research use only, “selling a platform”)
  • Discussion of currently available unregulated products (e.g., treatment apps)

Necessary Adaptations of Regulatory Frameworks for the Regulation of AI-CDSS medical devices in Mental Health

  • what needs to be done to make the regulatory process more efficient?

Study Limitations and Future Directions

  • Review limitations
  • Sources of bias
  • Unavailability of data
  • Lack of responses from vendors
  • What products or companies may have been excluded and why?
  • Future directions for regulated AI-CDSS for mental health (which areas, how approach the regulation process)

Conclusion

References

Aafjes-van Doorn, K., Kamsteeg, C., Bate, J., & Aafjes, M. (2021). A scoping review of machine learning in psychotherapy research. Psychotherapy Research, 31(1), 92–116. https://doi.org/10.1080/10503307.2020.1808729
Carolan, J. E., McGonigle, J., Dennis, A., Lorgelly, P., & Banerjee, A. (2022). Technology-enabled, evidence-driven, and patient-centered: The way forward for regulating software as a medical device. JMIR Medical Informatics, 10(1), e34038. https://doi.org/10.2196/34038
Chekroud, A. M., Bondar, J., Delgadillo, J., Doherty, G., Wasil, A., Fokkema, M., Cohen, Z., Belgrave, D., DeRubeis, R., Iniesta, R., Dwyer, D., & Choi, K. (2021). The promise of machine learning in predicting treatment outcomes in psychiatry. World Psychiatry, 20(2), 154–170. https://doi.org/10.1002/wps.20882
Cho, G., Yim, J., Choi, Y., Ko, J., & Lee, S.-H. (2019). Review of machine learning algorithms for diagnosing mental illness. Psychiatry Investigation, 16(4), 262–269. https://doi.org/10.30773/pi.2018.12.21.2
Drug Administration, U. F. {and}. (2022, October 5). Artificial intelligence and machine learning (AI/ML)-enabled medical devices. FDA. https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-aiml-enabled-medical-devices
European Commission. (2019). MDCG 2019-11 guidance on qualification and classification of software in regulation (EU) 2017/745 – MDR and regulation (EU) 2017/746 – IVDR. https://ec.europa.eu/docsroom/documents/37581
European Commission. (2022). Medical devices – sector. https://ec.europa.eu/health/medical-devices-sector/overview_en
Gerke, S., Babic, B., Evgeniou, T., & Cohen, G. I. (2020). The need for a system view to regulate artificial intelligence/machine learning-based software as medical device. NPJ Digital Medicine. https://doi.org/10.1038/s41746-020-0262-2
Graham, S., Depp, C., Lee, E. E., Nebeker, C., Tu, X., Kim, H.-C., & Jeste, D. V. (2019). Artificial intelligence for mental health and mental illnesses: An overview. Current Psychiatry Reports, 21(11), 116. https://doi.org/10.1007/s11920-019-1094-0
HealthSkouts. (2023). Certified apps library – healthskouts digital health advisory company. https://www.healthskouts.com/certified-apps/
International Organization for Standardization. (2020). ISO/IEC TR 24028:2020(en), information technology — artificial intelligence — overview of trustworthiness in artificial intelligence. https://www.iso.org/obp/ui/#iso:std:iso-iec:tr:24028:ed-1:v1:en
Kleine, A.-K., Lermer, E., Cecil, J., Heinrich, A., & Gaube, S. (2023). Advancing mental health care with AI-enabled precision psychiatry tools: A patent review. PsyArXiv. https://doi.org/10.31234/osf.io/wmr38
Leeuwen, K. G. van, Schalekamp, S., Rutten, M. J. C. M., Ginneken, B. van, & Rooij, M. de. (2021). Artificial intelligence in radiology: 100 commercially available products and their scientific evidence. European Radiology, 31(6), 3797–3804. https://doi.org/10.1007/s00330-021-07892-z
Liu, X., Cruz Rivera, S., Moher, D., Calvert, M. J., & Denniston, A. K. (2020). Reporting guidelines for clinical trial reports for interventions involving artificial intelligence: The CONSORT-AI extension. Nature Medicine, 26(9), 1364–1374. https://doi.org/10.1038/s41591-020-1034-x
Mashar, M., Chawla, S., Chen, F., Lubwama, B., Patel, K., Kelshiker, M. A., Bachtiger, P., & Peters, N. S. (2023). Artificial intelligence algorithms in health care: Is the current food and drug administration regulation sufficient? JMIR AI, 2(1), e42940. https://doi.org/10.2196/42940
Moher, D., Hopewell, S., Schulz, K. F., Montori, V., Gøtzsche, P. C., Devereaux, P. J., Elbourne, D., Egger, M., & Altman, D. G. (2012). CONSORT 2010 explanation and elaboration: Updated guidelines for reporting parallel group randomised trials. International Journal of Surgery, 10(1), 28–55. https://doi.org/10.1016/j.ijsu.2011.10.001
Muehlematter, U. J., Daniore, P., & Vokinger, K. N. (2021). Approval of artificial intelligence and machine learning-based medical devices in the USA and europe (2015–20): A comparative analysis. The Lancet Digital Health, 3(3), e195–e203. https://doi.org/10.1016/S2589-7500(20)30292-2
ResearchandMarkets. (2020). Medical devices in 2025: An AI impact analysis. https://store.frost.com/medical-devices-in-2025-an-ai-impact-analysis.html
Shatte, A. B. R., Hutchinson, D. M., & Teague, S. J. (2019). Machine learning in mental health: A scoping review of methods and applications. Psychological Medicine, 49(9), 1426–1448. https://doi.org/10.1017/S0033291719000151
Shen, Y., Shamout, F. E., Oliver, J. R., Witowski, J., Kannan, K., Park, J., Wu, N., Huddleston, C., Wolfson, S., Millet, A., Ehrenpreis, R., Awal, D., Tyma, C., Samreen, N., Gao, Y., Chhor, C., Gandhi, S., Lee, C., Kumari-Subaiya, S., … Geras, K. J. (2021). Artificial intelligence system reduces false-positive findings in the interpretation of breast ultrasound exams. Nature Communications, 12(1), 5645. https://doi.org/10.1038/s41467-021-26023-2
The Medical Futurist. (2023). AI-based algorithms. https://medicalfuturist.com/fda-approved-ai-based-algorithms//
Tornero-Costa, R., Martinez-Millana, A., Azzopardi-Muscat, N., Lazeri, L., Traver, V., & Novillo-Ortiz, D. (2023). Methodological and quality flaws in the use of artificial intelligence in mental health research: Systematic review. JMIR Mental Health, 10(1), e42045. https://doi.org/10.2196/42045
United states code: Federal food, drug, and cosmetic act, 21 u.s.c. §§ 321-392 suppl. 5., (1964). https://www.loc.gov/item/uscode1964-023021009/
US Food and Drug Administration. (2013). Software as a medical device (SaMD): Key definitions.
US Food and Drug Administration. (2019). Proposed regulatory framework for modifications to artificial intelligence/machine learning (AI/ML)-based software as a medical device (SaMD): Discussion paper and request for feedback. https:/​/www.​fda.gov/​files/​medical%20devices/​published/​US-FDA-Artificial-Intelligence-and-Machine-Learning-Discussion-Paper.​pdf
US Food and Drug Administration. (2021). Artificial intelligence/machine learning (AI/ML)-based software as a medical device (SaMD) action plan. https://www.fda.gov/media/145022/download [accessed 2022-12-10
World Health Organization. (2017). WHO global model regulatory framework for medical devices including in vitro diagnostic medical devices. World Health Organization. https://apps.who.int/iris/handle/10665/255177
Wu, E., Wu, K., Daneshjou, R., Ouyang, D., Ho, D. E., & Zou, J. (2021). How medical AI devices are evaluated: Limitations and recommendations from an analysis of FDA approvals. Nature Medicine, 27(4), 582–584. https://doi.org/10.1038/s41591-021-01312-x
Yang, Q., Steinfeld, A., Rosé, C., & Zimmerman, J. (2020). Re-examining whether, why, and how human-AI interaction is uniquely difficult to design. Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, 1–13. https://doi.org/10.1145/3313831.3376301