International Journal of Reliability, Risk and Safety: Theory and Application

International Journal of Reliability, Risk and Safety: Theory and Application

Artificial Intelligence in Psychotherapy: Risks, Trustworthiness, and Safety Concerns from the Perspective of Clinical Professionals

Document Type : Original Research Article

Author
Associate Professor Psychology, Islamshahr branch, Islamic Azad University, Islamshahr, Iran.
Abstract
The rapid integration of Artificial Intelligence (AI) into psychotherapy has transformed psychological evaluation and intervention, while simultaneously raising critical concerns regarding clinical reliability, ethical integrity, and patient safety. This qualitative case study examined the perspectives of an experienced clinical psychologist on the risks, trustworthiness, and safety considerations associated with AI-based psychological evaluations. Data were collected through a semi-structured, in-depth interview and were analyzed thematically following Braun and Clarke’s (2006) framework. Three overarching themes emerged: (1) clinical and social risks, including algorithmic mislabeling, decontextualized interpretation, and erosion of the therapeutic alliance; (2) technical and clinical reliability, reflecting the discrepancy between AI accuracy in objective testing and its limitations in capturing emotional and cultural nuance; and (3) ethical and justice-oriented concerns, encompassing algorithmic bias, data confidentiality, client dependency, and the need for professional training and regulatory oversight. The resulting conceptual model, described as the “Risk–Trust–Ethics Triad,” demonstrates a dynamic, interdependent relationship among these dimensions, suggesting that any imbalance—technical, moral, or human—undermines the stability of the system. Findings highlight the necessity of human supervision, ethical governance, and culturally sensitive design as prerequisites for the deployment of trustworthy AI in psychotherapy. The study concludes that the sustainable integration of AI requires a human-centered framework in which technology functions not as a replacement for clinical judgment, but as an ethical partner that enhances therapeutic insight and judgment.
Keywords
Subjects

[1] A. Spiegler, M. Maas, and T. Suisse, Artificial Intelligence and the Future of Defense, M.A. Faghih and F. Nazarizadeh, transl., Tehran, Iran, Defense Industries Education and Research Institute, 2020, (in Persian).
[2] F. Nazarizadeh, A. Farsijani, and H. Khazaeil, “The future of artificial intelligence in strategic management of organizations,” Life Sciences and Emerging Fields, vol. 1, no. 1, pp. 119-156, 2023, (in Persian).
[3] J. Canals and F. Heukamp, Ed. The Future of Management in an AI World: Redefining Purpose and Strategy in the Fourth Industrial Revolution, Cham, Switzerland: Palgrave Macmillan, 2020, https://doi.org/10.1007/978-3-030-20680-2.
[4] F. Rezaei, “Investigating the impact of artificial intelligence on psychological well-being and stress reduction: Analyzing the impact of new technology on improving mental health,” Journal of Cognition, Behavior, Learning, vol. 1, no. 1,  pp. 167-181, 2024, (in Persian), https://doi.org/10.61838/jcbl.1.1.13.  
[5] S.S. Hosseini and M.R. Yamaghani, “multimodal emotional recognition by artificial intelligence and its application in psychology,” Journal of Arak University Medical Sciences, vol. 27, no. 4,  pp. 194-204, http://dx.doi.org/10.61186/jams.27.4.194.  
[6] F. Bagheri Sheykhangafshe, A. Abolghasemi, and S.M. Kafi Masouleh, “Predicting resilience based on dark triad personality and psychological wellbeing in athletes students,” Journal of Arak University Medical Sciences, vol. 24, no. 2, pp. 230-45, 2021, (in Persian), http://dx.doi.org/10.32598/jams.24.2.6151.1.
[7] T. Anvarjon, M. Mustaqeem, and S. Kwon, “Deep-net: A lightweight CNN-based speech emotion recognition system using deep frequency features,” Sensors (Basel), vol. 20, no. 18, 2020, Art. no. 5212, https://doi.org/10.3390/s20185212.
[8] S.K. Bharti et al., “Text-Based Emotion Recognition Using Deep Learning Approach,” Computational Intelligence and Neuroscience, 2022, Art. no. 2645381, https://doi.org/10.1155/2022/2645381.  
[9] H.J. Rumpf, D. Browne, D. Brandt, and F. Rehbein, “Addressing taxonomic challenges for Internet use disorders in light of changing technologies and diagnostic classifications,” Journal of behavioral addictions, vol. 9, no. 4, 2020, pp. 942-944, https://doi.org/10.1556/2006.2020.00094.
[10] X. Kong and K. Zhang, “A novel text sentiment analysis system using improved depthwise separable convolution neural networks,” PeerJ Computer Science, vol. 9, 2023, Art. no. e1236, https://doi.org/10.7717/peerj-cs.1236.  
[11] S. Rauschert, K. Raubenheimer, P.E. Melton, R.C. Huang, “Machine learning and clinical epigenetics a review of challenges for diagnosis and classification,” Clin Epigenetics, vol. 12, 2020, Art. no. 51, http://doi:10.1186/s13148-020-00842-4.
[12] A. Das, J. Mock, Y. Huang, E. Golob, and P. Najafirad, “interpretable self-supervised facial micro-expression learning to predict cognitive state and neurological disorders,” in AIAA Conference on Artificial Intelligence, vol. 35, no. 1, pp. 818-826, https://doi.org/10.1609/aaai.v35i1.16164.  
[13] Y. Ozkanca, M.G. Öztürk, M.N. Ekmekci, D.C. Atkins, C. Demiroglu, and R. Hosseini Ghomi, “Depression screening from voice samples of patients affected by Parkinson's disease,” Digit Biomark, vol. 3, no. 2, pp. 72-82, 2019,  http://doi:10.1159/000500354.
[14] J. Dong, Z. Wu, H. Xu, and D. Ouyang, “Formulation AI a novel web-based platform for drug formulation design driven by artificial intelligence,” Brief Bioinform, vol. 25, no. 1, 2023 Art. no.  bbad419, http://doi:10.1093/bib/bbad419.
[15] Y. Zhai, X. Song, Y. Chen, W. Lu, “A study of mobile medical app user satisfaction incorporating theme analysis and review sentiment tendencies,” International journal of environmental research and public health, vol. 19, no. 12, 2022, Art. no. 7466, https://doi.org/10.3390/ijerph19127466.  
[16] M. Amirkhani, M.H. Sadeghian, A. Hajbi, and A.A. Shojaei, “Investigating the challenges of using artificial intelligence in suicide counseling with an interdisciplinary approach,” Medicine and spiritual health, vol. 33, no. 4, 2024, pp. 183-170, (in Persian).
[17] M. Tamtaji and A. Ekrami Kivaj, “The role of artificial intelligence in implementing HSE standards in the petrochemical industry,” International Journal of Reliability, Risk and Safety: Theory and Application, vol. 8, no. 1, pp. 30-46, 2025, https://doi.org/10.22034/IJRRS.2025.8.1.3.  
[18] M. Karimi, “The Role of artificial intelligence in reducing risks and managing risk in the future,” Research in safety, heaith and environment, vol. 2, no. 7, pp. 63-71, 2024, (in Persian).
[19] F. Rahmati and J. Mashayekhi, “Ethical considerations of the use of artificial intelligence (AI) in health care system,” Payesh, vol. 23, no. 5, 2024, 23(5): 793-796. http://payeshjournal.ir/article-1-2248-fa.html.
[20] S. Heydari, “Normalization of the artificial intelligence literacy scale based on classical test theory (CTT),” International Journal of Reliability, Risk and Safety: Theory and App lication, vol. 8, no. 1, pp. 47-54, 2025, https://doi.org/10.22034/IJRRS.2025.8.1.4.
Volume 9, Issue 1
February 2026
Pages 1-8

  • Receive Date 06 November 2025
  • Revise Date 12 December 2025
  • Accept Date 31 January 2026