Exploring the perceptions of judges and prison professionals regarding the use of Artificial Intelligence for the performance of their duties
Abstract
Artificial Intelligence (AI) is increasingly being integrated into the criminal justice system, promising greater efficiency. However, its success depends not only on technological sophistication but also on how professionals in the field perceive it. Comparative research reveals that, while the usefulness of AI is widely acknowledged, fundamental concerns persist regarding algorithmic transparency, data bias, and the necessity of human oversight. These concerns are, in fact, key challenges addressed by national and international regulations and recommendations, which seek to balance technological progress with the rule of law. In this context, the present study addresses a research gap in Spain by exploring how judicial and correctional professionals perceive the impact of AI on their roles. Using the Nominal Group Technique, the study identified AI applications with the greatest perceived potential as well as the main challenges for its advancement. This work provides an important empirical foundation for designing implementation strategies, training programs, and regulatory frameworks that promote responsible AI, while ensuring alignment with the needs of those who will use it
Keywords
Downloads
How to Cite
License
Copyright (c) 2025 María Contreras Román, Sandra Pérez Domínguez

This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
References
BARABAS, C., VIRZA, M., DINAKAR, K., ITO, J., y ZITTRAIN, J., “Interventions over predictions: Reframing the ethical debate for actuarial risk assessment”, Proceedings of Machine Learning Research, vol. 81, 2018, pp. 62-76 (https://proceedings.mlr.press/v81/barabas18a.html, consulta: 29/09/2025).
BARONA VILAR, S. (2023). Algoritmización de la prueba y la decisión judicial en el proceso penal:¿ utopía o distopía?. In El proceso penal ante una nueva realidad tecnológica europea (pp. 133-161). Thomson Reuters Aranzadi. (https://dialnet.unirioja.es/descarga/articulo/8857602.pdf, consulta: 29/09/2025)
BARYSĖ, D., y SAREL, R., “Algorithms in the court: does it matter which part of the judicial decision-making is automated?”, Artificial Intelligence and Law, vol. 32, n.º 1, 2024, pp. 117-146. (https://link.springer.com/article/10.1007/s10506-022-09343-6, consulta: 29/09/2025)
BORGESIUS, F., “Strengthening legal protection against discrimination by algorithms and artificial intelligence”, The International Journal of Human Rights, 2020, vol. 24, no 10, pp. 1572-1593. https://www.tandfonline.com/doi/abs/10.1080/13642987.2020.1743976 (consulta: 13/10/2025)
BUSKEY, B., y WOODS, A., “Making sense of pretrial risk assessments”, National Association of Defense Lawyers, 2018 (https://www.nacdl.org/Article/June2018-MakingSenseofPretrialRiskAssesde, consulta: 14/10/2025).
CASALEIRO, P., ROACH ANLEU, S., y DIAS, J. P., “Introduction. Empirical research with judicial professionals and courts: Methods and practices. Oñati Socio-Legal Series, 2023, 13(S1), S1-S9. https://doi.org/10.35295/osls.iisl.1935 (consulta: 10/10/2025).
CHEN, F., WANG, L., HONG, J., JIANG, J., y ZHOU, L., “Unmasking bias in artificial
intelligence: a systematic review of bias detection and mitigation strategies
in electronic health record-based models”, Journal of Biomedical Informatics
Association, 2024, vol. 31, no 5, p. 1172-1183. https://arxiv.org/abs/2310.19917
(consulta: 10/10/2025). .
COMISIÓN EUROPEA PARA LA EFICACIA DE LA JUSTICIA (CEPEJ), Carta ética europea sobre el uso de la inteligencia artificial en los sistemas judiciales, Consejo de Europa, Estrasburgo, 2018 (https://protecciondata.es/wp-content/uploads/2021/12/Carta-Etica-Europea-sobre-el-uso-de-la-Inteligencia-
Artificial-en-los-sistemas-judiciales-y-su-entorno.pdf, consulta: 14/10/2025).
DE FINE LICHT, K., y DE FINE LICHT, J., “Artificial intelligence, transparency, and public decision-making: Why explanations are key when trying to produce perceived legitimacy”, AI & Society, 2020, vol. 35, n.º 4, pp. 917-926. https://link.springer.com/article/10.1007/s00146-020-00960-w (consulta: 15/10/2025).
DELBECQ, A. L., y VAN DE VEN, A. H., “A Group Process Model for Problem Identification and Program Planning”, The Journal of Applied Behavioral Science, vol. 7, n.º 4, 1971, pp. 467-492. (https://doi.org/10.1177/002188637100700404, consulta: 29/09/2025)
FANEGA, M., FRESNEDA PORTILLO, C., y BELTRÁN CAMACHO, Á., “Publicación del algoritmo penitenciario TVR”, Revista Española de Investigación Criminológica, 2025. (https://doi.org/10.46381/reic.v22i1.917, consulta: 12/10/2025)
FINE, A., LE, S., y MILLER, M. K., “Content Analysis of Judges’ Sentiments Toward Artificial Intelligence Risk Assessment Tools” Criminology, Criminal Justice, Law & Society 24 (2): 31–46. (https:/ / doi.org/ 10.54555/ CCJLS.8169.84869, consulta: 13/10/2015).
GALLI, F., y SARTOR, G., “AI Approaches to Predictive Justice: A Critical Assessment”, Human(ities) and Rights – Global Network Journal, vol. 5, n.º 2, 2024, pp. 217-233 (https://www.humanitiesandrights.com/journal/index.php/har/article/view/118, consulta: 13/10/2025)
GOBIERNO DE ESPAÑA, Estrategia Nacional de Inteligencia Artificial, Secretaría de Estado de Digitalización e Inteligencia Artificial, Madrid, 2020 (https://portal.mineco.gob.es/eses/Ministerio/Documents/20201202_ENIA.pdf,
consulta: 14/10/2025).
GUILLÉN ZANÓN, Á., “La Técnica del Grupo Nominal”, Documentación Administrativa, n.º 223, 1990, pp. 45-68. (https://revistasonline.inap.es/index.
php/DA/article/download/5196/5250, consulta: 29/09/2025)
HOFF, K. A., y BASHIR, M., “Trust in automation: Integrating empirical evidence on factors that influence trust”, Human Factors, vol. 57, n.º 3, 2015, pp. 407-434
(https://doi.org/10.1177/0018720814547570, consulta: 29/09/2025).
IVERSON, J., “Surveilling Potential Uses and Abuses of Artificial Intelligence in Correctional Spaces”, Lincoln Memorial University Law Review, vol. 9, n.º 1,
https://digitalcommons.lmunet.edu/lmulrev/vol9/iss3/1, consulta:
/10/2025)

