When AI Weakens Evidence Traceability: Ethical Challenges for Credibility and Responsibility in Digital Forensics
Abstract
Artificial intelligence (AI) is increasingly used in digital forensic workflows to support the analysis, interpretation, and reporting of large datasets. While this integration offers efficiency benefits, it also raises ethical concerns. This paper examines the weakening of evidence traceability in AI-assisted digital forensic workflows as a challenge in digital forensics and argues that reduced traceability threatens two core requirements of forensic practice: evidence credibility and responsibility attribution. Unlike conventional technical errors, traceability breakdowns may generate analytic statements that appear plausible but lack explicit grounding in identifiable forensic artifacts, thereby weakening the link between evidence and inference.
Through an ethics- and governance-focused analysis, the paper shows how AI-assisted workflows can undermine evidentiary trust, distribute responsibility across human and institutional actors, and produce accountability gaps that conflict with forensic norms. The analysis also examines human and institutional effects of AI use, including automation bias, institutional reliance on AI outputs, and the displacement of responsibility.
The paper concludes that technical validation alone cannot address these issues. Responsible use of AI in digital forensics requires clear human–AI role definition, procedures that support auditability and contestability, and governance frameworks that assign accountability across the AI lifecycle. These conditions are necessary to maintain the credibility and legitimacy of digital forensic evidence.
Keywords:
Artificial Intelligence; Digital Forensics; Evidence Traceability; Algorithmic Accountability; AI GovernanceData Availability Statement
Data sharing is not applicable to this article as the study is based on normative ethical analysis and literature review. No datasets were generated or analysed during the research.
Copyright Notice & License:
All articles published in AI & Future Society are licensed under the Creative Commons Attribution 4.0 International License (CC BY 4.0). Authors retain copyright and grant the journal the right of first publication. This license permits anyone to copy, distribute, remix, adapt, and build upon the work—even commercially—provided proper credit is given to the original author(s) and the source, a link to the license is provided, and any changes are indicated.

This work is licensed under a Creative Commons Attribution 4.0 International License.
References
Aleke, N. T., & Trigui, M. (2025). Legal and ethical challenges in digital forensics investigations. In Digital Forensics in the Age of AI (pp. 147–176). IGI Global.
Ananny, M., & Crawford, K. (2018). Seeing without knowing: Limitations of the transparency ideal and its application to algorithmic accountability. New Media & Society, 20(3), 973–989.
Arrieta, A. B., Díaz-Rodríguez, N., Del Ser, J., Bennetot, A., Tabik, S., Barbado, A., ... & Herrera, F. (2020). Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion, 58, 82–115.
Casey, E. (2009). Handbook of digital forensics and investigation. Academic Press.
Coltri, M. A., Uys, W. R., & Joubert, K. (2025). Investing AI ethics in forensic investigations: Development, policies, and best practices. Athens Journal of Sciences, 355.
Dignum, V. (2019). Responsible artificial intelligence: How to develop and use AI in a responsible way (Vol. 2156). Springer.
Dunsin, D., Ghanem, M. C., Ouazzane, K., & Vassilev, V. (2024). A comprehensive analysis of the role of artificial intelligence and machine learning in modern digital forensics and incident response. Forensic Science International: Digital Investigation, 48, 301675.
Dzindolet, M. T., Pierce, L. G., Beck, H. P., & Dawe, L. A. (2002). The perceived utility of human and automated aids in a visual detection task. Human Factors, 44(1), 79–94.
Elish, M. C. (2025). Moral crumple zones: Cautionary tales in human–robot interaction. In Robot Law: Volume II (pp. 83–105). Edward Elgar Publishing.
Endsley, M. R. (2017). From here to autonomy: Lessons learned from human–automation research. Human Factors, 59(1), 5–27.
Floridi, L., Cowls, J., Beltrametti, M., Chatila, R., Chazerand, P., Dignum, V., ... & Vayena, E. (2018). AI4People—An ethical framework for a good AI society: Opportunities, risks, principles, and recommendations. Minds and Machines, 28(4), 689–707.
Hefetz, I. (2025). Evaluating bias in forensic evidence: From expert analysis to AI-based decision tools. Forensic Science International: Synergy, 11, 100645.
Horsman, G. (2022). Technical reporting in digital forensics. Journal of Forensic Sciences, 67(6), 2458–2468.
Howell, B. A. (2004). Ambiguities in US law for investigators. Digital Investigation, 1(2), 106–111.
Hu, Y. (2024). The market strategy of the international education industry: A differentiation perspective in research. SHS Web of Conferences, 187, 04011. EDP Sciences.
Hu, Y., Jusoh, S., Abd Razak, M. F., Yusof, A. R. B. M., Xie, Q., Song, C., & Zong, X. (2026). Understanding the digital transformation of SMEs: A systematic review from a dynamic capabilities perspective across business, technology, and societal domains. Global Review of Humanities, Arts, and Society, 2(1), 1–17.
Huang, L., Yu, W., Ma, W., Zhong, W., Feng, Z., Wang, H., ... & Liu, T. (2025). A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions. ACM Transactions on Information Systems, 43(2), 1–55.
Ji, Z., Lee, N., Frieske, R., Yu, T., Su, D., Xu, Y., ... & Fung, P. (2023). Survey of hallucination in natural language generation. ACM Computing Surveys, 55(12), 1–38.
Kent, K., Chevalier, S., Grance, T., & Dang, H. (2006). Guide to integrating forensic techniques into incident response (NIST Special Publication 800-86). National Institute of Standards and Technology. https://doi.org/10.6028/NIST.SP.800-86
Khan, A. A., Zhang, X., Hajjej, F., Yang, J., Ku, C. S., & Por, L. Y. (2024). ASMF: Ambient social media forensics chain of custody with an intelligent digital investigation process using federated learning. Heliyon, 10(1).
Kroll, J. A., Huey, J., Barocas, S., Felten, E. W., Reidenberg, J. R., Robinson, D. G., & Yu, H. (2017). Accountable algorithms. University of Pennsylvania Law Review, 165(3), 633–705.
Liu, E. (2025). Building trustworthy AI: Transparency, fairness, and governance in the digital age. AI & Future Society, 1(1), 3–5.
Magesh, V., Surani, F., Dahl, M., Suzgun, M., Manning, C. D., & Ho, D. E. (2025). Hallucination‐free? Assessing the reliability of leading AI legal research tools. Journal of Empirical Legal Studies, 22(2), 216–242.
Marsico, F., & Amigo, M. (2025). Ethical and security challenges in AI for forensic genetics: From bias to adversarial attacks. Forensic Science International: Genetics, 76, 103225.
Marturana, F., & Tacconi, S. (2013). A machine learning-based triage methodology for automated categorization of digital media. Digital Investigation, 10(2), 193–204.
Mata v. Avianca, Inc., 22-cv-1461 (S.D.N.Y. 2023).
Maynez, J., Narayan, S., Bohnet, B., & McDonald, R. (2020). On faithfulness and factuality in abstractive summarization. arXiv preprint arXiv:2005.00661.
Mittelstadt, B. D., Allo, P., Taddeo, M., Wachter, S., & Floridi, L. (2016). The ethics of algorithms: Mapping the debate. Big Data & Society, 3(2).
Monasterio Astobiza, A., Toboso, M., Aparicio, M., Ausín, T., López, D., Morte, R., & Pons, J. L. (2019). Bringing inclusivity to robotics with INBOTS. Nature Machine Intelligence, 1(4), 164.
National Institute of Standards and Technology. (2023). AI risk management framework (AI RMF 1.0). https://doi.org/10.6028/NIST.AI.100-1
Ngejane, C. H., Eloff, J. H., Sefara, T. J., & Marivate, V. N. (2021). Digital forensics supported by machine learning for the detection of online sexual predatory chats. Forensic Science International: Digital Investigation, 36, 301109.
OECD. (2019). OECD principles on artificial intelligence. OECD Publishing.
Parasuraman, R., & Riley, V. (1997). Humans and automation: Use, misuse, disuse, abuse. Human Factors, 39(2), 230–253.
Resnik, D. B., & Hosseini, M. (2025). The ethics of using artificial intelligence in scientific research: New guidance needed for a new tool. AI and Ethics, 5(2), 1499–1521.
Rudin, C. (2019). Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5), 206–215.
Scientific Working Group on Digital Evidence. (2018). Requirements for report writing in digital and multimedia forensics (18-Q-002-1.0). https://www.swgde.org/wp-content/uploads/2023/11/2018-11-20-SWGDE-Requirements-for-Report-Writin.pdf
Sharma, B., Ghawaly, J., McCleary, K., Webb, A. M., & Baggili, I. (2025). ForensicLLM: A local large language model for digital forensics. Forensic Science International: Digital Investigation, 52, 301872.
Singh, S., & Devi, L. (2025). Reliability and admissibility of AI-generated forensic evidence in criminal trials. arXiv preprint arXiv:2601.06048.
Studiawan, H., Breitinger, F., & Scanlon, M. (2025). Towards a standardized methodology and dataset for evaluating LLM-based digital forensic timeline analysis. arXiv preprint arXiv:2505.03100.
Supreme Court of the United States. (1993). Daubert v. Merrell Dow Pharmaceuticals, Inc., 509 U.S. 579.
Tiwari, V., Dasari, V. S. R., & Wang, J. (n.d.). The future of artificial intelligence in forensics: Advancements, challenges, and ethical considerations.
Williams v. City of Detroit, No. 2:21-cv-10827 (E.D. Mich. 2024).
Xie, Q., Hanafiah, M. H., Yang, H., Abdul Hamid, R., & Juhdi, N. H. (2025). Mediating effects of sustainability-oriented innovation on the relationship between organizational creativity and sustainable performance among Chinese manufacturing SMEs. Multidisciplinary Reviews, 8(12), 2025373. https://doi.org/10.31893/multirev.2025373

