Skip navigation

Detecting and preventing data poisoning attacks on AI models

Detecting and preventing data poisoning attacks on AI models

Kure, Halima I., Sarkar, Pradipta, Ndanusa, Ahmed B. and Nwajana, Augustine O. ORCID logoORCID: https://orcid.org/0000-0001-6591-5269 (2025) Detecting and preventing data poisoning attacks on AI models. In: 2025 PhotonIcs & Electromagnetics Research Symposium. IEEE Xplore . Institute of Electrical and Electronics Engineers (IEEE). (In Press)

[thumbnail of Accepted Conference Paper]
Preview
PDF (Accepted Conference Paper)
50474 NWAJANA_Detecting_And_Preventing_Data_Poisoning_Attacks_On_AI_Models_(CONFERENCE PAPER AAM)_2025.pdf - Accepted Version

Download (663kB) | Preview

Abstract

This paper investigates the critical issue of data poisoning attacks on AI models, a growing concern in the ever-evolving landscape of artificial intelligence and cybersecurity. As advanced technology systems become increasingly prevalent across various sectors, the need for robust defence mechanisms against adversarial attacks becomes paramount. The study aims to develop and evaluate novel techniques for detecting and preventing data poisoning attacks, focusing on both theoretical frameworks and practical applications. Through a comprehensive literature review, experimental validation using the CIFAR-10 and Insurance Claims datasets, and the development of innovative algorithms, this paper seeks to enhance the resilience of AI models against malicious data manipulation. The study explores various methods, including anomaly detection, robust optimization strategies, and ensemble learning, to identify and mitigate the effects of poisoned data during model training. Experimental results indicate that data poisoning significantly degrades model performance, reducing classification accuracy by up to 27% in image recognition tasks (CIFAR-10) and 22% in fraud detection models (Insurance Claims dataset). The proposed defence mechanisms, including statistical anomaly detection and adversarial training, successfully mitigated poisoning effects, improving model robustness and restoring accuracy levels by an average of 15-20%. The findings further demonstrate that ensemble learning techniques provide an additional layer of resilience, reducing false positives and false negatives caused by adversarial data injections. The findings of this paper contribute to the broader understanding of adversarial machine learning and provide actionable insights for practitioners in industries facing emerging threats to AI-based systems. By addressing the challenges of data poisoning attacks, this study aims to improve the security and reliability of AI models, ultimately fostering greater trust in AI technologies across critical applications.

Item Type: Conference Proceedings
Title of Proceedings: 2025 PhotonIcs & Electromagnetics Research Symposium
Uncontrolled Keywords: data poisoning, Artificial Intelligence, cybersecurity, machine learning, adversarial attacks, model security
Subjects: Q Science > Q Science (General)
Q Science > QA Mathematics > QA75 Electronic computers. Computer science
T Technology > T Technology (General)
Faculty / School / Research Centre / Research Group: Faculty of Engineering & Science
Faculty of Engineering & Science > School of Engineering (ENG)
Related URLs:
Last Modified: 21 May 2025 11:37
URI: http://gala.gre.ac.uk/id/eprint/50474

Actions (login required)

View Item View Item

Downloads

Downloads per month over past year

View more statistics