Detecting and preventing data poisoning attacks on AI models
Kure, Halima I., Sarkar, Pradipta, Ndanusa, Ahmed B. and Nwajana, Augustine O. ORCID: https://orcid.org/0000-0001-6591-5269
(2025)
Detecting and preventing data poisoning attacks on AI models.
In: 2025 PhotonIcs & Electromagnetics Research Symposium.
IEEE Xplore
.
Institute of Electrical and Electronics Engineers (IEEE).
(In Press)
Preview |
PDF (Accepted Conference Paper)
50474 NWAJANA_Detecting_And_Preventing_Data_Poisoning_Attacks_On_AI_Models_(CONFERENCE PAPER AAM)_2025.pdf - Accepted Version Download (663kB) | Preview |
Abstract
This paper investigates the critical issue of data poisoning attacks on AI models, a growing concern in the ever-evolving landscape of artificial intelligence and cybersecurity. As advanced technology systems become increasingly prevalent across various sectors, the need for robust defence mechanisms against adversarial attacks becomes paramount. The study aims to develop and evaluate novel techniques for detecting and preventing data poisoning attacks, focusing on both theoretical frameworks and practical applications. Through a comprehensive literature review, experimental validation using the CIFAR-10 and Insurance Claims datasets, and the development of innovative algorithms, this paper seeks to enhance the resilience of AI models against malicious data manipulation. The study explores various methods, including anomaly detection, robust optimization strategies, and ensemble learning, to identify and mitigate the effects of poisoned data during model training. Experimental results indicate that data poisoning significantly degrades model performance, reducing classification accuracy by up to 27% in image recognition tasks (CIFAR-10) and 22% in fraud detection models (Insurance Claims dataset). The proposed defence mechanisms, including statistical anomaly detection and adversarial training, successfully mitigated poisoning effects, improving model robustness and restoring accuracy levels by an average of 15-20%. The findings further demonstrate that ensemble learning techniques provide an additional layer of resilience, reducing false positives and false negatives caused by adversarial data injections. The findings of this paper contribute to the broader understanding of adversarial machine learning and provide actionable insights for practitioners in industries facing emerging threats to AI-based systems. By addressing the challenges of data poisoning attacks, this study aims to improve the security and reliability of AI models, ultimately fostering greater trust in AI technologies across critical applications.
Item Type: | Conference Proceedings |
---|---|
Title of Proceedings: | 2025 PhotonIcs & Electromagnetics Research Symposium |
Uncontrolled Keywords: | data poisoning, Artificial Intelligence, cybersecurity, machine learning, adversarial attacks, model security |
Subjects: | Q Science > Q Science (General) Q Science > QA Mathematics > QA75 Electronic computers. Computer science T Technology > T Technology (General) |
Faculty / School / Research Centre / Research Group: | Faculty of Engineering & Science Faculty of Engineering & Science > School of Engineering (ENG) |
Related URLs: | |
Last Modified: | 21 May 2025 11:37 |
URI: | http://gala.gre.ac.uk/id/eprint/50474 |
Actions (login required)
![]() |
View Item |
Downloads
Downloads per month over past year