Adversarial Attacks in Machine Learning: Key Insights and Defense Approaches

Main Article Content

Yahya Layth Khaleel
Mustafa Abdulfattah Habeeb
Hussein Alnabulsi

Abstract

There is a considerable threat present in genres such as machine learning due to adversarial attacks which include purposely feeding the system with data that will alter the decision region. These attacks are committed to presenting different data to machine learning models in a way that the model would be wrong in its classification or prediction. The field of study is still relatively young and has to develop strong bodies of scientific research that would eliminate the gaps in the current knowledge. This paper provides the literature review of adversarial attacks and defenses based on the highly cited articles and conference published in the Scopus database. Through the classification and assessment of 128 systematic articles: 80 original papers and 48 review papers till May 15, 2024, this study categorizes and reviews the literature from different domains, such as Graph Neural Networks, Deep Learning Models for IoT Systems, and others. The review posits findings on identified metrics, citation analysis, and contributions from these studies while suggesting the area’s further research and development for adversarial robustness’ and protection mechanisms. The identified objective of this work is to present the basic background of adversarial attacks and defenses, and the need for maintaining the adaptability of machine learning platforms. In this context, the objective is to contribute to building efficient and sustainable protection mechanisms for AI applications in various industries

Downloads

Download data is not yet available.

Article Details

How to Cite
Khaleel , Y. L., Habeeb , M. A., & Alnabulsi , H. (2024). Adversarial Attacks in Machine Learning: Key Insights and Defense Approaches . Applied Data Science and Analysis, 2024, 121–147. https://doi.org/10.58496/ADSA/2024/011
Section
Articles

References

A. S. Albahri et al., “A systematic review of trustworthy artificial intelligence applications in natural disasters,” Comput. Electr. Eng., vol. 118, p. 109409, 2024, doi: 10.1016/j.compeleceng.2024.109409.

M. A. Habeeb, “Hate Speech Detection using Deep Learning Master thesis,” University of Miskolc, 2021. [Online]. Available: http://midra.uni-miskolc.hu/document/40792/38399.pdf

M. E. Alqaysi, A. S. Albahri, and R. A. Hamid, “Evaluation and benchmarking of hybrid machine learning models for autism spectrum disorder diagnosis using a 2-tuple linguistic neutrosophic fuzzy sets-based decision-making model,” Neural Comput. Appl., 2024, doi: 10.1007/s00521-024-09905-6.

A. H. Alamoodi, M. S. Al-Samarraay, O. S. Albahri, M. Deveci, A. S. Albahri, and S. Yussof, “Evaluation of energy economic optimization models using multi-criteria decision-making approach,” Expert Syst. Appl., vol. 255, p. 124842, 2024, doi: 10.1016/j.eswa.2024.124842.

A. S. Albahri et al., “Prioritizing complex health levels beyond autism triage using fuzzy multi-criteria decision-making,” Complex Intell. Syst., 2024, doi: 10.1007/s40747-024-01432-0.

S. Dadvandipour and Y. L. Khaleel, “Application of deep learning algorithms detecting fake and correct textual or verbal news,” Prod. Syst. Inf. Eng., vol. 10, no. 2, pp. 37–51, 2022, doi: 10.32968/psaie.2022.2.4.

M. A. Habeeb, Y. L. Khaleel, and A. S. Albahri, “Toward Smart Bicycle Safety: Leveraging Machine Learning Models and Optimal Lighting Solutions,” in Proceedings of the Third International Conference on Innovations in Computing Research (ICR’24), K. Daimi and A. Al Sadoon, Eds., Cham: Springer Nature Switzerland, 2024, pp. 120–131.

S. Ghazal, A. Munir, and W. S. Qureshi, “Computer vision in smart agriculture and precision farming: Techniques and applications,” Artif. Intell. Agric., vol. 13, pp. 64–83, 2024, doi: 10.1016/j.aiia.2024.06.004.

Y. L. Khaleel, “Fake News Detection Using Deep Learning,” University of Miskolc, 2021. doi: 10.1007/978-3-030-91305-2_19.

Z. T. Al-qaysi, A. S. Albahri, M. A. Ahmed, and M. M. Salih, “Dynamic decision-making framework for benchmarking brain–computer interface applications: a fuzzy-weighted zero-inconsistency method for consistent weights and VIKOR for stable rank,” Neural Comput. Appl., vol. 36, no. 17, pp. 10355–10378, 2024, doi: 10.1007/s00521-024-09605-1.

A. S. Albahri et al., “Fuzzy decision-making framework for explainable golden multi-machine learning models for real-time adversarial attack detection in Vehicular Ad-hoc Networks,” Inf. Fusion, vol. 105, p. 102208, 2024, doi: 10.1016/j.inffus.2023.102208.

A. H. Alamoodi et al., “Selection of electric bus models using 2-tuple linguistic T-spherical fuzzy-based decision-making model,” Expert Syst. Appl., vol. 249, p. 123498, 2024, doi: https://doi.org/10.1016/j.eswa.2024.123498.

F. K. H. Mihna, M. A. Habeeb, Y. L. Khaleel, Y. H. Ali, and L. A. E. Al-Saeedi, “Using Information Technology for Comprehensive Analysis and Prediction in Forensic Evidence,” Mesopotamian J. CyberSecurity, vol. 4, no. 1, pp. 4–16, 2024, doi: 10.58496/MJCS/2024/002.

A. S. Albahri, Y. L. Khaleel, and M. A. Habeeb, “The Considerations of Trustworthy AI Components in Generative AI; A Letter to Editor,” Appl. Data Sci. Anal., vol. 2023, pp. 108–109, 2023, doi: 10.58496/adsa/2023/009.

L. Alzubaidi et al., “MEFF – A model ensemble feature fusion approach for tackling adversarial attacks in medical imaging,” Intell. Syst. with Appl., vol. 22, 2024, doi: 10.1016/j.iswa.2024.200355.

D. Zügner and S. Günnemann, “Adversarial attacks on graph neural networks via meta learning,” in 7th International Conference on Learning Representations, ICLR 2019, International Conference on Learning Representations, ICLR, 2019. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083950290&partnerID=40&md5=4a963cbdc0747de26316af0093ca9ad3

O. Ibitoye, O. Shafiq, and A. Matrawy, “Analyzing adversarial attacks against deep learning for intrusion detection in IoT networks,” in Proceedings - IEEE Global Communications Conference, GLOBECOM, 2019. doi: 10.1109/GLOBECOM38437.2019.9014337.

M. Haghighatlari, J. Li, F. Heidar-Zadeh, Y. Liu, X. Guan, and T. Head-Gordon, “Learning to Make Chemical Predictions: The Interplay of Feature Representation, Data, and Machine Learning Methods,” Chem, vol. 6, no. 7, pp. 1527–1542, 2020, doi: 10.1016/j.chempr.2020.05.014.

J. G. Greener, S. M. Kandathil, L. Moffat, and D. T. Jones, “A guide to machine learning for biologists,” Nat. Rev. Mol. Cell Biol., vol. 23, no. 1, pp. 40–55, 2022, doi: 10.1038/s41580-021-00407-0.

W. E. Zhang, Q. Z. Sheng, A. Alhazmi, and C. Li, “Adversarial Attacks on Deep-learning Models in Natural Language Processing,” ACM Trans. Intell. Syst. Technol., vol. 11, no. 3, pp. 1–41, 2020, doi: 10.1145/3374217.

X. Ma et al., “Understanding adversarial attacks on deep learning based medical image analysis systems,” Pattern Recognit., vol. 110, p. 107332, 2021, doi: 10.1016/j.patcog.2020.107332.

M. Shen, H. Yu, L. Zhu, K. Xu, Q. Li, and J. Hu, “Effective and Robust Physical-World Attacks on Deep Learning Face Recognition Systems,” IEEE Trans. Inf. Forensics Secur., vol. 16, pp. 4063–4077, 2021, doi: 10.1109/TIFS.2021.3102492.

K. Roshan and A. Zafar, “Black-box adversarial transferability: An empirical study in cybersecurity perspective,” Comput. Secur., vol. 141, 2024, doi: 10.1016/j.cose.2024.103853.

K. Roshan, A. Zafar, and S. B. Ul Haque, “Untargeted White-box Adversarial Attack with Heuristic Defence Methods in Real-time Deep Learning based Network Intrusion Detection System,” Comput. Commun., 2023, doi: https://doi.org/10.1016/j.comcom.2023.09.030.

F. Yu, L. Wang, X. Fang, and Y. Zhang, “The defense of adversarial example with conditional generative adversarial networks,” Secur. Commun. Networks, vol. 2020, no. 1, p. 3932584, 2020, doi: 10.1155/2020/3932584.

H. Liang, E. He, Y. Zhao, Z. Jia, and H. Li, “Adversarial Attack and Defense: A Survey,” Electron., vol. 11, no. 8, 2022, doi: 10.3390/electronics11081283.

A. Chakraborty, M. Alam, V. Dey, A. Chattopadhyay, and D. Mukhopadhyay, “A survey on adversarial attacks and defences,” CAAI Trans. Intell. Technol., vol. 6, no. 1, pp. 25–45, 2021, doi: 10.1049/cit2.12028.

K. Mahmood, R. Mahmood, E. Rathbun, and M. Van Dijk, “Back in Black: A Comparative Evaluation of Recent State-Of-The-Art Black-Box Attacks,” IEEE Access, vol. 10, pp. 998–1019, 2022, doi: 10.1109/ACCESS.2021.3138338.

Y. Dong, S. Cheng, T. Pang, H. Su, and J. Zhu, “Query-Efficient Black-Box Adversarial Attacks Guided by a Transfer-Based Prior,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 44, no. 12, pp. 9536–9548, 2022, doi: 10.1109/TPAMI.2021.3126733.

G. Yang, Q. Ye, and J. Xia, “Unbox the black-box for the medical explainable AI via multi-modal and multi-centre data fusion: A mini-review, two showcases and beyond,” Inf. Fusion, vol. 77, pp. 29–52, 2022, doi: 10.1016/j.inffus.2021.07.016.

E. Mariotti, “A holistic perspective on designing and evaluating explainable AI models: from white-box additive models to post-hoc explanations for black-box models.” 2024.

S. Ai, A. S. Voundi Koe, and T. Huang, “Adversarial perturbation in remote sensing image recognition,” Appl. Soft Comput., vol. 105, p. 107252, 2021, doi: 10.1016/j.asoc.2021.107252.

C. Zhang, X. Costa-Perez, and P. Patras, “Adversarial Attacks Against Deep Learning-Based Network Intrusion Detection Systems and Defense Mechanisms,” IEEE/ACM Trans. Netw., vol. 30, no. 3, pp. 1294–1311, 2022, doi: 10.1109/TNET.2021.3137084.

B. Wu et al., “Attacking Adversarial Attacks as A Defense,” arXiv Prepr. arXiv2106.04938, 2021, [Online]. Available: http://arxiv.org/abs/2106.04938

N. Liu, M. Du, R. Guo, H. Liu, and X. Hu, “Adversarial Attacks and Defenses,” ACM SIGKDD Explor. Newsl., vol. 23, no. 1, pp. 86–99, May 2021, doi: 10.1145/3468507.3468519.

L. Griffin, “Evaluating Methods for Improving DNN Robustness Against Adversarial Attacks,” no. 1. University of South Florida, pp. 1–23, 2023. [Online]. Available: https://www.proquest.com/openview/0a3e9e510f3b25b0516f4b623af4423f/1?pq-origsite=gscholar&cbl=18750&diss=y

Y. L. Khaleel, M. A. Habeeb, A. S. Albahri, T. Al-Quraishi, O. S. Albahri, and A. H. Alamoodi, “Network and cybersecurity applications of defense in adversarial attacks: A state-of-the-art using machine learning and deep learning methods,” vol. 33, no. 1, 2024, doi: doi:10.1515/jisys-2024-0153.

M. Macas, C. Wu, and W. Fuertes, “Adversarial examples: A survey of attacks and defenses in deep learning-enabled cybersecurity systems,” Expert Syst. Appl., vol. 238, p. 122223, Mar. 2024, doi: 10.1016/j.eswa.2023.122223.

K. Ren, T. Zheng, Z. Qin, and X. Liu, “Adversarial Attacks and Defenses in Deep Learning,” Engineering, vol. 6, no. 3, pp. 346–360, 2020, doi: 10.1016/j.eng.2019.12.012.

J. Chen, X. Wu, Y. Guo, Y. Liang, and S. Jha, “Towards Evaluating the Robustness of Neural Networks Learned By Transduction,” in ICLR 2022 - 10th International Conference on Learning Representations, Ieee, 2022, pp. 39–57.

A. Ilyas, S. Santurkar, D. Tsipras, L. Engstrom, B. Tran, and A. Madry, “Adversarial examples are not bugs, they are features,” Adv. Neural Inf. Process. Syst., vol. 32, 2019.

X. Zhang and M. Zitnik, “GNNGUARD: Defending graph neural networks against adversarial attacks,” in Advances in Neural Information Processing Systems, Neural information processing systems foundation, 2020. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85104609818&partnerID=40&md5=7112b4afbafe82c4d30a023c4eeb8dc3

H. Wu, C. Wang, Y. Tyshetskiy, A. Docherty, K. Lu, and L. Zhu, “Adversarial examples for graph data: Deep insights into attack and defense,” in IJCAI International Joint Conference on Artificial Intelligence, K. S., Ed., International Joint Conferences on Artificial Intelligence, 2019, pp. 4816–4823. doi: 10.24963/ijcai.2019/669.

H. Dai et al., “Adversarial attack on graph structured data,” in 35th International Conference on Machine Learning, ICML 2018, K. A. and D. J., Eds., International Machine Learning Society (IMLS), 2018, pp. 1799–1808. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85057260187&partnerID=40&md5=f27682ae15830c7b87e1ffcf27e737aa

D. Zügner, A. Akbarnejad, and S. Günnemann, “Adversarial attacks on neural networks for graph data,” in Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Association for Computing Machinery, 2018, pp. 2847–2856. doi: 10.1145/3219819.3220078.

Y. Dong et al., “Efficient decision-based black-box adversarial attacks on face recognition,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, IEEE Computer Society, 2019, pp. 7706–7714. doi: 10.1109/CVPR.2019.00790.

S. G. Finlayson, J. D. Bowers, J. Ito, J. L. Zittrain, A. L. Beam, and I. S. Kohane, “Adversarial attacks on medical machine learning,” Science (80-. )., vol. 363, no. 6433, pp. 1287–1289, 2019, doi: 10.1126/science.aaw4399.

A. Madry, A. Makelov, L. Schmidt, D. Tsipras, and A. Vladu, “Towards deep learning models resistant to adversarial attacks,” in 6th International Conference on Learning Representations, ICLR 2018 - Conference Track Proceedings, International Conference on Learning Representations, ICLR, 2018. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083954061&partnerID=40&md5=84bf66031966d7b8f24a2260e59ff64c

R. Shao, X. Lan, J. Li, and P. C. Yuen, “Multi-adversarial discriminative deep domain generalization for face presentation attack detection,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, IEEE Computer Society, 2019, pp. 10015–10023. doi: 10.1109/CVPR.2019.01026.

M. Sadeghi and E. G. Larsson, “Adversarial attacks on deep-learning based radio signal classification,” IEEE Wirel. Commun. Lett., vol. 8, no. 1, pp. 213–216, 2019, doi: 10.1109/LWC.2018.2867459.

L. Schönherr, K. Kohls, S. Zeiler, T. Holz, and D. Kolossa, “Adversarial Attacks Against Automatic Speech Recognition Systems via Psychoacoustic Hiding,” in 26th Annual Network and Distributed System Security Symposium, NDSS 2019, The Internet Society, 2019. doi: 10.14722/ndss.2019.23288.

G. Goswami, N. Ratha, A. Agarwal, R. Singh, and M. Vatsa, “Unravelling robustness of deep learning based face recognition against adversarial attacks,” in 32nd AAAI Conference on Artificial Intelligence, AAAI 2018, AAAI press, 2018, pp. 6829–6836. doi: 10.1609/aaai.v32i1.12341.

S. Huang, N. Papernot, I. Goodfellow, Y. Duan, and P. Abbeel, “Adversarial attacks on neural network policies,” in 5th International Conference on Learning Representations, ICLR 2017 - Workshop Track Proceedings, International Conference on Learning Representations, ICLR, 2017. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85115871809&partnerID=40&md5=84171d54c437457e576e50692d1b7e04

M. Yu, M. Zhou, and W. Su, “A secure routing protocol against byzantine attacks for MANETs in adversarial environments,” IEEE Trans. Veh. Technol., vol. 58, no. 1, pp. 449–460, 2009, doi: 10.1109/TVT.2008.923683.

I. Corona, G. Giacinto, and F. Roli, “Adversarial attacks against intrusion detection systems: Taxonomy, solutions and open issues,” Inf. Sci. (Ny)., vol. 239, pp. 201–225, 2013, doi: 10.1016/j.ins.2013.03.022.

H. Qiu, T. Dong, T. Zhang, J. Lu, G. Memmi, and M. Qiu, “Adversarial Attacks against Network Intrusion Detection in IoT Systems,” IEEE Internet Things J., vol. 8, no. 13, pp. 10327–10335, 2021, doi: 10.1109/JIOT.2020.3048038.

S. Cheng, Y. Dong, T. Pang, H. Su, and J. Zhu, “Improving black-box adversarial attacks with a transfer-based prior,” in Advances in Neural Information Processing Systems, Neural information processing systems foundation, 2019. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090169867&partnerID=40&md5=9dde3ef3249f74f65c428131ed49734c

Q. Huang, I. Katsman, Z. Gu, H. He, S. Belongie, and S. N. Lim, “Enhancing adversarial example transferability with an intermediate level attack,” in Proceedings of the IEEE International Conference on Computer Vision, Institute of Electrical and Electronics Engineers Inc., 2019, pp. 4732–4741. doi: 10.1109/ICCV.2019.00483.

Y. Liu, X. Chen, C. Liu, and D. Song, “Delving into transferable adversarial examples and black-box attacks,” in 5th International Conference on Learning Representations, ICLR 2017 - Conference Track Proceedings, International Conference on Learning Representations, ICLR, 2017. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088225756&partnerID=40&md5=84bfcdfc146d3ce2983bbd860c3547ce

A. Demontis et al., “Why do adversarial attacks transfer? Explaining transferability of evasion and poisoning attacks,” in Proceedings of the 28th USENIX Security Symposium, USENIX Association, 2019, pp. 321–338. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85072900690&partnerID=40&md5=ebe100a708426ec0b1edeef53c787f14

S. Thys, W. Van Ranst, and T. Goedeme, “Fooling automated surveillance cameras: Adversarial patches to attack person detection,” in IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, IEEE Computer Society, 2019, pp. 49–55. doi: 10.1109/CVPRW.2019.00012.

N. Narodytska and S. Kasiviswanathan, “Simple Black-Box Adversarial Attacks on Deep Neural Networks,” in IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, IEEE Computer Society, 2017, pp. 1310–1318. doi: 10.1109/CVPRW.2017.172.

X. Zhou, W. Liang, W. Li, K. Yan, S. Shimizu, and K. I. K. Wang, “Hierarchical Adversarial Attacks Against Graph-Neural-Network-Based IoT Network Intrusion Detection System,” IEEE Internet Things J., vol. 9, no. 12, pp. 9310–9319, 2022, doi: 10.1109/JIOT.2021.3130434.

F. Croce and M. Hein, “Minimally distorted adversarial examples with a fast adaptive boundary attack,” in 37th International Conference on Machine Learning, ICML 2020, D. H. and S. A., Eds., International Machine Learning Society (IMLS), 2020, pp. 2174–2183. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85104186671&partnerID=40&md5=7a4457680c875323e63ed2b49b1b8402

M. Andriushchenko, F. Croce, N. Flammarion, and M. Hein, “Square Attack: A Query-Efficient Black-Box Adversarial Attack via Random Search,” Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics), vol. 12368 LNCS, pp. 484–501, 2020, doi: 10.1007/978-3-030-58592-1_29.

A. Ilyas, L. Engstrom, and A. Madry, “Prior convictions: Black-box adversarial attacks with bandits and priors,” in 7th International Conference on Learning Representations, ICLR 2019, International Conference on Learning Representations, ICLR, 2019. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083953118&partnerID=40&md5=d6ac28d6fc01771e2ac0f95812dae7b7

C. Guo, J. R. Gardner, Y. You, A. G. Wilson, and K. Q. Weinberger, “Simple black-box adversarial attacks,” in 36th International Conference on Machine Learning, ICML 2019, International Machine Learning Society (IMLS), 2019, pp. 4410 – 4423. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85073203396&partnerID=40&md5=dc3fedb1bf2c908a3b08ba70db8571eb

F. Tramèr, A. Kurakin, N. Papernot, I. Goodfellow, D. Boneh, and P. McDaniel, “Ensemble adversarial training: Attacks and defenses,” in 6th International Conference on Learning Representations, ICLR 2018 - Conference Track Proceedings, International Conference on Learning Representations, ICLR, 2018. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083953449&partnerID=40&md5=bf2f71b308c2b1ad53207565cf750d26

F. Croce and M. Hein, “Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks,” in 37th International Conference on Machine Learning, ICML 2020, D. H. and S. A., Eds., International Machine Learning Society (IMLS), 2020, pp. 2184–2194. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85105183209&partnerID=40&md5=f4d452a7ea209d57876bae3aed272167

P. Samangouei, M. Kabkab, and R. Chellappa, “Defense-Gan: Protecting classifiers against adversarial attacks using generative models,” in 6th International Conference on Learning Representations, ICLR 2018 - Conference Track Proceedings, International Conference on Learning Representations, ICLR, 2018. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083952288&partnerID=40&md5=89d3e0934167ea8524dd3ea46d84758e

H. Zhang and J. Wang, “Defense against adversarial attacks using feature scattering-based adversarial training,” in Advances in Neural Information Processing Systems, Neural information processing systems foundation, 2019. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085197234&partnerID=40&md5=ad449f48773632064dc822eef014bb04

S. Chen et al., “Automated poisoning attacks and defenses in malware detection systems: An adversarial machine learning approach,” Comput. Secur., vol. 73, pp. 326–344, 2018, doi: 10.1016/j.cose.2017.11.007.

F. Pierazzi, F. Pendlebury, J. Cortellazzi, and L. Cavallaro, “Intriguing properties of adversarial ML attacks in the problem space,” in Proceedings - IEEE Symposium on Security and Privacy, 2020, pp. 1332–1349. doi: 10.1109/SP40000.2020.00073.

N. Entezari, S. A. Al-Sayouri, A. Darvishzadeh, and E. E. Papalexakis, “All you need is Low (rank): Defending against adversarial attacks on graphs,” in WSDM 2020 - Proceedings of the 13th International Conference on Web Search and Data Mining, Association for Computing Machinery, Inc, 2020, pp. 169–177. doi: 10.1145/3336191.3371789.

K. Lee, K. Lee, H. Lee, and J. Shin, “A simple unified framework for detecting out-of-distribution samples and adversarial attacks,” in Advances in Neural Information Processing Systems, G. K., B. S., W. H., G. K., L. H., C.-B. N., and G. R., Eds., Neural information processing systems foundation, 2018, pp. 7167–7177. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85064822451&partnerID=40&md5=ed7d4489422d4e8f072fb9146ddf8761

A. Pattanaik, Z. Tang, S. Liu, G. Bommannan, and G. Chowdhary, “Robust Deep Reinforcement Learning with adversarial attacks,” in Proceedings of the International Joint Conference on Autonomous Agents and Multiagent Systems, AAMAS, International Foundation for Autonomous Agents and Multiagent Systems (IFAAMAS), 2018, pp. 2040–2042. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85054762037&partnerID=40&md5=30fcb05b34d27197cdd5dafbddd136ad

Y. Zhao, H. Zhu, R. Liang, Q. Shen, S. Zhang, and K. Chen, “Seeing isn’t believing: Towards more robust adversarial attack against real world object detectors,” in Proceedings of the ACM Conference on Computer and Communications Security, Association for Computing Machinery, 2019, pp. 1989–2004. doi: 10.1145/3319535.3354259.

Z. He, A. S. Rakin, and D. Fan, “Parametric noise injection: Trainable randomness to improve deep neural network robustness against adversarial attack,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, IEEE Computer Society, 2019, pp. 588–597. doi: 10.1109/CVPR.2019.00068.

F. Zhang, P. P. K. Chan, B. Biggio, D. S. Yeung, and F. Roli, “Adversarial Feature Selection Against Evasion Attacks,” IEEE Trans. Cybern., vol. 46, no. 3, pp. 766–777, 2016, doi: 10.1109/TCYB.2015.2415032.

P. Y. Chen, Y. Sharma, H. Zhang, J. Yi, and C. J. Hsieh, “EAD: Elastic-net attacks to deep neural networks via adversarial examples,” in 32nd AAAI Conference on Artificial Intelligence, AAAI 2018, AAAI press, 2018, pp. 10–17. doi: 10.1609/aaai.v32i1.11302.

Y. Dong, T. Pang, H. Su, and J. Zhu, “Evading defenses to transferable adversarial examples by translation-invariant attacks,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Jun. 2019, pp. 4307–4316. doi: 10.1109/CVPR.2019.00444.

D. Slack, S. Hilgard, E. Jia, S. Singh, and H. Lakkaraju, “Fooling LIME and SHAP: Adversarial attacks on post hoc explanation methods,” in AIES 2020 - Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, Association for Computing Machinery, Inc, 2020, pp. 180–186. doi: 10.1145/3375627.3375830.

Y. C. Lin, Z. W. Hong, Y. H. Liao, M. L. Shih, M. Y. Liu, and M. Sun, “Tactics of adversarial attack on deep reinforcement learning agents,” in 5th International Conference on Learning Representations, ICLR 2017 - Workshop Track Proceedings, International Conference on Learning Representations, ICLR, 2017. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85106092143&partnerID=40&md5=fdb1f5623b6062f9121010a12a700836

Y. C. Lin, Z. W. Hong, Y. H. Liao, M. L. Shih, M. Y. Liu, and M. Sun, “Tactics of adversarial attack on deep reinforcement learning agents,” in IJCAI International Joint Conference on Artificial Intelligence, S. C., Ed., International Joint Conferences on Artificial Intelligence, 2017, pp. 3756–3762. doi: 10.24963/ijcai.2017/525.

N. Carlini and D. Wagner, “Audio adversarial examples: Targeted attacks on speech-to-text,” in Proceedings - 2018 IEEE Symposium on Security and Privacy Workshops, SPW 2018, Institute of Electrical and Electronics Engineers Inc., 2018, pp. 1–7. doi: 10.1109/SPW.2018.00009.

S. Pawar, S. El Rouayheb, and K. Ramchandran, “Securing dynamic distributed storage systems against eavesdropping and adversarial attacks,” IEEE Trans. Inf. Theory, vol. 57, no. 10, pp. 6734–6753, 2011, doi: 10.1109/TIT.2011.2162191.

A. Eyas, L. Engstrom, A. Athalye, and J. Lin, “Black-box adversarial attacks with limited queries and information,” in 35th International Conference on Machine Learning, ICML 2018, D. J. and K. A., Eds., International Machine Learning Society (IMLS), 2018, pp. 3392–3401. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85055122654&partnerID=40&md5=d545b57a0d5af74d7ecf91c57218924e

Y. Cao et al., “Adversarial sensor attack on LiDAR-based perception in autonomous driving,” in Proceedings of the ACM Conference on Computer and Communications Security, Association for Computing Machinery, 2019, pp. 2267–2281. doi: 10.1145/3319535.3339815.

X. Wang and K. He, “Enhancing the Transferability of Adversarial Attacks through Variance Tuning,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, IEEE Computer Society, 2021, pp. 1924–1933. doi: 10.1109/CVPR46437.2021.00196.

A. Bojchevski and S. Günnemann, “Adversarial attacks on node embeddings via graph poisoning,” in 36th International Conference on Machine Learning, ICML 2019, International Machine Learning Society (IMLS), 2019, pp. 1112–1123. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071150199&partnerID=40&md5=fbaebf786559a6129c43664c9c5cba7e

R. Duan, X. Ma, Y. Wang, J. Bailey, A. K. Qin, and Y. Yang, “Adversarial camouflage: Hiding physical-world attacks with natural styles,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, IEEE Computer Society, 2020, pp. 997–1005. doi: 10.1109/CVPR42600.2020.00108.

Y. Dong et al., “Boosting Adversarial Attacks with Momentum,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2018, pp. 9185–9193. doi: 10.1109/CVPR.2018.00957.

J. Zhang et al., “Attacks which do not kill training make adversarial learning stronger,” in 37th International Conference on Machine Learning, ICML 2020, D. H. and S. A., Eds., International Machine Learning Society (IMLS), 2020, pp. 11214–11224. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85105322490&partnerID=40&md5=6e99952b1c309b7719291c84f91bd689

F. Tramèr, N. Carlini, W. Brendel, and A. Madry, “On adaptive attacks to adversarial example defenses,” in Advances in Neural Information Processing Systems, Neural information processing systems foundation, 2020. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85100497968&partnerID=40&md5=1a2f0668e1a16348fc1f94bd242ee078

M. S. Chong, M. Wakaiki, and J. P. Hespanha, “Observability of linear systems under adversarial attacks,” in Proceedings of the American Control Conference, Institute of Electrical and Electronics Engineers Inc., 2015, pp. 2439–2444. doi: 10.1109/ACC.2015.7171098.

Y. Lin, H. Zhao, X. Ma, Y. Tu, and M. Wang, “Adversarial Attacks in Modulation Recognition with Convolutional Neural Networks,” IEEE Trans. Reliab., vol. 70, no. 1, pp. 389–401, 2021, doi: 10.1109/TR.2020.3032744.

D. Zhu, P. Cui, Z. Zhang, and W. Zhu, “Robust graph convolutional networks against adversarial attacks,” in Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Association for Computing Machinery, 2019, pp. 1399–1407. doi: 10.1145/3292500.3330851.

D. J. Miller, Z. Xiang, and G. Kesidis, “Adversarial Learning Targeting Deep Neural Network Classification: A Comprehensive Review of Defenses against Attacks,” Proc. IEEE, vol. 108, no. 3, pp. 402–433, 2020, doi: 10.1109/JPROC.2020.2970615.

J. Jia, A. Salem, M. Backes, Y. Zhang, and N. Z. Gong, “Memguard: Defending against black-box membership inference attacks via adversarial examples,” in Proceedings of the ACM Conference on Computer and Communications Security, Association for Computing Machinery, 2019, pp. 259–274. doi: 10.1145/3319535.3363201.

A. Arnab, O. Miksik, and P. H. S. Torr, “On the robustness of semantic segmentation models to adversarial attacks,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 42, no. 12, pp. 3040–3053, 2020, doi: 10.1109/TPAMI.2019.2919707.

H. Xiao, H. Xiao, and C. Eckert, “Adversarial label flips attack on support vector machines,” Front. Artif. Intell. Appl., vol. 242, pp. 870–875, 2012, doi: 10.3233/978-1-61499-098-7-870.

S. Baluja and I. Fischer, “Learning to attack: Adversarial transformation networks,” in 32nd AAAI Conference on Artificial Intelligence, AAAI 2018, AAAI press, 2018, pp. 2687–2695. doi: 10.1609/aaai.v32i1.11672.

J. Zhang, J. Chen, D. Wu, B. Chen, and S. Yu, “Poisoning attack in federated learning using generative adversarial nets,” in Proceedings - 2019 18th IEEE International Conference on Trust, Security and Privacy in Computing and Communications/13th IEEE International Conference on Big Data Science and Engineering, TrustCom/BigDataSE 2019, Institute of Electrical and Electronics Engineers Inc., 2019, pp. 374–380. doi: 10.1109/TrustCom/BigDataSE.2019.00057.

J. Uesato, B. O’Donoghue, A. Van Den Oord, and P. Kohli, “Adversarial risk and the dangers of evaluating against weak attacks,” in 35th International Conference on Machine Learning, ICML 2018, K. A. and D. J., Eds., International Machine Learning Society (IMLS), 2018, pp. 7995–8007. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85057301673&partnerID=40&md5=53ebff6be03140645e37b4d3a505d486

J. Rony, L. G. Hafemann, L. S. Oliveira, I. Ben Ayed, R. Sabourin, and E. Granger, “Decoupling direction and norm for efficient gradient-based l2 adversarial attacks and defenses,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, IEEE Computer Society, 2019, pp. 4317–4325. doi: 10.1109/CVPR.2019.00445.

H. Fawzi, P. Tabuada, and S. Diggavi, “Secure estimation and control for cyber-physical systems under adversarial attacks,” IEEE Trans. Automat. Contr., vol. 59, no. 6, pp. 1454–1467, 2014, doi: 10.1109/TAC.2014.2303233.

S. T. Chen, C. Cornelius, J. Martin, and D. H. P. Chau, “ShapeShifter: Robust physical adversarial attack on faster R-CNN object detector,” Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics), vol. 11051 LNAI, pp. 52–68, 2019, doi: 10.1007/978-3-030-10925-7_4.

X. Yuan, P. He, Q. Zhu, and X. Li, “Adversarial Examples: Attacks and Defenses for Deep Learning,” IEEE Trans. neural networks Learn. Syst., vol. 30, no. 9, pp. 2805–2824, 2019, doi: 10.1109/TNNLS.2018.2886017.

M. Usama, M. Asim, S. Latif, J. Qadir, and Ala-Al-Fuqaha, “Generative adversarial networks for launching and thwarting adversarial attacks on network intrusion detection systems,” in 2019 15th International Wireless Communications and Mobile Computing Conference, IWCMC 2019, Institute of Electrical and Electronics Engineers Inc., 2019, pp. 78–83. doi: 10.1109/IWCMC.2019.8766353.

Z. Wang, H. Guo, Z. Zhang, W. Liu, Z. Qin, and K. Ren, “Feature Importance-aware Transferable Adversarial Attacks,” in Proceedings of the IEEE International Conference on Computer Vision, Institute of Electrical and Electronics Engineers Inc., 2021, pp. 7619–7628. doi: 10.1109/ICCV48922.2021.00754.

J. Sun, Y. Cao, Q. A. Chen, and Z. Morley Mao, “Towards robust LiDAR-based perception in autonomous driving: General black-box adversarial sensor attack and countermeasures,” in Proceedings of the 29th USENIX Security Symposium, USENIX Association, 2020, pp. 877–894. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091956673&partnerID=40&md5=d2c5942e1a7d98b8faa74b1761e19f25

A. Prakash, N. Moran, S. Garber, A. Dilillo, and J. Storer, “Deflecting Adversarial Attacks with Pixel Deflection,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, IEEE Computer Society, 2018, pp. 8571–8580. doi: 10.1109/CVPR.2018.00894.

W. Brendel, J. Rauber, and M. Bethge, “Decision-based adversarial attacks: Reliable attacks against black-box machine learning models,” in 6th International Conference on Learning Representations, ICLR 2018 - Conference Track Proceedings, International Conference on Learning Representations, ICLR, 2018. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083954048&partnerID=40&md5=d2646cdf5c872ab2eaf5119b5338131f

J. X. Morris, E. Lifland, J. Y. Yoo, J. Grigsby, D. Jin, and Y. Qi, “TextAttack: A Framework for Adversarial Attacks, Data Augmentation, and Adversarial Training in NLP,” in EMNLP 2020 - Conference on Empirical Methods in Natural Language Processing, Proceedings of Systems Demonstrations, L. Q. and S. D., Eds., Association for Computational Linguistics (ACL), 2020, pp. 119–126. doi: 10.18653/v1/2020.emnlp-demos.16.

L. Li, R. Ma, Q. Guo, X. Xue, and X. Qiu, “BERT-ATTACK: Adversarial attack against BERT using BERT,” in EMNLP 2020 - 2020 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference, Association for Computational Linguistics (ACL), 2020, pp. 6193–6202. doi: 10.18653/v1/2020.emnlp-main.500.

F. Liao, M. Liang, Y. Dong, T. Pang, X. Hu, and J. Zhu, “Defense Against Adversarial Attacks Using High-Level Representation Guided Denoiser,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Jun. 2018, pp. 1778–1787. doi: 10.1109/CVPR.2018.00191.

J. Lin, C. Song, K. He, L. Wang, and J. E. Hopcroft, “Nesterov Accelerated Gradient and Scale Invariance for Adversarial Attacks,” in 8th International Conference on Learning Representations, ICLR 2020, International Conference on Learning Representations, ICLR, 2020. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85150590715&partnerID=40&md5=732ac1f089eacc316281df94cc07ea5a

Y. Wang, Y. an Tan, W. Zhang, Y. Zhao, and X. Kuang, “An adversarial attack on DNN-based black-box object detectors,” J. Netw. Comput. Appl., vol. 161, 2020, doi: 10.1016/j.jnca.2020.102634.

S. M. K. A. Kazmi, N. Aafaq, M. A. Khan, M. Khalil, and A. Saleem, “From Pixel to Peril: Investigating Adversarial Attacks on Aerial Imagery Through Comprehensive Review and Prospective Trajectories,” IEEE Access, vol. 11, pp. 81256–81278, 2023, doi: 10.1109/ACCESS.2023.3299878.

J. Fang, Y. Jiang, C. Jiang, Z. L. Jiang, C. Liu, and S. M. Yiu, “State-of-the-art optical-based physical adversarial attacks for deep learning computer vision systems,” Expert Syst. Appl., vol. 250, 2024, doi: 10.1016/j.eswa.2024.123761.

T. Long, Q. Gao, L. Xu, and Z. Zhou, “A survey on adversarial attacks in computer vision: Taxonomy, visualization and future directions,” Comput. Secur., vol. 121, 2022, doi: 10.1016/j.cose.2022.102847.

X. Ling et al., “Adversarial attacks against Windows PE malware detection: A survey of the state-of-the-art,” Comput. Secur., vol. 128, 2023, doi: 10.1016/j.cose.2023.103134.

N. Akhtar, A. Mian, N. Kardan, and M. Shah, “Advances in Adversarial Attacks and Defenses in Computer Vision: A Survey,” IEEE Access, vol. 9, pp. 155161–155196, 2021, doi: 10.1109/ACCESS.2021.3127960.

C. Li, H. Wang, W. Yao, and T. Jiang, “Adversarial attacks in computer vision: a survey,” J. Membr. Comput., vol. 6, no. 2, pp. 130–147, 2024, doi: 10.1007/s41965-024-00142-3.

H. Zheng et al., “Survey of Adversarial Attack, Defense and Robustness Analysis for Natural Language Processing,” Jisuanji Yanjiu yu Fazhan/Computer Res. Dev., vol. 58, no. 8, pp. 1727–1750, 2021, doi: 10.7544/issn1000-1239.2021.20210304.

H. Tan, L. Wang, H. Zhang, J. Zhang, M. Shafiq, and Z. Gu, “Adversarial Attack and Defense Strategies of Speaker Recognition Systems: A Survey,” Electron., vol. 11, no. 14, 2022, doi: 10.3390/electronics11142183.

C. Wei, M. Sun, X. Zou, and X. Zhang, “Reviews on the Attack and Defense Methods of Voice Adversarial Examples,” J. Cyber Secur., vol. 7, no. 1, pp. 100–113, 2022, doi: 10.19363/J.cnki.cn10-1380/tn.2022.01.07.

J. Gao, D. Yan, and M. Dong, “Black-box adversarial attacks through speech distortion for speech emotion recognition,” Eurasip J. Audio, Speech, Music Process., vol. 2022, no. 1, 2022, doi: 10.1186/s13636-022-00254-7.

D. Xu, R. Fang, B. Jiang, and Q. Xuan, “A Review of Speech Adversarial Attack and Defense Methods,” J. Cyber Secur., vol. 7, no. 1, pp. 126–144, 2022, doi: 10.19363/J.cnki.cn10-1380/tn.2022.01.09.

A. Alotaibi and M. A. Rassam, “Adversarial Machine Learning Attacks against Intrusion Detection Systems: A Survey on Strategies and Defense,” Futur. Internet, vol. 15, no. 2, 2023, doi: 10.3390/fi15020062.

J. Li, Y. Wang, L. Luo, and Y. Wang, “A Survey of Adversarial Attack Techniques for Android Malware Detection,” J. Cyber Secur., vol. 6, no. 4, pp. 28–43, 2021, doi: 10.19363/J.cnki.cn10-1380/tn.2021.07.02.

H. Khazane, M. Ridouani, F. Salahdine, and N. Kaabouch, “A Holistic Review of Machine Learning Adversarial Attacks in IoT Networks,” Futur. Internet, vol. 16, no. 1, 2024, doi: 10.3390/fi16010032.

J. Vitorino, I. Praça, and E. Maia, “SoK: Realistic adversarial attacks and defenses for intelligent network intrusion detection,” Comput. Secur., vol. 134, 2023, doi: 10.1016/j.cose.2023.103433.

A. M. Zbrzezny and A. E. Grzybowski, “Deceptive Tricks in Artificial Intelligence: Adversarial Attacks in Ophthalmology,” J. Clin. Med., vol. 12, no. 9, 2023, doi: 10.3390/jcm12093266.

G. W. Muoka et al., “A Comprehensive Review and Analysis of Deep Learning-Based Medical Image Adversarial Attack and Defense,” Mathematics, vol. 11, no. 20, 2023, doi: 10.3390/math11204272.

V. Sorin, S. Soffer, B. S. Glicksberg, Y. Barash, E. Konen, and E. Klang, “Adversarial attacks in radiology – A systematic review,” Eur. J. Radiol., vol. 167, 2023, doi: 10.1016/j.ejrad.2023.111085.

V. Srinivasan, C. Rohrer, A. Marban, K. R. Müller, W. Samek, and S. Nakajima, “Robustifying models against adversarial attacks by Langevin dynamics,” Neural Networks, vol. 137, pp. 1–17, 2021, doi: 10.1016/j.neunet.2020.12.024.

M. Xia, Z. Ye, W. Zhao, R. Yi, and Y. Liu, “Adversarial attack and interpretability of the deep neural net- work from the geometric perspective,” Sci. Sin. Informationis, vol. 51, no. 9, pp. 1411–1437, 2021, doi: 10.1360/SSI-2020-0169.

X. Han, Y. Zhang, W. Wang, and B. Wang, “Text Adversarial Attacks and Defenses: Issues, Taxonomy, and Perspectives,” Secur. Commun. Networks, vol. 2022, 2022, doi: 10.1155/2022/6458488.

A. Kloukiniotis, A. Papandreou, A. Lalos, P. Kapsalas, D. V. Nguyen, and K. Moustakas, “Countering Adversarial Attacks on Autonomous Vehicles Using Denoising Techniques: A Review,” IEEE Open J. Intell. Transp. Syst., vol. 3, pp. 61–80, 2022, doi: 10.1109/OJITS.2022.3142612.

A. K. Sahu and S. Kar, “Decentralized Zeroth-Order Constrained Stochastic Optimization Algorithms: Frank-Wolfe and Variants with Applications to Black-Box Adversarial Attacks,” Proc. IEEE, vol. 108, no. 11, pp. 1890–1905, 2020, doi: 10.1109/JPROC.2020.3012609.

A. K. Nair, E. D. Raj, and J. Sahoo, “A robust analysis of adversarial attacks on federated learning environments,” Comput. Stand. Interfaces, vol. 86, 2023, doi: 10.1016/j.csi.2023.103723.

J. Chen, X. Lin, Z. Shi, and Y. Liu, “Link Prediction Adversarial Attack Via Iterative Gradient Attack,” IEEE Trans. Comput. Soc. Syst., vol. 7, no. 4, pp. 1081–1094, 2020, doi: 10.1109/TCSS.2020.3004059.

S. Almutairi and A. Barnawi, “Securing DNN for smart vehicles: an overview of adversarial attacks, defenses, and frameworks,” J. Eng. Appl. Sci., vol. 70, no. 1, 2023, doi: 10.1186/s44147-023-00184-x.

H. Xu et al., “Adversarial Attacks and Defenses in Images, Graphs and Text: A Review,” Int. J. Autom. Comput., vol. 17, no. 2, pp. 151–178, 2020, doi: 10.1007/s11633-019-1211-x.

Q. Li, C. Lin, Y. Yang, C. Shen, and L. Fang, “Adversarial Attacks and Defenses Against Deep Learning Under the Cloud-Edge-Terminal Scenes,” Jisuanji Yanjiu yu Fazhan/Computer Res. Dev., vol. 59, no. 10, pp. 2109–2129, 2022, doi: 10.7544/issn1000-1239.20220665.

S. Kaviani, K. J. Han, and I. Sohn, “Adversarial attacks and defenses on AI in medical imaging informatics: A survey,” Expert Syst. Appl., vol. 198, 2022, doi: 10.1016/j.eswa.2022.116815.

J. Li, Y. Liu, T. Chen, Z. Xiao, Z. Li, and J. Wang, “Adversarial attacks and defenses on cyber-physical systems: A survey,” IEEE Internet Things J., vol. 7, no. 6, pp. 5103–5115, 2020, doi: 10.1109/JIOT.2020.2975654.

Z. Zhai, P. Li, and S. Feng, “State of the art on adversarial attacks and defenses in graphs,” Neural Comput. Appl., vol. 35, no. 26, pp. 18851–18872, 2023, doi: 10.1007/s00521-023-08839-9.

S. A. Alsuhibany, “A Survey on Adversarial Perturbations and Attacks on CAPTCHAs,” Appl. Sci., vol. 13, no. 7, 2023, doi: 10.3390/app13074602.

A. S. Albahri et al., “A systematic review of trustworthy and explainable artificial intelligence in healthcare: Assessment of quality, bias risk, and data fusion,” Inf. Fusion, vol. 96, pp. 156–191, 2023, doi: 10.1016/j.inffus.2023.03.008.

A. S. Albahri et al., “A Trustworthy and Explainable Framework for Benchmarking Hybrid Deep Learning Models Based on Chest X-Ray Analysis in CAD Systems,” Int. J. Inf. Technol. Decis. Mak., vol. 0, no. 0, pp. 1–54, 2024, doi: 10.1142/S0219622024500019.

L. Alzubaidi et al., “Towards Risk-Free Trustworthy Artificial Intelligence: Significance and Requirements,” Int. J. Intell. Syst., vol. 2023, p. 4459198, 2023, doi: 10.1155/2023/4459198.

L. Alzubaidi et al., “Comprehensive review of deep learning in orthopaedics: Applications, challenges, trustworthiness, and fusion,” Artif. Intell. Med., vol. 155, p. 102935, 2024, doi: https://doi.org/10.1016/j.artmed.2024.102935.

M. A. Alsalem et al., “Evaluation of trustworthy artificial intelligent healthcare applications using multi-criteria decision-making approach,” Expert Syst. Appl., vol. 246, p. 123066, 2024, doi: 10.1016/j.eswa.2023.123066.

L. Alzubaidi et al., “Trustworthy deep learning framework for the detection of abnormalities in X-ray shoulder images,” PLoS One, vol. 19, no. 3 March, p. e0299545, 2024, doi: 10.1371/journal.pone.0299545.

A. S. Albahri et al., “A Systematic Review of Using Deep Learning Technology in the Steady-State Visually Evoked Potential-Based Brain-Computer Interface Applications: Current Trends and Future Trust Methodology,” Int. J. Telemed. Appl., vol. 2023, 2023, doi: 10.1155/2023/7741735.

M. G. Yaseen and A. S. Albahri, “Mapping the Evolution of Intrusion Detection in Big Data: A Bibliometric Analysis,” Mesopotamian J. Big Data, vol. 2023, pp. 138–148, 2023, doi: 10.58496/mjbd/2023/018.

S. Rani, A. Kataria, S. Kumar, and P. Tiwari, “Federated learning for secure IoMT-applications in smart healthcare systems: A comprehensive review,” Knowledge-Based Syst., vol. 274, p. 110658, 2023, doi: 10.1016/j.knosys.2023.110658.

M. R. Baker et al., “Comparison of Machine Learning Approaches for Detecting COVID-19-Lockdown-Related Discussions During Recovery and Lockdown Periods,” J. Oper. Intell., vol. 1, no. 1, pp. 11–29, 2023, doi: 10.31181/jopi1120233.