Thanks to visit codestin.com
Credit goes to www.scribd.com

0% found this document useful (0 votes)
7 views5 pages

Batch 8C Review1

The project report focuses on enhancing the robustness of AI systems against adversarial attacks, which manipulate input data to mislead AI models. It outlines the development of advanced defensive strategies using Python, including adversarial training and robust optimization, to improve the security and reliability of AI applications. The expected outcomes include a deployable framework for real-world applications and comprehensive evaluation metrics for adversarial robustness.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
7 views5 pages

Batch 8C Review1

The project report focuses on enhancing the robustness of AI systems against adversarial attacks, which manipulate input data to mislead AI models. It outlines the development of advanced defensive strategies using Python, including adversarial training and robust optimization, to improve the security and reliability of AI applications. The expected outcomes include a deployable framework for real-world applications and comprehensive evaluation metrics for adversarial robustness.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 5

ENHANCING ROBUSTNESS OF AI SYSTEMS

AGAINST ADVERSARIAL ATTACKS


A PROJECT REPORT
Submitted in partial fulfillment of the requirements for the award of the degree of

Bachelor of Technology

in

COMPUTER SCIENCE AND ENGINEERING

BY

8C

M.Sai Akhila Dunna Pranav


21331A05F4 21331A05J2

Yerrapati Sravya Pathivada Gowtham 21331A05J1 21331A05D9

Under the Supervision of


Mrs. N Ashwini
Assistant Professor

DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING


MAHARAJ VIJAYARAM GAJAPATHI RAJ COLLEGE OF ENGINEERING (Autonomous)
(Approved by AICTE, New Delhi, and permanently affiliated to JNTUGV, Vizianagaram), Listed u/s 2(f) & 12(B)
of UGC Act 1956.
Vijayaram Nagar Campus, Chintalavalasa,Vizianagaram-535005, Andhra Pradesh
APRIL, 2024
ABSTRACT
Artificial Intelligence and Machine Learning have revolutionized our lives, from autonomous
vehicles to language translations, facial recognition and fraud detection. However, these
pervasive technologies are not invulnerable to threats. One significant challenge that AI/ML
models today face is Adversarial attacks - a deliberate method of manipulating the input data
(data poisoning) to mislead the system. Adversarial attacks pose significant threats to the
reliability and security of modern artificial intelligence (AI) applications and networks. This
project focuses on developing and implementing advanced defensive strategies to protect AI
models from these malicious inputs. Utilizing Python as the sole programming language, we will
explore techniques such as robust optimization, adversarial training, and innovative defense
mechanisms. The project aims to enhance the resilience of AI systems against various forms of
adversarial manipulations, ensuring their secure deployment in real-world applications. The
expected outcomes include improved security and reliability of AI models, comprehensive
evaluation metrics for adversarial robustness, and a deploy-able framework for real-world AI
applications.

INTRODUCTION
Artificial Intelligence (AI) has revolutionized numerous fields, including image recognition,
natural language processing, autonomous driving, and cybersecurity. Despite these
advancements, AI models, particularly those utilizing deep learning techniques, are susceptible
to adversarial attacks. These attacks involve the introduction of malicious inputs that
intentionally mislead AI models, leading to incorrect predictions. Such vulnerabilities pose
significant threats to the reliability and security of AI systems, emphasizing the need for robust
defensive strategies to mitigate these risks.
LITERATURE SURVEY
Adversarial Attacks in AI : Adversarial attacks are attempts to trick AI systems by altering input
data, leading to incorrect predictions which cause AI models, especially deep learning ones, to
fail. For example, a slight adjustment to an image can make an AI system misidentify what’s in
the picture.
Types of Adversarial Attacks : There are two main types of attacks on AI models: evasion
attacks and poisoning attacks. Evasion attacks happen when an attacker changes the input data
during the model’s use to trick it into making the wrong decision, such as confusing an image
classifier by changing a few pixels (Szegedy et al., 2014). Poisoning attacks occur during the
training phase of the model, where the attacker adds harmful data to the training set, causing the
AI to learn incorrectly and perform poorly in the future (Biggio et al., 2012).
Impact on AI Applications : These attacks can affect many real-world AI applications. For
example, in facial recognition systems, adversarial attacks can lead to security risks by making the
system misidentify people
Existing Defensive Mechanisms : Several strategies have been developed to defend against
these attacks. One common method is adversarial training, where the AI model is trained using
both normal and adversarial examples so it can better recognize and resist attacks (Madry et al.,
2018). Another approach is robust optimization, which involves tweaking the model to make it
less sensitive to adversarial inputs (Wong & Kolter, 2018). Other techniques include reducing the
complexity of input data to make attacks less effective (feature squeezing) and modifying inputs
before the model processes them (input transformations) (Xu et al., 2017).
Evaluation of Defense Strategies : It is essential to evaluate how well these defense mechanisms
work. Researchers use different metrics to assess how resilient models are to attacks, including
how accurately the model can classify adversarial examples and how well it performs under
different types of attacks.
PROBLEM STATEMENT
Artificial Intelligence (AI) and Machine Learning (ML) systems have become integral in various
high-stakes applications, such as autonomous driving, facial recognition, and cybersecurity.
However, these systems are increasingly vulnerable to adversarial attacks, where small, often
imperceptible modifications to input data lead to incorrect or potentially dangerous outputs. This
weakness poses a significant threat to the reliability, security, and trustworthiness of AI models,
especially in critical real-world scenarios. Current defense strategies, including adversarial
training and robust optimization, either fall short in providing comprehensive protection or come
with trade-offs in performance and scalability. Therefore, there is a critical need for innovative,
efficient, and resilient defense mechanisms to protect AI models from adversarial manipulations,
ensuring their safe and secure deployment in diverse applications.
This project aims to address this challenge by developing advanced defensive strategies that
enhance the robustness of AI models against adversarial attacks.

REQUIREMENTS GATHERING
1. Software Requirements:
Programming Language : Python - The primary programming language for developing the
project.
Machine Learning Frameworks : For building and training deep learning models.
Adversarial Attack Libraries : A library for benchmarking machine learning systems against
adversarial attacks.
Web Frameworks : Flask - A web framework for deploying the defense framework as a web
application.
Visualization Tools : Matplotlib: For plotting graphs and visualizing data. Version Control :
Git: For version control to manage changes to the codebase and collaborate with others.
Operating System : Compatibility with Windows, Linux, or macOS.

2. Hardware Requirements:
CPU : A multi-core processor is recommended for training and testing
models. GPU : A dedicated GPU for faster training of deep learning models
RAM : A minimum of 8 GB of RAM is required
Storage : At least 256 GB of free storage space for datasets, models, and project files.
Networking : Stable internet connection for downloading libraries, datasets, and deploying the
web application.
PROPOSED WORK
1. Adversarial Attack Analysis : Conduct an in-depth study of various adversarial attack
methods
2. Adversarial Training : Implement adversarial training where the model is trained using
adversarial examples alongside normal data.
3. Robust Optimization Techniques : Apply robust optimization techniques to strengthen model
defenses against worst-case adversarial attacks.
4. Develop New Defense Strategies : Create new ways to defend AI, such as Simplifying input
data to make it harder to trick the model.
5. Certified Defenses and Evaluation : Implement defenses that can prove their effectiveness
against attacks.
6. Build a Defense Framework : Create a tool that integrates these defense strategies and can be
used to protect AI models in real-world applications.

7. Testing and Validation : Test these strategies on popular

REFERENCES
Goodfellow, I. J., Shlens, J., & Szegedy, C. (2014). Explaining and Harnessing Adversarial
Examples. https://arxiv.org/abs/1412.6572

Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I., & Fergus, R.
(2014). Intriguing properties of neural networks
https://arxiv.org/abs/1312.6199
Role Faculty Name Designation Signature

PROJECT Mrs. N Aswini Assistant


SUPERVIS Professor
OR

PANEL Mrs K. Sobha Rani Professor


MEMBERS

Mr K. Leela Prasad Assistant


Professor

Mrs Jyothi Pyla Assistant


Professor

Project Coordinator Head of the Department Computer Science and Engineering

You might also like