Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
-
Updated
Mar 3, 2025 - Python
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
TextAttack 🐙 is a Python framework for adversarial attacks, data augmentation, and model training in NLP https://textattack.readthedocs.io/en/master/
A Python toolbox to create adversarial examples that fool neural networks in PyTorch, TensorFlow, and JAX
A unified evaluation framework for large language models
PyTorch implementation of adversarial attacks [torchattacks]
Must-read Papers on Textual Adversarial Attack and Defense
A pytorch adversarial library for attack and defense methods on images and graphs
A collection of anomaly detection methods (iid/point-based, graph and time series) including active learning for anomaly detection/discovery, bayesian rule-mining, description for diversity/explanation/interpretability. Analysis of incorporating label feedback with ensemble and tree-based detectors. Includes adversarial attacks with Graph Convol…
An Open-Source Package for Textual Adversarial Attack.
Code relative to "Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks"
A Harder ImageNet Test Set (CVPR 2021)
A Model for Natural Language Attack on Text Classification and Inference
A Python library for adversarial machine learning focusing on benchmarking adversarial robustness.
Implementation of Papers on Adversarial Examples
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to adversarial prompt attacks. 🏆 Best Paper Awards @ NeurIPS ML Safety Workshop 2022
🔥🔥Defending Against Deepfakes Using Adversarial Attacks on Conditional Image Translation Networks
Implementation of the KDD 2020 paper "Graph Structure Learning for Robust Graph Neural Networks"
TrojanZoo provides a universal pytorch platform to conduct security researches (especially backdoor attacks/defenses) of image classification in deep learning.
Simple pytorch implementation of FGSM and I-FGSM
Add a description, image, and links to the adversarial-attacks topic page so that developers can more easily learn about it.
To associate your repository with the adversarial-attacks topic, visit your repo's landing page and select "manage topics."