Trending repositories for topic adversarial-machine-learning
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
TextAttack 🐙 is a Python framework for adversarial attacks, data augmentation, and model training in NLP https://textattack.readthedocs.io/en/master/
Fawkes, privacy preserving tool against facial recognition systems. More info at https://sandlab.cs.uchicago.edu/fawkes
RobustBench: a standardized adversarial robustness benchmark [NeurIPS'21 Benchmarks and Datasets Track]
A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).
auto_LiRPA: An Automatic Linear Relaxation based Perturbation Analysis Library for Neural Networks and General Computational Graphs
A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).
auto_LiRPA: An Automatic Linear Relaxation based Perturbation Analysis Library for Neural Networks and General Computational Graphs
RobustBench: a standardized adversarial robustness benchmark [NeurIPS'21 Benchmarks and Datasets Track]
TextAttack 🐙 is a Python framework for adversarial attacks, data augmentation, and model training in NLP https://textattack.readthedocs.io/en/master/
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
Fawkes, privacy preserving tool against facial recognition systems. More info at https://sandlab.cs.uchicago.edu/fawkes
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
TextAttack 🐙 is a Python framework for adversarial attacks, data augmentation, and model training in NLP https://textattack.readthedocs.io/en/master/
A curated list of useful resources that cover Offensive AI.
ChatGPT Jailbreaks, GPT Assistants Prompt Leaks, GPTs Prompt Injection, LLM Prompt Security, Super Prompts, Prompt Hack, Prompt Security, Ai Prompt Engineering, Adversarial Machine Learning.
Fawkes, privacy preserving tool against facial recognition systems. More info at https://sandlab.cs.uchicago.edu/fawkes
Papers and resources related to the security and privacy of LLMs 🤖
RobustBench: a standardized adversarial robustness benchmark [NeurIPS'21 Benchmarks and Datasets Track]
A curated list of adversarial attacks and defenses papers on graph-structured data.
A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).
TransferAttack is a pytorch framework to boost the adversarial transferability for image classification.
auto_LiRPA: An Automatic Linear Relaxation based Perturbation Analysis Library for Neural Networks and General Computational Graphs
Security and Privacy Risk Simulator for Machine Learning (arXiv:2312.17667)
ChatGPT Jailbreaks, GPT Assistants Prompt Leaks, GPTs Prompt Injection, LLM Prompt Security, Super Prompts, Prompt Hack, Prompt Security, Ai Prompt Engineering, Adversarial Machine Learning.
A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).
Papers and resources related to the security and privacy of LLMs 🤖
TransferAttack is a pytorch framework to boost the adversarial transferability for image classification.
A curated list of useful resources that cover Offensive AI.
auto_LiRPA: An Automatic Linear Relaxation based Perturbation Analysis Library for Neural Networks and General Computational Graphs
RobustBench: a standardized adversarial robustness benchmark [NeurIPS'21 Benchmarks and Datasets Track]
Security and Privacy Risk Simulator for Machine Learning (arXiv:2312.17667)
A curated list of adversarial attacks and defenses papers on graph-structured data.
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
TextAttack 🐙 is a Python framework for adversarial attacks, data augmentation, and model training in NLP https://textattack.readthedocs.io/en/master/
Fawkes, privacy preserving tool against facial recognition systems. More info at https://sandlab.cs.uchicago.edu/fawkes
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
ChatGPT Jailbreaks, GPT Assistants Prompt Leaks, GPTs Prompt Injection, LLM Prompt Security, Super Prompts, Prompt Hack, Prompt Security, Ai Prompt Engineering, Adversarial Machine Learning.
Papers and resources related to the security and privacy of LLMs 🤖
TextAttack 🐙 is a Python framework for adversarial attacks, data augmentation, and model training in NLP https://textattack.readthedocs.io/en/master/
A curated list of useful resources that cover Offensive AI.
Fawkes, privacy preserving tool against facial recognition systems. More info at https://sandlab.cs.uchicago.edu/fawkes
TransferAttack is a pytorch framework to boost the adversarial transferability for image classification.
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
A Python library for adversarial machine learning focusing on benchmarking adversarial robustness.
RobustBench: a standardized adversarial robustness benchmark [NeurIPS'21 Benchmarks and Datasets Track]
Security and Privacy Risk Simulator for Machine Learning (arXiv:2312.17667)
A curated list of academic events on AI Security & Privacy
A curated list of trustworthy deep learning papers. Daily updating...
The fastest && easiest LLM security and privacy guardrails for GenAI apps.
CTF challenges designed and implemented in machine learning applications
A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).
[NeurIPS 2020, Spotlight] Code for "Robust Deep Reinforcement Learning against Adversarial Perturbations on Observations"
A curated list of adversarial attacks and defenses papers on graph-structured data.
ChatGPT Jailbreaks, GPT Assistants Prompt Leaks, GPTs Prompt Injection, LLM Prompt Security, Super Prompts, Prompt Hack, Prompt Security, Ai Prompt Engineering, Adversarial Machine Learning.
Papers and resources related to the security and privacy of LLMs 🤖
The fastest && easiest LLM security and privacy guardrails for GenAI apps.
TransferAttack is a pytorch framework to boost the adversarial transferability for image classification.
A curated list of academic events on AI Security & Privacy
CTF challenges designed and implemented in machine learning applications
The code for ECCV2022 (Watermark Vaccine: Adversarial Attacks to Prevent Watermark Removal)
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).
[NeurIPS 2020, Spotlight] Code for "Robust Deep Reinforcement Learning against Adversarial Perturbations on Observations"
Security and Privacy Risk Simulator for Machine Learning (arXiv:2312.17667)
A curated list of trustworthy deep learning papers. Daily updating...
A Python library for adversarial machine learning focusing on benchmarking adversarial robustness.
RobustBench: a standardized adversarial robustness benchmark [NeurIPS'21 Benchmarks and Datasets Track]
A curated resource list of adversarial attacks and defenses for Windows PE malware detection.
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
Reading list for adversarial perspective and robustness in deep reinforcement learning.
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
ChatGPT Jailbreaks, GPT Assistants Prompt Leaks, GPTs Prompt Injection, LLM Prompt Security, Super Prompts, Prompt Hack, Prompt Security, Ai Prompt Engineering, Adversarial Machine Learning.
TransferAttack is a pytorch framework to boost the adversarial transferability for image classification.
Reading list for adversarial perspective and robustness in deep reinforcement learning.
The fastest && easiest LLM security and privacy guardrails for GenAI apps.
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
TextAttack 🐙 is a Python framework for adversarial attacks, data augmentation, and model training in NLP https://textattack.readthedocs.io/en/master/
Papers and resources related to the security and privacy of LLMs 🤖
A curated list of useful resources that cover Offensive AI.
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
Fawkes, privacy preserving tool against facial recognition systems. More info at https://sandlab.cs.uchicago.edu/fawkes
ChatGPT Jailbreaks, GPT Assistants Prompt Leaks, GPTs Prompt Injection, LLM Prompt Security, Super Prompts, Prompt Hack, Prompt Security, Ai Prompt Engineering, Adversarial Machine Learning.
A curated list of trustworthy deep learning papers. Daily updating...
TransferAttack is a pytorch framework to boost the adversarial transferability for image classification.
RobustBench: a standardized adversarial robustness benchmark [NeurIPS'21 Benchmarks and Datasets Track]
Security and Privacy Risk Simulator for Machine Learning (arXiv:2312.17667)
A curated list of adversarial attacks and defenses papers on graph-structured data.
Reading list for adversarial perspective and robustness in deep reinforcement learning.
The fastest && easiest LLM security and privacy guardrails for GenAI apps.
A curated list of academic events on AI Security & Privacy
💡 Adversarial attacks on explanations and how to defend them
A Python library for adversarial machine learning focusing on benchmarking adversarial robustness.
Backdoors Framework for Deep Learning and Federated Learning. A light-weight tool to conduct your research on backdoors.
CTF challenges designed and implemented in machine learning applications
A curated list of academic events on AI Security & Privacy
A curated list of trustworthy deep learning papers. Daily updating...
[ICLR 2023, Spotlight] Indiscriminate Poisoning Attacks on Unsupervised Contrastive Learning
A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).
A Paperlist of Adversarial Attack on Object Detection
A list of papers in NeurIPS 2022 related to adversarial attack and defense / AI security.
Implements Adversarial Examples for Semantic Segmentation and Object Detection, using PyTorch and Detectron2
Security and Privacy Risk Simulator for Machine Learning (arXiv:2312.17667)
The code for ECCV2022 (Watermark Vaccine: Adversarial Attacks to Prevent Watermark Removal)
Official Implementation of ICLR 2022 paper, ``Adversarial Unlearning of Backdoors via Implicit Hypergradient''
The official implementation of the CCS'23 paper, Narcissus clean-label backdoor attack -- only takes THREE images to poison a face recognition dataset in a clean-label way and achieves a 99.89% attack...
A curated list of useful resources that cover Offensive AI.
💡 Adversarial attacks on explanations and how to defend them
Code for the paper "A Light Recipe to Train Robust Vision Transformers" [SaTML 2023]
RobustBench: a standardized adversarial robustness benchmark [NeurIPS'21 Benchmarks and Datasets Track]
auto_LiRPA: An Automatic Linear Relaxation based Perturbation Analysis Library for Neural Networks and General Computational Graphs
Create adversarial attacks against machine learning Windows malware detectors
Backdoors Framework for Deep Learning and Federated Learning. A light-weight tool to conduct your research on backdoors.