14 results found Sort:

This repository is primarily maintained by Omar Santos (@santosomar) and includes thousands of resources related to ethical hacking, bug bounties, digital forensics and incident response (DFIR), artif...
Created 2017-06-19
3,601 commits to master branch, last one 15 days ago
215
3.4k
apache-2.0
27
🐢 Open-Source Evaluation & Testing for LLMs and ML models
Created 2022-03-06
9,675 commits to main branch, last one a day ago
103
1.0k
cc-by-sa-4.0
25
A curated list of useful resources that cover Offensive AI.
Created 2023-01-28
121 commits to main branch, last one 19 days ago
A list of backdoor learning resources
Created 2020-06-13
728 commits to master branch, last one 7 months ago
14
197
apache-2.0
3
RuLES: a benchmark for evaluating rule-following in language models
Created 2023-11-03
22 commits to main branch, last one about a month ago
A curated list of academic events on AI Security & Privacy
Created 2021-10-04
69 commits to main branch, last one 23 days ago
The official implementation of the CCS'23 paper, Narcissus clean-label backdoor attack -- only takes THREE images to poison a face recognition dataset in a clean-label way and achieves a 99.89% attack...
Created 2022-04-08
59 commits to main branch, last one about a year ago
Code for "Adversarial attack by dropping information." (ICCV 2021)
Created 2021-04-12
29 commits to main branch, last one 2 years ago
Performing website vulnerability scanning using OpenAI technologie
Created 2023-02-26
42 commits to main branch, last one about a month ago
Official Implementation of ICLR 2022 paper, ``Adversarial Unlearning of Backdoors via Implicit Hypergradient''
Created 2021-10-07
17 commits to main branch, last one about a year ago
[IJCAI 2024] Imperio is an LLM-powered backdoor attack. It allows the adversary to issue language-guided instructions to control the victim model's prediction for arbitrary targets.
Created 2024-01-02
9 commits to main branch, last one about a month ago
ATLAS tactics, techniques, and case studies data
Created 2021-12-30
274 commits to main branch, last one about a month ago
Website Prompt Injection is a concept that allows for the injection of prompts into an AI system via a website's. This technique exploits the interaction between users, websites, and AI systems to exe...
Created 2024-03-11
9 commits to main branch, last one 2 months ago
Learning to Identify Critical States for Reinforcement Learning from Videos (Accepted to ICCV'23)
Created 2023-07-18
6 commits to main branch, last one 9 months ago