21 results found Sort:

This repository is primarily maintained by Omar Santos (@santosomar) and includes thousands of resources related to ethical hacking, bug bounties, digital forensics and incident response (DFIR), artif...
Created 2017-06-19
3,705 commits to master branch, last one 17 hours ago
117
1.1k
cc-by-sa-4.0
27
A curated list of useful resources that cover Offensive AI.
Created 2023-01-28
134 commits to main branch, last one 18 days ago
A list of backdoor learning resources
Created 2020-06-13
734 commits to master branch, last one 4 months ago
15
211
apache-2.0
2
RuLES: a benchmark for evaluating rule-following in language models
Created 2023-11-03
31 commits to main branch, last one about a month ago
A curated list of academic events on AI Security & Privacy
Created 2021-10-04
70 commits to main branch, last one 3 months ago
Toolkits to create a human-in-the-loop approval layer to monitor and guide AI agents workflow in real-time.
Created 2024-10-13
137 commits to main branch, last one a day ago
18
106
apache-2.0
12
[CCS'24] SafeGen: Mitigating Unsafe Content Generation in Text-to-Image Models
Created 2023-12-04
17 commits to main branch, last one about a month ago
The official implementation of the CCS'23 paper, Narcissus clean-label backdoor attack -- only takes THREE images to poison a face recognition dataset in a clean-label way and achieves a 99.89% attack...
Created 2022-04-08
59 commits to main branch, last one about a year ago
Code for "Adversarial attack by dropping information." (ICCV 2021)
Created 2021-04-12
29 commits to main branch, last one 3 years ago
Performing website vulnerability scanning using OpenAI technologie
Created 2023-02-26
42 commits to main branch, last one 7 months ago
Official Implementation of ICLR 2022 paper, ``Adversarial Unlearning of Backdoors via Implicit Hypergradient''
Created 2021-10-07
17 commits to main branch, last one 2 years ago
ATLAS tactics, techniques, and case studies data
Created 2021-12-30
276 commits to main branch, last one about a month ago
Cyber-Security Bible! Theory and Tools, Kali Linux, Penetration Testing, Bug Bounty, CTFs, Malware Analysis, Cryptography, Secure Programming, Web App Security, Cloud Security, Devsecops, Ethical Hack...
Created 2024-01-11
33 commits to main branch, last one about a month ago
[IJCAI 2024] Imperio is an LLM-powered backdoor attack. It allows the adversary to issue language-guided instructions to control the victim model's prediction for arbitrary targets.
Created 2024-01-02
9 commits to main branch, last one 7 months ago
[NDSS'24] Inaudible Adversarial Perturbation: Manipulating the Recognition of User Speech in Real Time
Created 2023-12-08
7 commits to master branch, last one about a month ago
This repository provide the studies on the security of language models for code (CodeLMs).
Created 2023-11-04
160 commits to main branch, last one 26 days ago
Website Prompt Injection is a concept that allows for the injection of prompts into an AI system via a website's. This technique exploits the interaction between users, websites, and AI systems to exe...
Created 2024-03-11
9 commits to main branch, last one 8 months ago
The Prompt Injection Testing Tool is a Python script designed to assess the security of your AI system's prompt handling against a predefined list of user prompts commonly used for injection attacks. ...
Created 2024-03-20
4 commits to main branch, last one 8 months ago
Learning to Identify Critical States for Reinforcement Learning from Videos (Accepted to ICCV'23)
Created 2023-07-18
6 commits to main branch, last one about a year ago
安全手册,企业安全实践、攻防与安全研究知识库
Created 2023-11-22
38 commits to main branch, last one 13 days ago