21 results found Sort:
- Filter by Primary Language:
- Python (12)
- HTML (2)
- Jupyter Notebook (2)
- CSS (1)
- Svelte (1)
- +
This repository is primarily maintained by Omar Santos (@santosomar) and includes thousands of resources related to ethical hacking, bug bounties, digital forensics and incident response (DFIR), artif...
Created
2017-06-19
3,705 commits to master branch, last one 17 hours ago
🐢 Open-Source Evaluation & Testing for ML & LLM systems
Created
2022-03-06
10,112 commits to main branch, last one a day ago
A curated list of useful resources that cover Offensive AI.
Created
2023-01-28
134 commits to main branch, last one 18 days ago
A list of backdoor learning resources
Created
2020-06-13
734 commits to master branch, last one 4 months ago
RuLES: a benchmark for evaluating rule-following in language models
Created
2023-11-03
31 commits to main branch, last one about a month ago
A curated list of academic events on AI Security & Privacy
Created
2021-10-04
70 commits to main branch, last one 3 months ago
Toolkits to create a human-in-the-loop approval layer to monitor and guide AI agents workflow in real-time.
Created
2024-10-13
137 commits to main branch, last one a day ago
[CCS'24] SafeGen: Mitigating Unsafe Content Generation in Text-to-Image Models
Created
2023-12-04
17 commits to main branch, last one about a month ago
The official implementation of the CCS'23 paper, Narcissus clean-label backdoor attack -- only takes THREE images to poison a face recognition dataset in a clean-label way and achieves a 99.89% attack...
Created
2022-04-08
59 commits to main branch, last one about a year ago
Code for "Adversarial attack by dropping information." (ICCV 2021)
Created
2021-04-12
29 commits to main branch, last one 3 years ago
Performing website vulnerability scanning using OpenAI technologie
Created
2023-02-26
42 commits to main branch, last one 7 months ago
Official Implementation of ICLR 2022 paper, ``Adversarial Unlearning of Backdoors via Implicit Hypergradient''
Created
2021-10-07
17 commits to main branch, last one 2 years ago
ATLAS tactics, techniques, and case studies data
Created
2021-12-30
276 commits to main branch, last one about a month ago
Cyber-Security Bible! Theory and Tools, Kali Linux, Penetration Testing, Bug Bounty, CTFs, Malware Analysis, Cryptography, Secure Programming, Web App Security, Cloud Security, Devsecops, Ethical Hack...
Created
2024-01-11
33 commits to main branch, last one about a month ago
[IJCAI 2024] Imperio is an LLM-powered backdoor attack. It allows the adversary to issue language-guided instructions to control the victim model's prediction for arbitrary targets.
Created
2024-01-02
9 commits to main branch, last one 7 months ago
[NDSS'24] Inaudible Adversarial Perturbation: Manipulating the Recognition of User Speech in Real Time
Created
2023-12-08
7 commits to master branch, last one about a month ago
This repository provide the studies on the security of language models for code (CodeLMs).
Created
2023-11-04
160 commits to main branch, last one 26 days ago
Website Prompt Injection is a concept that allows for the injection of prompts into an AI system via a website's. This technique exploits the interaction between users, websites, and AI systems to exe...
Created
2024-03-11
9 commits to main branch, last one 8 months ago
The Prompt Injection Testing Tool is a Python script designed to assess the security of your AI system's prompt handling against a predefined list of user prompts commonly used for injection attacks. ...
Created
2024-03-20
4 commits to main branch, last one 8 months ago
Learning to Identify Critical States for Reinforcement Learning from Videos (Accepted to ICCV'23)
Created
2023-07-18
6 commits to main branch, last one about a year ago
安全手册,企业安全实践、攻防与安全研究知识库
Created
2023-11-22
38 commits to main branch, last one 13 days ago