29 results found Sort:
- Filter by Primary Language:
- Python (17)
- Jupyter Notebook (2)
- HTML (1)
- Java (1)
- JavaScript (1)
- CSS (1)
- Svelte (1)
- Go (1)
- +
This repository is primarily maintained by Omar Santos (@santosomar) and includes thousands of resources related to ethical hacking, bug bounties, digital forensics and incident response (DFIR), artif...
Created
2017-06-19
3,807 commits to master branch, last one a day ago
🐢 Open-Source Evaluation & Testing for AI & LLM systems
Created
2022-03-06
10,289 commits to main branch, last one 7 hours ago
A curated list of useful resources that cover Offensive AI.
Created
2023-01-28
140 commits to main branch, last one 9 days ago
A list of backdoor learning resources
Created
2020-06-13
734 commits to master branch, last one 8 months ago
a prompt injection scanner for custom LLM applications
Created
2023-07-15
28 commits to main branch, last one about a month ago
A security scanner for your LLM agentic workflows
Created
2025-02-12
59 commits to main branch, last one 4 days ago
RuLES: a benchmark for evaluating rule-following in language models
Created
2023-11-03
38 commits to main branch, last one 2 months ago
Toolkits to create a human-in-the-loop approval layer to monitor and guide AI agents workflow in real-time.
Created
2024-10-13
146 commits to main branch, last one 4 months ago
A curated list of academic events on AI Security & Privacy
Created
2021-10-04
70 commits to main branch, last one 7 months ago
[CCS'24] SafeGen: Mitigating Unsafe Content Generation in Text-to-Image Models
Created
2023-12-04
20 commits to main branch, last one 14 days ago
Whistleblower is a offensive security tool for testing against system prompt leakage and capability discovery of an AI application exposed through API. Built for AI engineers, security researchers and...
Created
2024-06-23
28 commits to main branch, last one 8 months ago
The official implementation of the CCS'23 paper, Narcissus clean-label backdoor attack -- only takes THREE images to poison a face recognition dataset in a clean-label way and achieves a 99.89% attack...
Created
2022-04-08
59 commits to main branch, last one about a year ago
Framework for testing vulnerabilities of large language models (LLM).
Created
2024-09-05
10 commits to release branch, last one 3 days ago
Cyber-Security Bible! Theory and Tools, Kali Linux, Penetration Testing, Bug Bounty, CTFs, Malware Analysis, Cryptography, Secure Programming, Web App Security, Cloud Security, Devsecops, Ethical Hack...
Created
2024-01-11
33 commits to main branch, last one 5 months ago
Code for "Adversarial attack by dropping information." (ICCV 2021)
Created
2021-04-12
29 commits to main branch, last one 3 years ago
Run and manage MCP servers easily and securely
Created
2025-03-12
354 commits to main branch, last one 6 hours ago
Performing website vulnerability scanning using OpenAI technologie
Created
2023-02-26
43 commits to main branch, last one 10 days ago
ATLAS tactics, techniques, and case studies data
Created
2021-12-30
277 commits to main branch, last one about a month ago
this.env defines, locks, and hashes the environment to establish a reliable and secure operational context. By detecting and responding to changes, it ensures consistency and integrity, especially for...
Created
2024-08-02
20 commits to main branch, last one 2 months ago
Unofficial pytorch implementation of paper: Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures
Created
2021-07-16
19 commits to main branch, last one 15 days ago
Code scanner to check for issues in prompts and LLM calls
Created
2025-03-14
51 commits to master branch, last one 9 days ago
[NDSS'24] Inaudible Adversarial Perturbation: Manipulating the Recognition of User Speech in Real Time
Created
2023-12-08
7 commits to master branch, last one 6 months ago
Official Implementation of ICLR 2022 paper, ``Adversarial Unlearning of Backdoors via Implicit Hypergradient''
Created
2021-10-07
17 commits to main branch, last one 2 years ago
AIHTTPAnalyzer revolutionizes web application security testing by bringing artificial intelligence capabilities to Burp Suite. This innovative extension harnesses the power of AI to automate vulnerabi...
Created
2024-08-27
84 commits to main branch, last one about a month ago
This repository provide the studies on the security of language models for code (CodeLMs).
Created
2023-11-04
165 commits to main branch, last one about a month ago
[IJCAI 2024] Imperio is an LLM-powered backdoor attack. It allows the adversary to issue language-guided instructions to control the victim model's prediction for arbitrary targets.
Created
2024-01-02
9 commits to main branch, last one 12 months ago
安全手册,企业安全实践、攻防与安全研究知识库
Created
2023-11-22
39 commits to main branch, last one 4 months ago
Benchmark evaluating LLMs on their ability to create and resist disinformation. Includes comprehensive testing across major models (Claude, GPT-4, Gemini, Llama, etc.) with standardized evaluation met...
Created
2024-10-22
12 commits to master branch, last one 26 days ago
Learning to Identify Critical States for Reinforcement Learning from Videos (Accepted to ICCV'23)
Created
2023-07-18
6 commits to main branch, last one about a year ago