20 results found Sort:

This repository is primarily maintained by Omar Santos (@santosomar) and includes thousands of resources related to ethical hacking, bug bounties, digital forensics and incident response (DFIR), artif...
Created 2017-06-19
3,746 commits to master branch, last one 4 days ago
118
1.1k
cc-by-sa-4.0
27
A curated list of useful resources that cover Offensive AI.
Created 2023-01-28
135 commits to main branch, last one 14 days ago
A list of backdoor learning resources
Created 2020-06-13
734 commits to master branch, last one 5 months ago
15
214
apache-2.0
2
RuLES: a benchmark for evaluating rule-following in language models
Created 2023-11-03
32 commits to main branch, last one 27 days ago
Toolkits to create a human-in-the-loop approval layer to monitor and guide AI agents workflow in real-time.
Created 2024-10-13
146 commits to main branch, last one 25 days ago
A curated list of academic events on AI Security & Privacy
Created 2021-10-04
70 commits to main branch, last one 4 months ago
19
108
apache-2.0
12
[CCS'24] SafeGen: Mitigating Unsafe Content Generation in Text-to-Image Models
Created 2023-12-04
17 commits to main branch, last one 2 months ago
The official implementation of the CCS'23 paper, Narcissus clean-label backdoor attack -- only takes THREE images to poison a face recognition dataset in a clean-label way and achieves a 99.89% attack...
Created 2022-04-08
59 commits to main branch, last one about a year ago
Code for "Adversarial attack by dropping information." (ICCV 2021)
Created 2021-04-12
29 commits to main branch, last one 3 years ago
Performing website vulnerability scanning using OpenAI technologie
Created 2023-02-26
42 commits to main branch, last one 8 months ago
Cyber-Security Bible! Theory and Tools, Kali Linux, Penetration Testing, Bug Bounty, CTFs, Malware Analysis, Cryptography, Secure Programming, Web App Security, Cloud Security, Devsecops, Ethical Hack...
Created 2024-01-11
33 commits to main branch, last one 2 months ago
ATLAS tactics, techniques, and case studies data
Created 2021-12-30
276 commits to main branch, last one 2 months ago
Official Implementation of ICLR 2022 paper, ``Adversarial Unlearning of Backdoors via Implicit Hypergradient''
Created 2021-10-07
17 commits to main branch, last one 2 years ago
[IJCAI 2024] Imperio is an LLM-powered backdoor attack. It allows the adversary to issue language-guided instructions to control the victim model's prediction for arbitrary targets.
Created 2024-01-02
9 commits to main branch, last one 8 months ago
This repository provide the studies on the security of language models for code (CodeLMs).
Created 2023-11-04
162 commits to main branch, last one 8 days ago
Framework for testing vulnerabilities of large language models (LLM).
Created 2024-09-05
4 commits to release branch, last one 5 days ago
[NDSS'24] Inaudible Adversarial Perturbation: Manipulating the Recognition of User Speech in Real Time
Created 2023-12-08
7 commits to master branch, last one 2 months ago
安全手册,企业安全实践、攻防与安全研究知识库
Created 2023-11-22
39 commits to main branch, last one 23 days ago
Learning to Identify Critical States for Reinforcement Learning from Videos (Accepted to ICCV'23)
Created 2023-07-18
6 commits to main branch, last one about a year ago