104 results for “topic:adversarial-defense”
Must-read Papers on Textual Adversarial Attack and Defense
auto_LiRPA: An Automatic Linear Relaxation based Perturbation Analysis Library for Neural Networks and General Computational Graphs
Code for our NeurIPS 2019 *spotlight* "Provably Robust Deep Learning via Adversarially Trained Smoothed Classifiers"
A curated list of papers on adversarial machine learning (adversarial examples and defense methods).
[ICML 2024] Unsupervised Adversarial Fine-Tuning of Vision Embeddings for Robust Large Vision-Language Models
This repository contains the implementation of three adversarial example attack methods FGSM, IFGSM, MI-FGSM and one Distillation as defense against all attacks using MNIST dataset.
A list of awesome resources for adversarial attack and defense method in deep learning
Provably defending pretrained classifiers including the Azure, Google, AWS, and Clarifai APIs
Certified defense to adversarial examples using CROWN and IBP. Also includes GPU implementation of CROWN verification algorithm (in PyTorch).
Adversarial attacks on Deep Reinforcement Learning (RL)
[ICLR 2021] "InfoBERT: Improving Robustness of Language Models from An Information Theoretic Perspective" by Boxin Wang, Shuohang Wang, Yu Cheng, Zhe Gan, Ruoxi Jia, Bo Li, Jingjing Liu
CVPR 2022 Workshop Robust Classification
Machine Learning Attack Series
Adversarial Distributional Training (NeurIPS 2020)
😎 A curated list of awesome real-world adversarial examples resources
This repository provide the studies on the security of language models for code (CodeLMs).
Code for the paper: Adversarial Training Against Location-Optimized Adversarial Patches. ECCV-W 2020.
pytorch implementation of Parametric Noise Injection for adversarial defense
Learnable Boundary Guided Adversarial Training (ICCV2021)
GARNET: Reduced-Rank Topology Learning for Robust and Scalable Graph Neural Networks
[IEEE TIP 2021] Self-Attention Context Network: Addressing the Threat of Adversarial Attacks for Hyperspectral Image Classification
Code for the paper "Consistency Regularization for Certified Robustness of Smoothed Classifiers" (NeurIPS 2020)
Feature Separation and Recalibration (CVPR 2023 Highlights)
Source Code for 'SECurity evaluation platform FOR Speaker Recognition' released in 'Defending against Audio Adversarial Examples on Speaker Recognition Systems'
Understanding Catastrophic Overfitting in Single-step Adversarial Training [AAAI 2021]
Adversarial Ranking Attack and Defense, ECCV, 2020.
Adversarial Attack and Defense in Deep Ranking, T-PAMI, 2024
[ECCV 2020] Pytorch codes for Open-set Adversarial Defense
Code for the paper "SmoothMix: Training Confidence-calibrated Smoothed Classifiers for Certified Robustness" (NeurIPS 2021)
Minimal implementation of Denoised Smoothing (https://arxiv.org/abs/2003.01908) in TensorFlow.