Skip to yearly menu bar Skip to main content


Search All 2023 Events
 

51 Results

<<   <   Page 1 of 5   >   >>
Workshop
Don't trust your eyes: on the (un)reliability of feature visualizations
Workshop
Don't trust your eyes: on the (un)reliability of feature visualizations
Robert Geirhos · Roland S. Zimmermann · Blair Bilodeau · Wieland Brendel · Been Kim
Poster
Tue 17:00 Multi-View Masked World Models for Visual Robotic Manipulation
Younggyo Seo · Junsu Kim · Stephen James · Kimin Lee · Jinwoo Shin · Pieter Abbeel
Workshop
Model-tuning Via Prompts Makes NLP Models Adversarially Robust
Workshop
A physics-orientd method for attacking SAR images using salient regions
Workshop
Robust Semantic Segmentation: Strong Adversarial Attacks and Fast Training of Robust Models
Workshop
Transferable Adversarial Perturbations between Self-Supervised Speech Recognition Models
Workshop
Establishing a Benchmark for Adversarial Robustness of Compressed Deep Learning Models after Pruning
Workshop
PIAT: Parameter Interpolation based Adversarial Training for Image Classification
Workshop
Accurate, Explainable, and Private Models: Providing Recourse While Minimizing Training Data Leakage
Workshop
Scoring Black-Box Models for Adversarial Robustness
Workshop
Adversarial Training in Continuous-Time Models and Irregularly Sampled Time-Series: A First Look