Skip to yearly menu bar Skip to main content


Search All 2023 Events
 

48 Results

<<   <   Page 4 of 4   >>   >
Workshop
Benchmarking Adversarial Robustness of Compressed Deep Learning Models
Brijesh Vora · Kartik Patwari · Syed Mahbub Hafiz · Zubair Shafiq · Chen-Nee Chuah
Workshop
On feasibility of intent obfuscating attacks
ZhaoBin Li · Patrick Shafto
Workshop
Robust Semantic Segmentation: Strong Adversarial Attacks and Fast Training of Robust Models
Francesco Croce · Naman Singh · Matthias Hein
Workshop
Identifying Adversarially Attackable and Robust Samples
Vyas Raina · Mark Gales
Workshop
Near Optimal Adversarial Attack on UCB Bandits
Shiliang Zuo
Workshop
Why do universal adversarial attacks work on large language models?: Geometry might be the answer
Workshop
Black Box Adversarial Prompting for Foundation Models
Workshop
Near Optimal Adversarial Attack on UCB Bandits
Workshop
Black Box Adversarial Prompting for Foundation Models
Natalie Maus · Patrick Chao · Eric Wong · Jacob Gardner
Workshop
Why do universal adversarial attacks work on large language models?: Geometry might be the answer
Varshini Subhash · Anna Bialas · Siddharth Swaroop · Weiwei Pan · Finale Doshi-Velez
Workshop
AdversNLP: A Practical Guide to Assessing NLP Robustness Against Text Adversarial Attacks
Othmane BELMOUKADAM
Workshop
AdversNLP: A Practical Guide to Assessing NLP Robustness Against Text Adversarial Attacks