Search this site
Embedded Files
Towards AGI
  • Home
  • Schedule
  • Projects
  • Topics&Papers
    • Adversarial Robustness
    • Alignment and Safety
    • CompPsych-FoMo
    • Compression and Fast Inference
    • Continual Learning at Scale
    • Emergence & Phase Transitions in ML
    • Foundation Models
    • Generalization (iid and ood)
    • High Performance Computing
    • Knowledge Fusion
    • Neural Scaling Laws
    • Out-of-Distribution Generalization
    • Scaling Laws in Nature
    • State Space Models
    • Time Series Foundation Models
  • Reading Group
Towards AGI
  • Home
  • Schedule
  • Projects
  • Topics&Papers
    • Adversarial Robustness
    • Alignment and Safety
    • CompPsych-FoMo
    • Compression and Fast Inference
    • Continual Learning at Scale
    • Emergence & Phase Transitions in ML
    • Foundation Models
    • Generalization (iid and ood)
    • High Performance Computing
    • Knowledge Fusion
    • Neural Scaling Laws
    • Out-of-Distribution Generalization
    • Scaling Laws in Nature
    • State Space Models
    • Time Series Foundation Models
  • Reading Group
  • More
    • Home
    • Schedule
    • Projects
    • Topics&Papers
      • Adversarial Robustness
      • Alignment and Safety
      • CompPsych-FoMo
      • Compression and Fast Inference
      • Continual Learning at Scale
      • Emergence & Phase Transitions in ML
      • Foundation Models
      • Generalization (iid and ood)
      • High Performance Computing
      • Knowledge Fusion
      • Neural Scaling Laws
      • Out-of-Distribution Generalization
      • Scaling Laws in Nature
      • State Space Models
      • Time Series Foundation Models
    • Reading Group

Adversarial Robustness at Scale 

Scaling Compute Is Not All You Need for Adversarial Robustness


On the Adversarial Robustness of Multi-Modal Foundation Models ICCV’23


On the Robustness of Large Multimodal Models Against Image Adversarial Attacks ArXiv Dec’23 


On Evaluating Adversarial Robustness of Large Vision-Language Models NeurIPS’23


Jailbreak in pieces: Compositional Adversarial Attacks on Multi-Modal Language Models ICLR’24 submission


Non-targeted Adversarial Attacks on Vision-Language Models via Maximizing Information Entropy ICLR’24 submission


VLATTACK: Multimodal Adversarial Attacks on Vision-Language Tasks via Pre-trained Models NeurIPS’23 


Visual Adversarial Examples Jailbreak Aligned Large Language Models ArXiv July’23


How Robust is Google's Bard to Adversarial Image Attacks? ArXiv Sept’23

Baseline Defenses for Adversarial Attacks Against Aligned Language Models ICLR’24 submission 

Google Sites
Report abuse
Page details
Page updated
Google Sites
Report abuse