Categories: FAANG

Interpreting CLIP: Insights on the Robustness to ImageNet Distribution Shifts

What distinguishes robust models from non-robust ones? While for ImageNet distribution shifts it has been shown that such differences in robustness can be traced back predominantly to differences in training data, so far it is not known what that translates to in terms of what the model has learned. In this work, we bridge this gap by probing the representation spaces of 16 robust zero-shot CLIP vision encoders with various backbones (ResNets and ViTs) and pretraining sets (OpenAI, LAION-400M, LAION-2B, YFCC15M, CC12M and DataComp), and comparing them to the representation spaces of less…
AI Generated Robotic Content

Recent Posts

Everyone Has Given Up on AI Safety, Now What?

The End of the AI Safety DebateFor years, a passionate contingent of researchers, ethicists, and…

1 day ago

The rise of browser-use agents: Why Convergence’s Proxy is beating OpenAI’s Operator

A new wave of AI-powered browser-use agents is emerging, promising to transform how enterprises interact…

1 day ago

Elon Musk Threatens FBI Agents and Air Traffic Controllers With Forced Resignation If They Don’t Respond to an Email

Employees throughout the federal government have until 11:59pm ET Monday to detail five things they…

1 day ago

How to get a robot collective to act like a smart material

Researchers are blurring the lines between robotics and materials, with a proof-of-concept material-like collective of…

1 day ago

Understanding RAG Part VI: Effective Retrieval Optimization

Be sure to check out the previous articles in this series: •

2 days ago

PR Agencies in the Age of AI

TL;DR We compared Grok 3 and o3-mini’s results on this topic. They both passed. Since…

2 days ago