Categories: FAANG

Interpreting CLIP: Insights on the Robustness to ImageNet Distribution Shifts

What distinguishes robust models from non-robust ones? While for ImageNet distribution shifts it has been shown that such differences in robustness can be traced back predominantly to differences in training data, so far it is not known what that translates to in terms of what the model has learned. In this work, we bridge this gap by probing the representation spaces of 16 robust zero-shot CLIP vision encoders with various backbones (ResNets and ViTs) and pretraining sets (OpenAI, LAION-400M, LAION-2B, YFCC15M, CC12M and DataComp), and comparing them to the representation spaces of less…
AI Generated Robotic Content

Recent Posts

AI comes alive: From bartenders to surgical aides to puppies, tomorrow’s robots are on their way

Beyond performing tasks, machines will integrate into our social fabric, requiring us to navigate new…

21 hours ago

TikTok Is Already Back Online

Less than 24 hours after the app went dark, TikTok says President-elect Donald Trump has…

21 hours ago

On the Modeling Capabilities of Large Language Models for Sequential Decision Making

Large pretrained models are showing increasingly better performance in reasoning and planning tasks across different…

2 days ago

Anthropomorphizing AI: Dire consequences of mistaking human-like for human have already emerged

Anthropomorphizing AI creates dangerous blind spots in business decision-making beyond simple operational inefficiencies.Read More

2 days ago

The Roadmap for Mastering Machine Learning in 2025

Machine learning (ML) is now a part of our daily lives, from the voice assistants…

3 days ago

Beyond RAG: How cache-augmented generation reduces latency, complexity for smaller workloads

As LLMs become more capable, many RAG applications can be replaced with cache-augmented generation that…

3 days ago