Categories: FAANG

The “Super Weight:” How Even a Single Parameter can Determine a Large Language Model’s Behavior

A recent paper from Apple researchers, “The Super Weight in Large Language Models,” reveals that an extremely small subset of parameters in LLMs (in some cases, a single parameter) can exert a disproportionate influence on an LLM’s overall functionality (see Figure 1). This work highlights the critical role of these “super weights” and their corresponding “super activations,” offering a new insight into LLM architecture and avenues for efficient model compression. The paper provides full technical details and experimental results; in this post, we provide a high-level overview of the key…
AI Generated Robotic Content

Recent Posts

Qwen Image Edit 2511 — Coming next week

submitted by /u/Queasy-Carrot-7314 [link] [comments]

3 hours ago

BERT Models and Its Variants

This article is divided into two parts; they are: • Architecture and Training of BERT…

3 hours ago

Lean4: How the theorem prover works and why it’s the new competitive edge in AI

Large language models (LLMs) have astounded the world with their capabilities, yet they remain plagued…

4 hours ago

13 Best MagSafe Power Banks for iPhones (2025), Tested and Reviewed

Keep your iPhone or Qi2 Android phone topped up with one of these WIRED-tested Qi2…

4 hours ago

I love Qwen

It is far more likely that a woman underwater is wearing at least a bikini…

1 day ago

100% Unemployment is Inevitable*

TL;DR AI is already raising unemployment in knowledge industries, and if AI continues progressing toward…

1 day ago