Categories: FAANG

Careful With That Scalpel: Improving Gradient Surgery With an EMA

Beyond minimizing a single training loss, many deep learning estimation pipelines rely on an auxiliary objective to quantify and encourage desirable properties of the model (e.g. performance on another dataset, robustness, agreement with a prior). Although the simplest approach to incorporating an auxiliary loss is to sum it with the training loss as a regularizer, recent works have shown that one can improve performance by blending the gradients beyond a simple sum; this is known as gradient surgery. We cast the problem as a constrained minimization problem where the auxiliary objective is…
AI Generated Robotic Content

Recent Posts

LLM Evaluation Metrics Made Easy

Metrics are a cornerstone element in evaluating any AI system, and in the case of…

1 hour ago

Part 2: A Survey of Analytics Engineering Work at Netflix

This article is the second in a multi-part series sharing a breadth of Analytics Engineering…

1 hour ago

The Death of Net Neutrality Is a Bad Omen

While Americans might not mourn the loss of net neutrality, an appeals court’s ruling sets…

2 hours ago

7 Machine Learning Trends to Watch in 2025

Machine learning is now the cornerstone of recent technological progress, which is especially true for…

1 day ago

Inside the AI agent revolution: How data-driven automation transformed the enterprise in 2024

In a recent survey, 82% of tech executives said they intend to integrate AI agents…

1 day ago

The Cumulus Machine Review: Fast and Frothy Cold Brew

An innovative new coldbrew machine makes for wonderfully frothy nitro and espresso martinis. But is…

1 day ago