hapag eta image 1

How Hapag-Lloyd improved schedule reliability with ML-powered vessel schedule predictions using Amazon SageMaker

This post is cowritten with Thomas Voss and Bernhard Hersberger from Hapag-Lloyd. Hapag-Lloyd is one of the world’s leading shipping companies with more than 308 modern vessels, 11.9 million TEUs (twenty-foot equivalent units) transported per year, and 16,700 motivated employees in more than 400 offices in 139 countries. They connect continents, businesses, and people through …

image 1 croppedmax 1000x1000 1

Gemini CLI extension for PostgreSQL in action: Build a fuzzy search feature in minutes

Adding features to an app can be hard. One minute you’re writing code, the next you’re switching to the PostgreSQL database client to run a query, and then it’s over to the console to check on your instances. For example, let’s say you wanted to add search capabilities. This can mean adding the right extensions …

Compute-Optimal Quantization-Aware Training

Quantization-aware training (QAT) is a leading technique for improving the accuracy of quantized neural networks. Previ- ous work has shown that decomposing training into a full-precision (FP) phase followed by a QAT phase yields superior accuracy compared to QAT alone. However, the optimal allocation of compute between the FP and QAT phases remains unclear. We …

ML 19181 Arch

Modernize fraud prevention: GraphStorm v0.5 for real-time inference

Fraud continues to cause significant financial damage globally, with U.S. consumers alone losing $12.5 billion in 2024—a 25% increase from the previous year according to the Federal Trade Commission. This surge stems not from more frequent attacks, but from fraudsters’ increasing sophistication. As fraudulent activities become more complex and interconnected, conventional machine learning approaches fall short …

2 architecturemax 1000x1000 1

Forecasts and data insights come to BigQuery’s MCP and Agent Development Kit tools

For AI agents to be really useful, they need to be able to securely interact with enterprise data. In July, we introduced a toolset to help AI agents interact with and analyze business data in BigQuery through natural language, and with just a few lines of code. Today, we’re taking the next step, with “Ask …

Enabling Differentially Private Federated Learning for Speech Recognition: Benchmarks, Adaptive Optimizers, and Gradient Clipping

While federated learning (FL) and differential privacy (DP) have been extensively studied, their application to automatic speech recognition (ASR) remains largely unexplored due to the challenges in training large transformer models. Specifically, large models further exacerbate issues in FL as they are particularly susceptible to gradient heterogeneity across layers, unlike the relatively uniform gradient behavior …

1q871tK1C7Y8VSAXVOmu4Ig

100X Faster: How We Supercharged Netflix Maestro’s Workflow Engine

By Jun He, Yingyi Zhang, Ely Spears TL;DR We recently upgraded the Maestro engine to go beyond scalability and improved its performance by 100X! The overall overhead is reduced from seconds to milliseconds. We have updated the Maestro open source project with this improvement! Please visit the Maestro GitHub repository to get started. If you find …

Announcing Claude Sonnet 4.5 on Vertex AI

Today, we’re announcing the general availability of Claude Sonnet 4.5, Anthropic’s most intelligent model and its best-performing model for complex agents, coding, and computer use, on Vertex AI. Claude Sonnet 4.5 is built to work independently for hours, maintaining clarity while orchestrating tools and coordinating multiple agents to solve complex problems. It’s designed to excel …

Scaling Laws for Optimal Data Mixtures

Large foundation models are typically trained on data from multiple domains, with the data mixture—the proportion of each domain used—playing a critical role in model performance. The standard approach to selecting this mixture relies on trial and error, which becomes impractical for large-scale pretraining. We propose a systematic method to determine the optimal data mixture …

1tfnrbP7oD r9iEesLhACpA

Building a Resilient Data Platform with Write-Ahead Log at Netflix

By Prudhviraj Karumanchi, Samuel Fu, Sriram Rangarajan, Vidhya Arvind, Yun Wang, John Lu Introduction Netflix operates at a massive scale, serving hundreds of millions of users with diverse content and features. Behind the scenes, ensuring data consistency, reliability, and efficient operations across various services presents a continuous challenge. At the heart of many critical functions lies …