The Next Step in Personalization: Dynamic Sizzles

Authors:Bruce Wobbe, Leticia Kwok Additional Credits:Sanford Holsapple, Eugene Lok, Jeremy Kelly Introduction At Netflix, we strive to give our members an excellent personalized experience, helping them make the most successful and satisfying selections from our thousands of titles. We already personalize artwork and trailers, but we hadn’t yet personalized sizzle reels — until now. A sizzle reel is a montage …

Building on a year of focus to help IBM Power clients grow with hybrid cloud and AI

At the beginning of the year, we laid out a new strategy for IBM Power under the leadership of Ken King, who will be retiring by the end of 2023 after forty years with IBM. It is with immense gratitude that I thank Ken for his leadership not only across IBM Power, but for his …

monaideploy arch

Build a medical imaging AI inference pipeline with MONAI Deploy on AWS

This post is cowritten with Ming (Melvin) Qin, David Bericat and Brad Genereaux from NVIDIA. Medical imaging AI researchers and developers need a scalable, enterprise framework to build, deploy, and integrate their AI applications. AWS and NVIDIA have come together to make this vision a reality. AWS, NVIDIA, and other partners build applications and solutions …

image2 unQaZiP.max 1000x1000 1

Introducing Accurate Quantized Training (AQT) for accelerated ML training on TPU v5e

AI models continue to get bigger, requiring larger compute clusters with exa-FLOPs (10^18 FLOPs) of computing. While large-scale models continue to unlock new capabilities, driving down the cost of training and serving these models is the key to sustaining the pace of this innovation. Typically, the tensor operations (ops)1 are the most compute-intensive part of …

altup 1

Alternating updates for efficient transformers

Posted by Xin Wang, Software Engineer, and Nishanth Dikkala, Research Scientist, Google Research Contemporary deep learning models have been remarkably successful in many domains, ranging from natural language to computer vision. Transformer neural networks (transformers) are a popular deep learning architecture that today comprise the foundation for most tasks in natural language processing and also …

Putting data storage at the forefront of cloud security

We live in an era of unprecedented technology breakthroughs and opportunities. Recent advances in areas like AI and quantum computing offer transformative potential for businesses, but may also bring new risks and security challenges. IBM is working to address these challenges and evolving threats by helping organizations support highly secure, resilient and durable storage through …

altup

Alternating updates for efficient transformers

Posted by Xin Wang, Software Engineer, and Nishanth Dikkala, Research Scientist, Google Research Contemporary deep learning models have been remarkably successful in many domains, ranging from natural language to computer vision. Transformer neural networks (transformers) are a popular deep learning architecture that today comprise the foundation for most tasks in natural language processing and also …

ML 15684 00 1

Harnessing the power of enterprise data with generative AI: Insights from Amazon Kendra, LangChain, and large language models

Large language models (LLMs) with their broad knowledge, can generate human-like text on almost any topic. However, their training on massive datasets also limits their usefulness for specialized tasks. Without continued learning, these models remain oblivious to new data and trends that emerge after their initial training. Furthermore, the cost to train new LLMs can …

1 xc1yMqn.max 1000x1000 1

Using FFmpeg with Google Cloud Speech-to-Text

Google Cloud Speech-to-Text is a fully managed service that converts speech to text in real time. It can be used to transcribe audio and video files, create subtitles for videos, and build voice-activated applications. The service supports a wide range of audio formats, including WAV, MP3, and AAC. It can also transcribe audio in a …

Turing’s Mill: AI Supercomputer Revs UK’s Economic Engine

The home of the first industrial revolution just made a massive investment in the next one. The U.K. government has announced it will spend £225 million ($273 million) to build one of the world’s fastest AI supercomputers. Called Isambard-AI, it’s the latest in a series of systems named after a legendary 19th century British engineer …