Categories: FAANG

Neural Transducer Training: Reduced Memory Consumption with Sample-wise Computation

The neural transducer is an end-to-end model for automatic speech recognition (ASR). While the model is well-suited for streaming ASR, the training process remains challenging. During training, the memory requirements may quickly exceed the capacity of state-of-the-art GPUs, limiting batch size and sequence lengths. In this work, we analyze the time and space complexity of a typical transducer training setup. We propose a memory-efficient training method that computes the transducer loss and gradients sample by sample. We present optimizations to increase the efficiency and parallelism of the…
AI Generated Robotic Content

Recent Posts

Top Megelin Deals for Laser and LED Therapy Devices (2026)

This Mother's Day, Megelin is slashing prices on its best-selling laser and LED devices.

23 mins ago

New AI tool predicts airport traffic to avert devastating collisions

In managing airport traffic, small errors can cause catastrophe. A group from the CMU Robotics…

23 mins ago

Had to keep it going

Continuing the music video u/optimisoprimeo posted: https://www.reddit.com/r/StableDiffusion/comments/1t64gni/so_far_this_is_my_favorite_usecase_for_ltx/ submitted by /u/hidden2u [link] [comments]

23 hours ago

What Matters in Practical Learned Image Compression

One of the major differentiators unlocked by learned codecs relative to their hard-coded traditional counterparts…

23 hours ago

Secure short-term GPU capacity for ML workloads with EC2 Capacity Blocks for ML and SageMaker training plans

As companies of various sizes adopt graphic processing units (GPU)-based machine learning (ML) training, fine-tuning…

23 hours ago