Categories: FAANG

TRACT: Denoising Diffusion Models with Transitive Closure Time-Distillation

Denoising Diffusion models have demonstrated their proficiency for generative sampling. However, generating good samples often requires many iterations. Consequently, techniques such as binary time-distillation (BTD) have been proposed to reduce the number of network calls for a fixed architecture. In this paper, we introduce TRAnsitive Closure Time-distillation (TRACT), a new method that extends BTD. For single step diffusion,TRACT improves FID by up to 2.4x on the same architecture, and achieves new single-step Denoising Diffusion Implicit Models (DDIM) state-of-the-art FID (7.4 for…
AI Generated Robotic Content

Recent Posts

VibeVoice is crazy good (first try, no cherry-picking)

Installed VibeVoice using the wrapper this dude created. https://www.reddit.com/r/comfyui/comments/1n20407/wip2_comfyui_wrapper_for_microsofts_new_vibevoice/ Workflow is the multi-voice example one…

2 hours ago

7 Pandas Tricks for Efficient Data Merging

Data merging is the process of combining data from different sources into a unified dataset.

2 hours ago

How to Decide Between Random Forests and Gradient Boosting

When working with machine learning on structured data, two algorithms often rise to the top…

2 hours ago

Meet Boti: The AI assistant transforming how the citizens of Buenos Aires access government information with Amazon Bedrock

This post is co-written with Julieta Rappan, Macarena Blasi, and María Candela Blanco from the…

2 hours ago

In crowded voice AI market, OpenAI bets on instruction-following and expressive speech to win enterprise adoption

OpenAI's new speech model, gpt-realtime, hopes that its more naturalistic voices would make enterprises use…

3 hours ago

Watch Our Livestream Replay: Back to School in the Age of AI

We explored our latest investigations into how tech is shaping education today.

3 hours ago