Categories: FAANG

Layer-Wise Data-Free CNN Compression

We present an efficient method for compressing a trained neural network without using any data. Our data-free method requires 14x-450x fewer FLOPs than comparable state-of-the-art methods. We break the problem of data-free network compression into a number of independent layer-wise compressions. We show how to efficiently generate layer-wise training data, and how to precondition the network to maintain accuracy during layer-wise compression. We show state-of-the-art performance on MobileNetV1 for data-free low-bit-width quantization. We also show state-of-the-art performance on data-free…
AI Generated Robotic Content

Recent Posts

When she says she only likes open source dudes

submitted by /u/Jack_Fryy [link] [comments]

12 hours ago

SceneScout: Towards AI Agent-driven Access to Street View Imagery for Blind Users

People who are blind or have low vision (BLV) may hesitate to travel independently in…

12 hours ago

Why We Serve: Palantirians Reflect on Duty, Honor & Innovation

In honor of Independence Day, Palantir Veterans and Intelligence Community (IC) alums offer reflections on…

12 hours ago

Transforming network operations with AI: How Swisscom built a network assistant using Amazon Bedrock

In the telecommunications industry, managing complex network infrastructures requires processing vast amounts of data from…

12 hours ago

How to build a simple multi-agentic system using Google’s ADK

Agents are top of mind for enterprises, but often we find customers building one “super”…

12 hours ago

Sakana AI’s TreeQuest: Deploy multi-model teams that outperform individual LLMs by 30%

Sakana AI's new inference-time scaling technique uses Monte-Carlo Tree Search to orchestrate multiple LLMs to…

13 hours ago