EPISODE · May 25, 2020 · 52 MIN
Rethinking Model Size: Train Large, Then Compress with Joseph Gonzalez - #378
from The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) · host Sam Charrington
Today we’re joined by Joseph Gonzalez, Assistant Professor in the EECS department at UC Berkeley. In our conversation, we explore Joseph’s paper “Train Large, Then Compress: Rethinking Model Size for Efficient Training and Inference of Transformers,” which looks at compute-efficient training strategies for models. We discuss the two main problems being solved; 1) How can we rapidly iterate on variations in architecture? And 2) If we make models bigger, is it really improving any efficiency?
NOW PLAYING
Rethinking Model Size: Train Large, Then Compress with Joseph Gonzalez - #378
No transcript for this episode yet
Similar Episodes
No similar episodes found.