Home
Categories
EXPLORE
True Crime
Comedy
Business
Society & Culture
Sports
News
History
About Us
Contact Us
Copyright
© 2024 PodJoint
00:00 / 00:00
Sign in

or

Don't have an account?
Sign up
Forgot password
https://is1-ssl.mzstatic.com/image/thumb/Podcasts211/v4/aa/6e/c5/aa6ec535-4bf6-6000-c0c7-d14efada8206/mza_9225735319382486240.jpg/600x600bb.jpg
The AI Concepts Podcast
Sheetal ’Shay’ Dhar
49 episodes
1 week ago
Show more...
Technology
Education,
Courses,
Science
RSS
All content for The AI Concepts Podcast is the property of Sheetal ’Shay’ Dhar and is served directly from their servers with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
Show more...
Technology
Education,
Courses,
Science
https://is1-ssl.mzstatic.com/image/thumb/Podcasts211/v4/aa/6e/c5/aa6ec535-4bf6-6000-c0c7-d14efada8206/mza_9225735319382486240.jpg/600x600bb.jpg
Module 2: The Transformer Architecture: History - The Bottleneck That Broke Language Models
The AI Concepts Podcast
7 minutes
1 week ago
Module 2: The Transformer Architecture: History - The Bottleneck That Broke Language Models
Shay breaks down why recurrent neural networks (RNNs) struggled with long-range dependencies in language: fixed-size hidden states and the vanishing gradient caused models to forget early context in long texts. He explains how LSTMs added gates (forget, input, output) to manage memory and improve short-term performance but remained serial, creating a training and scaling bottleneck that prevented using massive parallel compute. The episode frames this fundamental bottleneck in NLP and sets up the next episode on attention, ending with a brief reflection on persistence and steady effort.
The AI Concepts Podcast