695: NLP with Transformers, feat. Hugging Face's Lewis Tunstall

Published: July 11, 2023, 11 a.m.

What are transformers in AI, and how do they help developers to run LLMs efficiently and accurately? This is a key question in this week\u2019s episode, where Hugging Face\u2019s ML Engineer Lewis Tunstall sits down with host Jon Krohn to discuss encoders and decoders, and the importance of continuing to foster democratic environments like GitHub for creating open-source models.This episode is brought to you by the AWS Insiders Podcast, by WithFeeling.ai, the company bringing humanity into AI, and by Modelbit, for deploying models in seconds. Interested in sponsoring a SuperDataScience Podcast episode? Visit JonKrohn.com/podcast for sponsorship information.In this episode you will learn:\u2022 What a transformer is, and why it is so important for NLP [04:34]\u2022 Different types of transformers and how they vary [11:39]\u2022 Why it\u2019s necessary to know how a transformer works [31:52]\u2022 Hugging Face\u2019s role in the application of transformers [57:10]\u2022 Lewis Tunstall\u2019s experience of working at Hugging Face [1:02:08]\u2022 How and where to start with Hugging Face libraries [1:18:27]\u2022 The necessity to democratize ML models in the future [1:25:25]Additional materials: www.superdatascience.com/695