Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Deep Learning with PyTorch Lightning

You're reading from   Deep Learning with PyTorch Lightning Swiftly build high-performance Artificial Intelligence (AI) models using Python

Arrow left icon
Product type Paperback
Published in Apr 2022
Publisher Packt
ISBN-13 9781800561618
Length 366 pages
Edition 1st Edition
Languages
Tools
Arrow right icon
Authors (2):
Arrow left icon
Dheeraj Arremsetty Dheeraj Arremsetty
Author Profile Icon Dheeraj Arremsetty
Dheeraj Arremsetty
Kunal Sawarkar Kunal Sawarkar
Author Profile Icon Kunal Sawarkar
Kunal Sawarkar
Arrow right icon
View More author details
Toc

Table of Contents (15) Chapters Close

Preface 1. Section 1: Kickstarting with PyTorch Lightning
2. Chapter 1: PyTorch Lightning Adventure FREE CHAPTER 3. Chapter 2: Getting off the Ground with the First Deep Learning Model 4. Chapter 3: Transfer Learning Using Pre-Trained Models 5. Chapter 4: Ready-to-Cook Models from Lightning Flash 6. Section 2: Solving using PyTorch Lightning
7. Chapter 5: Time Series Models 8. Chapter 6: Deep Generative Models 9. Chapter 7: Semi-Supervised Learning 10. Chapter 8: Self-Supervised Learning 11. Section 3: Advanced Topics
12. Chapter 9: Deploying and Scoring Models 13. Chapter 10: Scaling and Managing Training 14. Other Books You May Enjoy

Summary

Transfer learning is one of the most common ways used to cut compute costs, save time, and get the best results. In this chapter, we learned how to build models with ResNet-50 and pre-trained BERT architectures using PyTorch Lightning.

We have built an image classifier and a text classifier, and along the way, we have covered some useful PyTorch Lightning life cycle methods. We have learned how to make use of pre-trained models to work on our customized datasets with less effort and a smaller number of training epochs. Even with very little model tuning, we were able to achieve decent accuracy.

While transfer learning methods work great, their limitations should also be borne in mind. They work incredibly well for language models because the given dataset's text is usually made up of the same English words as in your core training set. When the core training set is very different from your given dataset, performance suffers. For example, if you want to build an image...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime
Banner background image