Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Applied Machine Learning and High-Performance Computing on AWS

You're reading from   Applied Machine Learning and High-Performance Computing on AWS Accelerate the development of machine learning applications following architectural best practices

Arrow left icon
Product type Paperback
Published in Dec 2022
Publisher Packt
ISBN-13 9781803237015
Length 382 pages
Edition 1st Edition
Tools
Arrow right icon
Authors (4):
Arrow left icon
Trenton Potgieter Trenton Potgieter
Author Profile Icon Trenton Potgieter
Trenton Potgieter
Shreyas Subramanian Shreyas Subramanian
Author Profile Icon Shreyas Subramanian
Shreyas Subramanian
Farooq Sabir Farooq Sabir
Author Profile Icon Farooq Sabir
Farooq Sabir
Mani Khanuja Mani Khanuja
Author Profile Icon Mani Khanuja
Mani Khanuja
Arrow right icon
View More author details
Toc

Table of Contents (20) Chapters Close

Preface 1. Part 1: Introducing High-Performance Computing
2. Chapter 1: High-Performance Computing Fundamentals FREE CHAPTER 3. Chapter 2: Data Management and Transfer 4. Chapter 3: Compute and Networking 5. Chapter 4: Data Storage 6. Part 2: Applied Modeling
7. Chapter 5: Data Analysis 8. Chapter 6: Distributed Training of Machine Learning Models 9. Chapter 7: Deploying Machine Learning Models at Scale 10. Chapter 8: Optimizing and Managing Machine Learning Models for Edge Deployment 11. Chapter 9: Performance Optimization for Real-Time Inference 12. Chapter 10: Data Visualization 13. Part 3: Driving Innovation Across Industries
14. Chapter 11: Computational Fluid Dynamics 15. Chapter 12: Genomics 16. Chapter 13: Autonomous Vehicles 17. Chapter 14: Numerical Optimization 18. Index 19. Other Books You May Enjoy

Observing the results

The recommendation provided by Inference Recommender includes instance metrics, performance metrics, and cost metrics.

Instance metrics include InstanceType, InitialInstanceCount, and EnvironmentParameters, which are tuned according to the job for better performance.

Performance metrics include MaxInvocations and ModelLatency, whereas cost metrics include CostPerHour and CostPerInference.

These metrics enable you to make informed trade-offs between cost and performance. For example, if your business requirement is overall price performance with an emphasis on throughput, then you should focus on CostPerInference. If your requirement is a balance between latency and throughput, then you should focus on ModelLatency and MaxInvocations metrics.

You can view the results of the Inference Recommender job either through an API call or in the SageMaker Studio UI.

The following is the code snippet for observing the results:

…
data = [
  ...
lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime
Banner background image