Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Generative AI with Amazon Bedrock

You're reading from   Generative AI with Amazon Bedrock Build, scale, and secure generative AI applications using Amazon Bedrock

Arrow left icon
Product type Paperback
Published in Jul 2024
Publisher Packt
ISBN-13 9781803247281
Length 384 pages
Edition 1st Edition
Tools
Arrow right icon
Authors (2):
Arrow left icon
Shikhar Kwatra Shikhar Kwatra
Author Profile Icon Shikhar Kwatra
Shikhar Kwatra
Bunny Kaushik Bunny Kaushik
Author Profile Icon Bunny Kaushik
Bunny Kaushik
Arrow right icon
View More author details
Toc

Table of Contents (18) Chapters Close

Preface 1. Part 1: Amazon Bedrock Foundations FREE CHAPTER
2. Chapter 1: Exploring Amazon Bedrock 3. Chapter 2: Accessing and Utilizing Models in Amazon Bedrock 4. Chapter 3: Engineering Prompts for Effective Model Usage 5. Chapter 4: Customizing Models for Enhanced Performance 6. Chapter 5: Harnessing the Power of RAG 7. Part 2: Amazon Bedrock Architecture Patterns
8. Chapter 6: Generating and Summarizing Text with Amazon Bedrock 9. Chapter 7: Building Question Answering Systems and Conversational Interfaces 10. Chapter 8: Extracting Entities and Generating Code with Amazon Bedrock 11. Chapter 9: Generating and Transforming Images Using Amazon Bedrock 12. Chapter 10: Developing Intelligent Agents with Amazon Bedrock 13. Part 3: Model Management and Security Considerations
14. Chapter 11: Evaluating and Monitoring Models with Amazon Bedrock 15. Chapter 12: Ensuring Security and Privacy in Amazon Bedrock 16. Index 17. Other Books You May Enjoy

What are FMs?

Most of the generative AI models today are powered by the transformer-based architecture. In general, these generative AI models, also widely known as FMs, employ transformers due to their ability to process text one token at a time or entire sequences of text at once using self-attention. FMs are trained on massive amounts of data with millions or billions of parameters, allowing them to understand relationships between words in context to predict subsequent sequences. While models based on the transformer architecture currently dominate the field, not all FMs rely on this architecture. Some models are built using alternative techniques, such as generative adversarial networks (GANs) or variational autoencoders.

GANs utilize two neural networks pitted against each other in competition. The first network is known as the generator and is tasked with generating synthetic samples that mimic real data. For example, the generator could produce new images, texts, or audio...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime
Banner background image