Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Building AI Applications with Microsoft Semantic Kernel

You're reading from   Building AI Applications with Microsoft Semantic Kernel Easily integrate generative AI capabilities and copilot experiences into your applications

Arrow left icon
Product type Paperback
Published in Jun 2024
Publisher Packt
ISBN-13 9781835463703
Length 252 pages
Edition 1st Edition
Languages
Tools
Concepts
Arrow right icon
Author (1):
Arrow left icon
Lucas A. Meyer Lucas A. Meyer
Author Profile Icon Lucas A. Meyer
Lucas A. Meyer
Arrow right icon
View More author details
Toc

Table of Contents (14) Chapters Close

Preface 1. Part 1:Introduction to Generative AI and Microsoft Semantic Kernel
2. Chapter 1: Introducing Microsoft Semantic Kernel FREE CHAPTER 3. Chapter 2: Creating Better Prompts 4. Part 2: Creating AI Applications with Semantic Kernel
5. Chapter 3: Extending Semantic Kernel 6. Chapter 4: Performing Complex Actions by Chaining Functions 7. Chapter 5: Programming with Planners 8. Chapter 6: Adding Memories to Your AI Application 9. Part 3: Real-World Use Cases
10. Chapter 7: Real-World Use Case – Retrieval-Augmented Generation 11. Chapter 8: Real-World Use Case – Making Your Application Available on ChatGPT 12. Index 13. Other Books You May Enjoy

Using memory within chats and LLMs

As we have seen before, models have a size limit called a context window. The size limit includes both the prompt with the user request and the response. The default context window for a model such as GPT-3.5, for example, is 4,096 bytes, meaning that both your prompt, including the user request, and the answer that GPT-3.5 provides can have at most 4,096 bytes; otherwise, you will get an error, or the response will cut off in the middle.

If your application uses a lot of text data, for example, a 10,000-page operating manual, or allows people to search and ask questions about a database of hundreds of documents with each one having 50 pages, you need to find a way of including just the relevant portion of this large dataset with your prompt. Otherwise, the prompt alone could be larger than the context window, resulting in an error, or the remaining context window could be so short that there would be no space for the model to provide a good answer...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime
Banner background image