Part 3: Common Generative AI Architectures
Many LLMs have limitations on how many tokens they can process. If the size of the input document exceeds the token limit of the model, additional steps need to be taken to bypass this limit. Even though many models increase their token limits and larger and larger texts could be fed into the model to generate a summary, the techniques described in this part are still relevant to optimize your applications for speed and performance.
This part has the following chapters:
- Chapter 7, Working with Long Context
- Chapter 8, Building Chatbots
- Chapter 9, Tools and Function Calling
- Chapter 10, Agents in Generative AI
- Chapter 11, Agentic Workflows