Enhancing Document Extraction with Azure AI Document Intelligence and LangChain for RAG Workflows.
18/07/2024
Introduction One of the ways to optimize cost and performance of Large Language Models (LLMs) is to cache the responses from LLMs, this is sometimes referred to as “semantic caching”. In this blog, we will discuss the approaches, benefits, common scenarios and key considerations for using semantic caching. What…