We use cookies to personalize content and to analyze our traffic. Please decide if you are willing to accept cookies from our website.

Just Cache It (Part 1): Maintaining Context with APIs and LLMs

Mon., 20. January 2025 | 5 min read

In 2023, GitHub hosted 1.8 million AI projects compared to approximately 650,000 in 2020. AI applications are rapidly increasing due to model improvements and API use. APIs allow developers to easily use LLMs in their applications allowing for fast development and deployment. However, high cost is one concern with API calls to LLMs. AI service providers usually charge users for API use based on the number of processed tokens. Money goes down the drain when AI applications send frequent API calls with similar content. Prompt caching or context caching solves this issue by caching context determined by the developer. This reduces the number of tokens sent via API calls. AI engineers can turn to prompt caching to decrease inference fees and reduce latency in their AI applications.

How Context Was Handled Before Prompt Caching

LLMs from major AI service providers are usually stateful …

Tactive Research Group Subscription

To access the complete article, you must be a member. Become a member to get exclusive access to the latest insights, survey invitations, and tailored marketing communications. Stay ahead with us.

Become a Client!

Similar Articles

Locking down LLMs to Combat Jailbreaks

Locking down LLMs to Combat Jailbreaks

LLM jail-breaking (also known as LLM manipulation) forces LLMs to exhibit unwanted behavior. These LLMs may become examples of irresponsible and unethical AI, depending on what they are forced to do. Cybersecurity teams can ensure that their LLMs are responsible and ethical through resilience testing for jailbreaks and implementing multiple guardrails to combat jailbreaks.
Navigate the Technology Trends of 2025 – Compliance

Navigate the Technology Trends of 2025 – Compliance

From the EU AI Act to emerging state-level AI laws in the US, 2025 promises heightened scrutiny and demands on IT systems. Organizations must adopt forward-thinking strategies, leveraging emerging technologies like LLMs and governance tools, to navigate this terrain effectively. CIOs should prioritise proactive compliance measures to safeguard operations and maintain competitive advantage.
EU Regulations, Technologies, AI Realities and Cyber Risks: 2024 Tech Insights

EU Regulations, Technologies, AI Realities and Cyber Risks: 2024 Tech Insights

2024 saw significant shifts in technology, with the EU's AI Act and DMA impacting businesses alongside the rise of modular laptops and the persistent threat of cyber attacks. This review highlights some of the developments that interested IT leaders. This list suggests CIOs and IT executives should continue to prioritise compliance, evaluate new technologies, and strengthen cybersecurity in 2025.