Titan AI LogoTitan AI

GPTCache

7,823
566
Python

Project Description

Semantic cache for LLMs. Fully integrated with LangChain and llama_index.

GPTCache: Semantic cache for LLMs. Fully integrated with LangChain and llama_index.

Project Title

GPTCache — Semantic Cache for Large Language Models (LLMs) to Reduce API Costs and Improve Speed

Overview

GPTCache is a Python-based library designed to create a semantic cache for storing responses from large language models (LLMs), such as ChatGPT. It aims to slash LLM API costs by up to 10x and boost response speed by 100x. Fully integrated with LangChain and llama_index, GPTCache offers a powerful solution for developers looking to optimize their LLM-based applications.

Key Features

  • Semantic caching to store and retrieve LLM responses
  • Integration with LangChain and llama_index for seamless operation
  • Support for multiple programming languages through Docker image
  • Automatic installation of additional libraries for extended features
  • Extensive documentation and community support

Use Cases

  • Developers using LLMs in applications to reduce API costs and improve response times
  • Enterprises handling high traffic levels and large volumes of LLM requests
  • Researchers and data scientists utilizing LLMs for semantic search and analysis

Advantages

  • Significant reduction in LLM API costs
  • Dramatic improvement in response speed for LLM services
  • Easy integration with existing LLM-based applications
  • Flexibility to use with any language through Docker

Limitations / Considerations

  • The project is under active development, and the API may change
  • Support for new API or models is not actively added due to the evolving nature of LLMs
  • Requires Python 3.8.1 or higher for optimal performance

Similar / Related Projects

  • LangChain: A framework for building LLM applications, with which GPTCache is fully integrated.
  • llama_index: Another tool for indexing and检索LLMs, also compatible with GPTCache.
  • OpenAI API: The original API for accessing LLMs like ChatGPT, which GPTCache aims to optimize.

Basic Information

  • GitHub: GPTCache
  • Stars: 7,789
  • License: MIT
  • Last Commit: 2025-10-10

📊 Project Information

  • Project Name: GPTCache
  • GitHub URL: https://github.com/zilliztech/GPTCache
  • Programming Language: Python
  • ⭐ Stars: 7,789
  • 🍴 Forks: 566
  • 📅 Created: 2023-03-24
  • 🔄 Last Updated: 2025-10-10

🏷️ Project Topics

Topics: [, ", a, i, g, c, ", ,, , ", a, u, t, o, g, p, t, ", ,, , ", b, a, b, y, a, g, i, ", ,, , ", c, h, a, t, b, o, t, ", ,, , ", c, h, a, t, g, p, t, ", ,, , ", c, h, a, t, g, p, t, -, a, p, i, ", ,, , ", d, o, l, l, y, ", ,, , ", g, p, t, ", ,, , ", l, a, n, g, c, h, a, i, n, ", ,, , ", l, l, a, m, a, ", ,, , ", l, l, a, m, a, -, i, n, d, e, x, ", ,, , ", l, l, m, ", ,, , ", m, e, m, c, a, c, h, e, ", ,, , ", m, i, l, v, u, s, ", ,, , ", o, p, e, n, a, i, ", ,, , ", r, e, d, i, s, ", ,, , ", s, e, m, a, n, t, i, c, -, s, e, a, r, c, h, ", ,, , ", s, i, m, i, l, a, r, i, t, y, -, s, e, a, r, c, h, ", ,, , ", v, e, c, t, o, r, -, s, e, a, r, c, h, ", ]


📚 Documentation

  • [Release
  • [pip download
  • [Codecov
  • [License
  • [Twitter

This article is automatically generated by AI based on GitHub project information and README content analysis

Titan AI Explorehttps://www.titanaiexplore.com/projects/gptcache-618276831en-USTechnology

Project Information

Created on 3/24/2023
Updated on 11/11/2025