Semantic Caching for LLMs: FastAPI, Redis, and Embeddings