Semantic Caching for LLMs: TTLs, Confidence, and Cache Safety