Reduce LLM API bills: prompt compaction, model-switching, caching | saasbrowser.ai