Reduce LLM eval costs with smart proxies, caching, and sampling | saasbrowser.ai