Visualize LLM context usage to avoid token overruns and optimize prompts | saasbrowser.ai