Reduce LLM token waste by compressing repo context before code AI | saasbrowser.ai