Using `llm` to handle large input context
Today, I ran into an issue where I wanted to use repomix to pack a large codebase into a single file to pass to an LLM, but I couldn't paste the output into any of the UIs I typically use. The React apps all became sluggish as I waited for ~500,000 tokens to paste.