Wrote a small script to test how LLMs handle growing context sizes. It sends larger prompts step by step and measures slowdown or failure. Pretty rough but useful so far. Curious if others have ideas for improving it.
ChatGPT for example almost always uses them. I'm sure they are more common in academic writing, but its now super common on boards like Reddit.
- No one ever