SWE-grep was able to hit ~700tokens/s and Cursor ~300token/s, hard to compare the precision/recall and cost effectiveness though, considering SWE-grep also adopted a "hack" of running it on Cerebras.
I'm trying to kickstart a RL-based code search project called "op-grep" here[1], still pretty early, but looking for collaborators!
[0]: https://cognition.ai/blog/swe-grep [1]: https://github.com/aperoc/op-grep
I mean it's not much, but the concept just resonates with me and I want to share it. Sad I can't share even simple opinion nowadays ...
I've drafted an architecture, with the steps mainly as so: 1. Collect actions (grep/glob/read) policies either from usage logs or open datasets 2. Optimize by removing redundant actions or parallelization 3. Train model on optimized action policy 4. Release model as a single file, MCP tool (Refer to repo for visual diagram of the architecture)
I've just released the base model and added `openai_forwarder.py` to start collecting action policies.
Looking for more eyes and contributors to make this a reality, thanks!