e.g. with mitmproxy and llama-cpp-python server
python -m llama_cpp.server --n_ctx 4096 --n_gpu_layers 1 --model ./path/to/..gguf
and then with mitmproxy in another terminal mitmproxy -p 5001 --mode reverse:http://127.0.0.1:8000
and then set this in your vscode settings.json (the same as for localpilot): "github.copilot.advanced": {
"debug.testOverrideProxyUrl": "http://localhost:5001",
"debug.overrideProxyUrl": "http://localhost:5001"
}
works way better for me than localpilot
they pushed the rocket to the limits in this flight