YakGPT is a simple, frontend-only, ChatGPT UI you can use to either chat normally, or, more excitingly, use your mic + OpenAI's Whisper API to chat hands-free.
Some features:
* A few fun characters pre-installed
* No tracking or analytics, OpenAI is the only thing it calls out to
* Optimized for mobile use via hands-free mode and cross-platform compressed audio recording
* Your API key and chat history are stored in browser local storage only
* Open-source, you can either use the deployed version at Vercel, or run it locally
Planned features:
* Integrate Eleven Labs & other TTS services to enable full hands-free conversation
* Implement LangChain and/or plugins
* Integrate more ASR services that allow for streaming
Source code: https://github.com/yakGPT/yakGPT
I’d love for you to try it out and hear your feedback!
Most people can talk faster than they can type, but they can read faster than other people can talk. So an interface where I speak but read the response is an ideal way of interfacing with ChatGPT.
What would be nice is if I didn't have to press the mic button to speak -- if it could just tell when I was speaking (perhaps by saying "hey YakGPT"). But I see how that might be hard to implement.
Would love to hook this up to some smart glasses with a heads-up display where I could speak and read the response.
Most people I know type faster than they can talk. Also more accurate. I find talking a horrible interface to a computer while sitting down. On the move it is another story entirely of course.
By the way, chatgpt is not very fast either, so usually I type something in the chat and continue working while it generates the response.
> smart glasses
I just tried that; it works quite well, however, pressing the mic button kind of messes up that experience.
I gave up at
Creating an optimized production build ...TypeError: Cannot read properties of null (reading 'useRef')
edit: hacketyfixey, let's punch the thing in the face until it works:
(I am sorry)It uses 2 external calls to a javascript CDN for the microphone package and something else. It would probably be best if it was localhost calls only since it handles an API key
Any chance you could integrate the backend-api, and let me paste in my Bearer token from there?
* GPT-4 is decently faster when talking straight to the API
* The API is so stupidly cheap that it's basically a rounding error for me. Half an hour of chatting to GPT3.5 costs me $0.02
Would be curious what you mean by integrating the backend-api?
But the benefit from using the API is that you can change the model on the fly, so you chat with 3.5 until you notice that it's not responding properly and, with all the history you have (probably stored in your database), you can send a bigger request with a probably better response once with GPT-4 as the selected model.
I really wish the interface on chat.openai.org would allow me to switch between models in the same conversation in order to 1) not use up your quota of GPT-4 interactions per 3 hours as quickly and 2) not strain the backend unnecessarily when you know that starting a conversation with GPT-3.5 is efficient enough until you notice that you better switch models.
OpenAI already has this implemented: When you use up your quota of GPT-4 chats, it offers you to drop down into GPT-3.5 in that same conversation.
Maybe I’ll have to try this for a month and see if it end up costing more than $20. Thanks for creating it!
[0] https://sublimegpt.com
https://help.openai.com/en/articles/5722486-how-your-data-is...
Deleted Comment
That's not what "run locally" means. This isn't any more "local" than talking to chatgpt directly, which is never running locally.
I'm thinking: What would a GPT project manager do? What would a GPT money manager do? What would a GPT logistics manager do? GPT Data Analyst, Etc.
> Please enter your OpenAI key
...
Do people just not get it?
I would in fact rather give all my company secrets to this random dude than OpenAI.
It seems they are genuine, and they phrase it exactly as it is. The only thing I would have maybe wanted to see in the title is "open-source" or free software.
Deleted Comment
One feature I am missing from all these front ends is the ability to edit your text and generate new response from that point. Official chat gpt UI is the only one that seems to do that.