Just chill out and try not doing that for a few seconds. Not saying this is always the wrong thing to do, but if your intention is to have your friends and colleagues appreciate your listening skills I think this is the starting point. Be more present to the conversation; responding to what they are actually saying and not just habitually reacting.
What other CLI tools are people using to work with LLMs in the terminal?
There one comment here about https://github.com/paul-gauthier/aider and Ollama is probably the most widely used CLI tool at the moment: https://github.com/ollama/ollama/blob/main/README.md#quickst...
npm i mktute
You can select between local model (ollama), claude 3.5 sonnet, or gpt-4. I've been surprised to find sonnet much better in performance and price for this task.
Quick landing page feedback: Could you update it so that the first two screen captures are not the same content? And then pick a different second example that isn't also about leaving a note? More variation in the landing page examples would give me a better intro.
Is it possible to just call them "researchers at xyz institute" and link to their lab rather than "Scandinavian researchers"?
Just a starting point. But if you are interested in this space, fork away and build it into something useful!
My personal take is that the current tools on the market are too expensive. The cost should go way, way down. This should stay open source. Patients should have easy access to full audio recordings and transcriptions of their medical appointments. One can dream!
This extension simply pulls down a transcript for the video you are currently viewing and sends this to the OpenAI API with your task request.
"Generate a vegetarian version of this recipe."
"Extract the materials list for this project."
"What was the book they recommended?"
This extension does not use a speech to text model to transcribe the video or an image recognition model to actually pass along information about what took place in the video. Automatically generated YouTube transcripts are not great quality, but they tend to be totally fine for the lightweight use cases when you just have a simple question on a video with dialogue.
gpt-3.5-turbo-1106 with the 16k token context window is used by default for videos under about ten minutes. For longer videos, the gpt-4-1106-preview model with a 128k context window is automatically selected. You should be able to use this for videos of to about four hours in length, but this is highly dependent on the volume of dialogue.
Best thought of as a very simple starting point for configuring a diy extension of your own. I actually use this on a regular basis now! Its nice for just getting around ads and not suffering through long youtube videos sent by well-meaning friends. Would love to see how other people configure this to be even better. Fork away!
Not saying these are all bad all the time. Just things to maybe start noticing. See what happens when you just hold off on these types of reactions for a moment.
You don't even have to do anything. Just... chill. Be silent. Think about what the person said.
I've found this is a more actionable place to start than something like "read this book" or "take this course"