Curious what people have been building with LLMs.
I worked on a chrome extension a few weeks ago that skips sponsorship sections in YouTube videos by reading through the transcript. Also was trying to experiment with an LLM to explain a function call chain across languages (in this case MakeFile, Python, Bash). I've tried running a few telegram bots that are PRE prompted to do certain things like help you with taxes.
What are you building?
What does the stack look like? How do you deploy it?
I ended up using it for more general purpose things because being able to have a hands-free phone call with an AI turned out to be pretty useful.
It's offline now, but here's the code with all the stack and deployment info: https://github.com/kevingduck/ChatGPT-phone/
Edit: forgot to mention this was all running off a $35 raspberry pi.
But it sounds damn creative as a project.
And you're right, I'm not a sales guy. This project is for people like me who want a risk-free place to learn the basics of sales so that when I do talk to an actual human, I won't panic and freeze up like I always do.
Dead Comment
Deleted Comment
For the dictated recipes, I told him to dictate just "flat" the words and numbers. So that I had paragraphs of recipes.
For the scanned recipes, I used Google OCR (I found out it was the best one quality wise).
For both sets of recipes, I then used GPT4 to "format" the unformatted recipes into well formatted Markdown. It successfully fixed typos and bad OCR from Google.
We then pasted all that well formatted text into a big Google Docs, and added images. Using OpenAI image generation I generated images for each of the 250+ recipes. For some of them I had to manually curate it, given that some of the recipes are for typical Mexican food: For example there's a (delicious) recipe called "PibiPollo" that for the unitiated it may look like a stew, so I had to tell something like "large corn tamale with thick hard crust".
In the end, the book was pretty nice! We distributed digital copies within the family and everybody was amazed :) . I loved spending time doing that.
https://drive.google.com/file/d/1OGE-zfNHHDnALbhgmf3lykBjcSg...
It is in Spanish though.
Are you using the "Record Audio" action or something else? Ideally the shortcut would stop listening after a pause like the native Dictate feature does it. At a minimum Record Audio seems to require hitting spacebar to stop - not great but not terrible.
Wondering: How big is you monthly OpenAI bill when using all these tools? Only a few $$$, or is it higher?
The backend is a Python FastAPI that uses ChromaDB to store my resume and Q&A pairs, OpenAI, and Airtable to log requests and responses. The UI is Sveltekit.
I'm currently building a different tool and will apply some learnings to my Interactive Resume AI. Instead of Airtable, I am going to use LangSmith for observability.
I started writing and my Substack articles are also linked to via my website. I'm currently working on applying sentence window retrieval and that article will be out shortly. This is part of a #buildinpublic effort to help build my brand as well.
I've been unemployed since Sept as a Senior Software Engineer. The market is tough so I'm focusing on the above to help get employment or a contract.
I also built Resume Chatbot but using slightly different stack: Python, Langchain, Faiss as vector store, MongoDB to store chat logs and Streamlit for UI. Here is a link: https://www.artkreimer.com/resume/ or you can try it on streamlit https://art-career-bot.streamlit.app/. Code is available here https://github.com/kredar/data_analytics/tree/master/career_.... Great thread and I got some ideas for my next project. Thanks a lot everyone.
The actual questions I got did not provide a response that is to my liking. Most of that is due in part because I'm using gpt3.5 since gpt4-turbo is a lot more expensive, and I can learn a lot more by using an inferior LLM.
For example, using an llm router to analyze the query and route to a specific helper function with a specific prompt would be helpful. Sometimes a user starts with a greeting but the response is a pre-written "Sorry an answer cannot be found". Questions are typically grouped into a category such as skills, experience, project, personal (ie: where are you located), preferences (ie: favorite language), and general interview questions (ie: why should I hire you). Questions in categories can be better answered by using a different prompt and/or RAG technique.
I was contacted by a company recruiter for a small healthcare SaaS in California and had 3 interviews recently. When I looked up the job, only 7 people had applied in 2 weeks on LinkedIn. They are a very real company with very real people, but their job post is not getting seen (it's not a promoted post).
My next AI project will be to scrape LinkedIn jobs, analyze it for repost/promoted behavior, group it by consulting/headhunters vs company job post, eliminate duplicates, and filter based on my skillset and hard-no qualities (such as can't work if I live in California, must be in EST but I'm in PST timezone, requires Java experience, etc).
* I built a real life Pokedex to recognize Pokemon [video] https://www.youtube.com/watch?v=wVcerPofkE0
* I used ChatGPT to filter nice comments and print them in my office [video] https://www.youtube.com/watch?v=AonMzGUN9gQ
* I built a general purpose chat assistant into an old intercom [video] https://www.youtube.com/watch?v=-zDdpeTdv84
Again, nothing terribly useful, but all fun.
1. sketch (in notebook, ai for pandas) https://github.com/approximatelabs/sketch
2. datadm (open source, "chat with data", with support for the open source LLMs (https://github.com/approximatelabs/datadm)
3. Our main product: julyp. https://julyp.com/ (currently under very active rebrand and cleanup) -- but a "chat with data" style app, with a lot of specialized features. I'm also streaming me using it (and sometimes building it) every weekday on twitch to solve misc data problems (https://www.twitch.tv/bluecoconut)
For your next question, about the stack and deploy: We're using all sorts of different stacks and tooling. We made our own tooling at one point (https://github.com/approximatelabs/lambdaprompt/), but have more recently switched to just using the raw requests ourselves and writing out the logic ourselves in the product. For our main product, the code just lives in our next app, and deploys on vercel.
The thing I'm working on now is AI mock interviewing. It's basically scratching my own itch, since I hate leetcode prep, and have found I can learn better through interaction. To paste a blurb from an earlier comment of mine:
I'm building https://comp.lol. It's AI powered mock coding interviews, FAANG style. Looking for alpha testers when I release, sign up if you wanna try it out or just wanna try some mock coding. If its slow to load, sorry, everything runs on free tiers right now.
I really dislike doing leetcode prep, and I can't intuitively understand the solutions by just reading them. I've found the best way for me to learn is to seriously try the problem (timed, interview like conditions), and be able to 'discuss' with the interviewer without just jumping to reading the solution. Been using and building this as an experiment to try prepping in a manner I like.
It's not a replacement for real mock interviews - I think those are still the best, but they're expensive and time consuming. I'm hoping to get 80% of the benefit in an easier package.
I just put a waitlist in case anyone wants to try it out and give me feedback when I get it out
Gonna apologize in advance about the copywriting. Was more messing around for my own amusement, will probably change later
Thanks for signing up, will send out an email once its ready to take for a spin!
Runs on a local LLM, because even using GPT3 costs would have added up quickly.
Currently requires CUDA and uses a 10.7B model but if anyone wants to try a smaller one and report results let me know on github and I can give some help.
https://github.com/thomasj02/AiFilter