Readit News logoReadit News
joshstrange · a year ago
I could not agree more with this. 90% of AI features feel tacked on and useless and that’s before you get to the price. Some of the services out here are wanting to charge 50% to 100% more for their sass just to enable “AI features”.

I’m actually having a really hard time thinking of an AI feature other than coding AI feature that I actually enjoy. Copilot/Aider/Claude Code are awesome but I’m struggling to think of another tool I use where LLMs have improved it. Auto completing a sentence for the next word in Gmail/iMessage is one example, but that existed before LLMs.

I have not once used the features in Gmail to rewrite my email to sound more professional or anything like that. If I need help writing an email, I’m going to do that using Claude or ChatGPT directly before I even open Gmail.

petekoomen · a year ago
One of the interesting things I've noticed is that the best experiences I've had with AI are with simple applications that don't do much to get in the way of the model, e.g. chatgpt and cursor/windsurf.

I'm hopeful that as devs figure out how to build better apps with AI we'll have have more and more "cursor moments" in other areas in our lives

dangus · a year ago
Perhaps the real takeaway is that there really is only one product, two if you count image generation.

Perhaps the only reason Cursor is so good is because editing code is so similar to the basic function of an LLM without anything wrapped around it.

Like, someone prove me wrong by linking 3 transformative AI products that:

1. Have nothing to do with "chatting" to a thin wrapper (couldn't just be done inside a plain LLM with a couple of file uploads added for additional context)

2. Don't involve traditional ML that has existed for years and isn't part of the LLM "revolution."

3. Has nothing to do with writing code

For example, I recently used an AI chatbot that was supposed to help me troubleshoot a consumer IoT device. It basically regurgitated steps from the manual and started running around in circles because my issue was simply not covered by documentation. I then had to tell it to send me to a human. The human had more suggestions that the AI couldn't think of but still couldn't help because the product was a piece of shit.

Or just look at Amazon Q. Ask it a basic AWS question and it'll just give you a bogus "sorry I can't help with that" answer where you just know that running over to chatgpt.com will actually give you a legitimate answer. Most AI "products" seem to be castrated versions of ChatGPT/Claude/Gemini.

That sort of overall garbage experience seems to be what is most frequently associated with AI. Basically, a futile attempt to replace low-wage employees that didn't end up delivering any value to anyone, especially since any company interested in eliminating employees just because "fuck it why not" without any real strategy probably has a busted-ass product to begin with.

Putting me on hold for 15 minutes would have been more effective at getting me to go away and no compute cycles would have been necessary.

teeray · a year ago
> This demo uses AI to read emails instead of write them

LLMs are so good at summarizing that I should basically only ever read one email—from the AI:

You received 2 emails today that need your direct reply from X and Y. 1 is still outstanding from two days ago, _would you like to send an acknowledgment_? You received 6 emails from newsletters you didn’t sign up for but were enrolled after you bought something _do you want to unsubscribe from all of them_ (_make this a permanent rule_).

namaria · a year ago
I have fed LLMs PDF files, asked about the content and gotten nonsense. I would be very hesitant to trust them to give me an accurate summary of my emails.
nradov · a year ago
LLMs are terrible at summarizing technical emails where the details matter. But you might get away with it, at least for a while, in low performing organizations that tolerate preventable errors.
FabHK · a year ago
I got an email from the restaurant saying "We will confirm your dinner reservation as soon as we can", and Apple Intelligence summarizing it as "Dinner reservation confirmed." Maybe it can not only summarize, but also see the future??
amrocha · a year ago
I fed an LLM the record of a chat between me and a friend, and asked it to summarize the times that we met in the past 3 months.

Every time it gave me different results, and not once did it actually get it all right.

LLMs are horrible for summarizing things. Summarizing is the art of turning low information density text into high information density text. LLMs can’t deal in details, so they can never accurately summarize anything.

koolba · a year ago
> LLMs are so good at summarizing that I should basically only ever read one email—from the AI

This could get really fun with some hidden text prompt injection. Just match the font and background color.

Maybe these tools should be doing the classic air gap approach of taking a picture of the rendered content and analyzing that.

joshstrange · a year ago
What system are you using to do this? I do think that this would provide value for me. Currently, I barely read my emails, which I'm not exactly proud of, but it's just the reality. So something that summarized the important things every day would be nice.
meroes · a year ago
Do you ever check its work?
throwaway290 · a year ago
What is the reason to unsub ever in that world? Are you saying the LLM can't skip emails? Seems like an arbitrary rule
danielbln · a year ago
I enjoy Claude as a general purpose "let's talk about this niche thing" chat bot, or for general ideation. Extracting structured data from videos (via Gemini) is quite useful as well, though to be fair it's not a super frequent use case for me.

That said, coding and engineering is by far the most common usecase I have for gen AI.

joshstrange · a year ago
Oh, I'm sorry if it wasn't clear. I use Claude and ChatGPT to talk to about a ton of topics. I'm mostly referring to AI features being added to existing SaaS or software products. I regularly find that moving the conversation to ChatGPT or Claude is much better than trying to use anything that they may have built into their existing product.
sanderjd · a year ago
I think the other application besides code copiloting that is already extremely useful is RAG-based information discovery a la Notion AI. This is already a giant improvement over "search google docs, and slack, and confluence, and jira, and ...".

Just integrated search over all the various systems at a company was an improvement that did not require LLMs, but I also really like the back and forth chat interface for this.

knightscoop · a year ago
I wonder sometime if this is why there is such an enthusiasm gap over AI between tech people and the general public. It's not just that your average person can't program; it's that they don't even conceptually understand why programming could unlock.
rcarmo · a year ago
The e-mail agent example is so good that it makes everything else I’ve seen and used pointless by comparison. I wonder why nobody’s done it that way yet.
bamboozled · a year ago
Have you ever been cooking and asked Siri to set a timer? That's basically the most used AI feature outside of "coding" I can think of.
joshstrange · a year ago
Setting a timer and setting a reminder. Occasionally converting units of measure. That's all I can rely on Siri (or Alexa) for and even then sometimes Siri doesn't make it clear if it did the thing. Most importantly, "set a reminder", it shows the text, and then the UI disappears, sometimes the reminder was created, sometimes not. It's maddening since I'm normally asking to be reminded about something important that I need to get recorded/tracked so I can "forget" it.

The number of times I've had 2 reminders fire back-to-back because I asked Siri again to create one since I was _sure_ it didn't create the first one.

Siri is so dumb and it's insane that more heads have not rolled at Apple because of it (I'm aware of the recent shakeup, it's about a decade too late). Lastly, whoever decided to ship the new Siri UI without any of the new features should lose their job. What a squandered opportunity and effectively fraud IMHO.

More and more it's clear that Tim Cook is not the person that Apple needs at the helm. My mom knows Siri sucks, why doesn't the CEO and/or why is he incapable of doing anything to fix it. Get off your Trump-kissing, over-relying-on-China ass and fix your software! (Siri is not the only thing rotten)

dale_glass · a year ago
I find that ChatGPT o3 (and the other advanced reasoning models) are decently good at answering questions with a "but".

Google is great at things like "Top 10 best rated movies of 2024", because people make lists of that sort of thing obsessively.

But Google is far less good at queries like "Which movies look visually beautiful but have been critically panned?". For that sort of thing I have far more luck with chatgpt because it's much less of a standard "top 10" list.

joshstrange · a year ago
o3 has been a big improvement on Deep Research IMHO. o1 (or whatever model I originally used with it) was interesting but the results weren't always great. o3 has done some impressive research tasks for me and, unlike the last model I used, when I "check its work" it has always been correct.
nicolas_t · a year ago
I like perplexity when I need a quick overview of a topic with references to relevant published studies. I often use it when researching what the current research says on parenting questions or education. It's not perfect but because the answers link to the relevant studies it's a good way to get a quick overview of research on a given topic
tomjen3 · a year ago
I really like my speech-to-text program, and I find using ChatGPT to look up things and answer questions is a much superior experience to Google, but otherwise, I completely agree with you.

Companies see that AI is a buzzword that means your stock goes up. So they start looking at it as an answer to the question: "How can I make my stock go up?" instead of "How can I create a better product", and then let the stock go up from creating a better product.

bigstrat2003 · a year ago
Honestly I don't even enjoy coding AI features. The only value I get out of AI is translation (which I take with a grain of salt because I don't know the other language and can't spot hallucinations, but it's the best tool I have), and shitposting (e.g. having chatGPT write funny stories about my friends and sending it to them for a laugh). I can't say there's an actual productive use case for me personally.
genewitch · a year ago
I've anecdotally tested translations by ripping the video with subtitles and having whisper subtitle it, and also asking several AI to translate the .srt or .vtt file (subtotext I think does this conversion if you don't wanna waste tokens on the metadata)

Whisper large-v3, the largest model I have, is pretty good, getting nearly identical translations to chatgpt or whatever, Google's default speech to text. The fun stuff is when you ask for text to text translations from LLMs.

I did a real small writeup with an example but I don't have a place to publish nor am I really looking for one.

I used whisper to transcribe nearly every "episode" of the Love Line syndicated radio show from 1997-2007 or so. It took, iirc, several days. I use it to grep the audio, as it were. I intend to do the same with my DVDs and such, just so I never have to Google "what movie / tv show is that line from?" I also have a lot of art bell shows, and a few others to transcribe.

apwell23 · a year ago
garmin wants me to pay for some gen-ai workout messages on connect plus. Its the most absurd AI slop of all. Same with strava. I workout for mental relaxation and i just hate this AI stuff being crammed in there.

Atleast clippy was kind of cute.

nradov · a year ago
Strava employees claim that casual users like the AI activity summaries. Supposedly users who don't know anything about exercise physiology didn't know how to interpret the various metrics and charts. I don't know if I believe that but it's at least plausible.

Personally I wish I could turn off the AI features, it's a waste of space.

danielbln · a year ago
Strava's integration is just so lackluster. It literally turns four numbers from right above the slop message into free text. Thanks Strava, I'm a pro user for a decade, finally I can read "This was a hard workout" after my run. Such useful, much AI.
sandspar · a year ago
I use AI chatbots for 2+ hours a day but the Garmin thing was too much for me. The day they released their AI Garmin+ subscription, I took off my Forerunner and put it in a drawer. The whole point of Garmin is that it feels emotionally clean to use. Garmin adding a scammy subscription makes the ecosystem feel icky, and I'm not going to wear a piece of clothing that makes me feel icky. I don't think I'll buy a Garmin watch again.

(Since taking off the watch, I miss some of the data but my overall health and sleep haven't changed.)

bigstrat2003 · a year ago
At this point, "we aren't adding any AI features" is a selling point for me. I've gotten real tired of AI slop and hype.
Ntrails · a year ago
> Auto completing a sentence for the next word in Gmail/iMessage is one example

Interestingly, I despise that feature. It breaks the flow of what is actually a very simple task. Now I'm reading, reconsidering if the offered thing is the same thing I wanted over and over again.

The fact that I know this and spend time repeatedly disabling the damned things is awfully tiresome (but my fault for not paying for my own email etc etc)

genewitch · a year ago
I've been using Fastmail in lieu of gmail for ten or eleven years. If you have a domain and control the DNS, I recommend it. At least you're not on Google anymore, and you're paying for fastmail, so it feels better - less like something is reading your emails.
Andugal · a year ago
> I’m actually having a really hard time thinking of an AI feature other than coding AI feature that I actually enjoy.

If you attend a lot of meetings, having an AI note-taker take notes for you and generate a structured summary, follow-up email, to-do list, and more will be an absolute game changer.

(Disclaimer, I'm the CTO of Leexi, an AI note-taker)

AlexandrB · a year ago
The catch is: does anyone actually read this stuff? I've been taking meeting notes for meetings I run (without AI) for around 6 months now and I suspect no one other than myself has looked at the notes I've put together. I've only looked back at those notes once or twice.

A big part of the problem is even finding this content in a modern corporate intranet (i.e. Confluence) and having a bunch of AI-generated text in there as well isn't going to help.

soco · a year ago
I'm not a CTO so maybe your wold is not my world, but for me the advantage of taking the notes myself is that only I know what's important to me, or what was news to me. Teams Premium - you can argue it's so much worse than your product - takes notes like "they discussed about the advantages of ABC" but maybe exactly those advantages are advantageous to know right? And so on. Then like others said, I will review my notes once to see if there's a followup, or a topic to research, and off they go to the bin. I have yet to need the meeting notes of last year. Shortly put: notes apps are to me a solution in search of a problem.
yesfitz · a year ago
Is Leexi's AI note-taker able to raise its hand in a meeting (or otherwise interrupt) and ask for clarification?

As a human note-taker, I find the most impactful result of real-time synthesis is the ability to identify and address conflicting information in the moment. That ability is reliant on domain knowledge and knowledge of the meeting attendees.

But if the AI could participate in the meeting in real time like I can, it'd be a huge difference.

bluGill · a year ago
But that isn't writing for me, it is taking notes for me. There is a difference. I don't need something to write for me - I know how to write. What I need is someone to clean up grammar, fact check the details, and otherwise clean things up. I have dysgraphia - a writing disorder - so I need help more than most, but I still don't need something to write my drafts for me: I can get that done well enough.
yoyohello13 · a year ago
We've had the built-in Teams summary AI for a while now and it absolutely misses important details and nuance that causes problems later.
joshstrange · a year ago
I've used multiple of these types of services and I'll be honest, I just don't really get the value. I'm in a ton of meetings and I run multiple teams but I just take notes myself in the meetings. Every time I've compared my own notes to the notes that the the AI note taker took, it's missing 0-2 critical things or it focuses on the wrong thing in the meeting. I've even had the note taker say essentially the opposite of what we decided on because we flip-flopped multiple times during the meeting.

Every mistake the AI makes is completely understandable, but it's only understandable because I was in the meeting and I am reviewing the notes right after the meeting. A week later, I wouldn't remember it, which is why I still just take my own notes in meetings. That said, having having a recording of the meeting and or some AI summary notes can be very useful. I just have not found that I can replace my note-taking with an AI just yet.

One issue I have is that there doesn't seem to be a great way to "end" the meeting for the note taker. I'm sure this is configurable, but some people at work use Supernormal and I've just taken to kicking it out of of meetings as soon as it tries to join. Mostly this is because I have meetings that run into another meeting, and so I never end the Zoom call between the meetings (I just use my personal Zoom room for all meetings). That means that the AI note taker will listen in on the second meeting and attribute it to the first meeting by accident. That's not the end of the world, but Supernormal, at least by default, will email everyone who was part of the the meeting a rundown of what happened in the meeting. This becomes a problem when you have a meeting with one group of people and then another group of people, and you might be talking about the first group of people in the second meeting ( i.e. management issues). So far I have not been burned badly by this, but I have had meeting notes sent out to to people that covered subjects that weren't really something they needed to know about or shouldn't know about in some cases.

Lastly, I abhor people using an AI notetaker in lieu of joining a meeting. As I said above, I block AI note takers from my zoom calls but it really frustrates me when an AI joins but the person who configured the AI does not. I'm not interested in getting messages "You guys talked about XXX but we want to do YYY" or "We shouldn't do XXX and it looks like you all decided to do that". First, you don't get to weigh in post-discussion, that's incredibly rude and disrespectful of everyone's time IMHO. Second, I'm not going to help explain what your AI note taker got wrong, that's not my job. So yeah, I'm not a huge fan of AI note takers though I do see where they can provide some value.

Yizahi · a year ago
In my company have a few "summaries" made by Zoom neural net, which we share for memes on the joke chats, they are so hilariously bad. No one uses that functionality seriously. I don't know about your app, but I've yet to see a working note taker in the wild.
UncleMeat · a year ago
You do you.

I attend a lot of meetings and I have reviewed the results of an AI note taker maybe twice ever. Getting an email with a todo-list saves a bit of time of writing down action items during a meeting, but I'd hardly consider it a game changer. "Wait, what'd we talk about in that meeting" is just not a problem I encounter often.

My experience with AI note takers is that they are useful for people who didn't attend the meeting and people who are being onboarded and want to be able to review what somebody was teaching them in the meeting and much much much less useful for other situations.

tlogan · a year ago
At the end of the day, it comes down to one thing: knowing what you want. And AI can’t solve that for you.

We’ve experimented heavily with integrating AI into our UI, testing a variety of models and workflows. One consistent finding emerged: most users don’t actually know what they want to accomplish. They struggle to express their goals clearly, and AI doesn’t magically fill that gap—it often amplifies the ambiguity.

Sure, AI reduces the learning curve for new tools. But paradoxically, it can also short-circuit the path to true mastery. When AI handles everything, users stop thinking deeply about how or why they’re doing something. That might be fine for casual use, but it limits expertise and real problem-solving.

So … AI is great—but the current diarrhea of “let’s just add AI here” without thinking through how it actually helps might be a sign that a lot of engineers have outsourced their thinking to ChatGPT.

petekoomen · a year ago
> They struggle to express their goals clearly, and AI doesn’t magically fill that gap—it often amplifies the ambiguity.

One surprising thing I've learned is that a fast feedback loop like this:

1. write a system prompt 2. watch the agent do the task, observe what it gets wrong 3. update the system prompt to improve the instructions

is remarkably useful in helping people write effective system prompts. Being able to watch the agent succeed or fail gives you realtime feedback about what is missing in your instructions in a way that anyone who has ever taught or managed professionally will instantly grok.

serpix · a year ago
What I've found with agents is that they stray from the task and even start to flip flop on implementations, going back and forth on a solution. They never admit they don't know something and just brute force a solution even though the answer cannot be found without trial and error or actually studying the problem. I repeatedly fall back to reading the docs and just finishing the job myself as the agent just does not know what to do.
kristjank · a year ago
I have also experienced this in the specific domain of well-learned idiots finding pseudo-explanations for why a technical choice should be taken, despite not knowing anything about the topic.

I have witnessed a colleague look up a component datasheet on ChatGPT and repeating whatever it told him (despite the points that it made weren't related to our use case). The knowledge monopoly in about 10 years when the old-guard programming crowd finally retires and/or unfortunately dies will be in the hands of people that will know what they don't know and be able to fill the gaps using appropriate information sources (including language models). The rest will probably resemble Idiocracy on a spectrum from frustrating to hilarious.

hnthrow90348765 · a year ago
In the process of finding out what customers or a PM/PO wants, developers ask clarifying questions given an ambiguous start. An AI could be made to also ask these questions. It may do this reasonably better than some engineers by having access to a ton of questions in its training data.

By using an AI, you might be making a reasonable guess that your problem has been solved before, but maybe not the exact details. This is true for a lot of technical tasks as I don't need to reinvent database access from first principles for every project. I google ORMs or something in my particular language and consider the options.

Even if the AI doesn't give you a direct solution, it's still a prompt for your brain as if you were in a conversation.

Deleted Comment

mNovak · a year ago
Just want to say the interactive widgets being actually hooked up to an LLM was very fun.

To continue bashing on gmail/gemini, the worst offender in my opinion is the giant "Summarize this email" button, sitting on top of a one-liner email like "Got it, thanks". How much more can you possibly summarize that email?

petekoomen · a year ago
Thank you! @LewisJEllis and I wrote a little framework for "vibe writing" that allows for writing in markdown and adding vibe-coded react components. It's a lot of fun to use!
ipaddr · a year ago
Can we all quickly move to a point in time where vibe-code is not a word
DesaiAshu · a year ago
My websites have this too with MDX, it's awesome. Reminds me of the old Bret Victor interactive tutorials back around when YC Research was funding HCI experiments
carterschonwald · a year ago
Very nice example of an actually usefully interactive essay.
bambax · a year ago
It is indeed a working demo, hitting

  https://llm.koomen.dev/v1/chat/completions
in the OpenAI API format, and it responds to any prompt without filtering. Free tokens, anyone?

More seriously, I think the reason companies don't want to expose the system prompt is because they want to keep some of the magic alive. Once most people understand that the universal interface to AI is text prompts, then all that will remain is the models themselves.

petekoomen · a year ago
That's right. llm.koomen.dev is a cloudflare worker that forwards requests to openai. I was a little worried about getting DDOSed but so far that hasn't been an issue, and the tokens are ridiculously cheap.
amiantos · a year ago
Blog author seems smart (despite questionable ideas about how much real world users would want to interact with any of his elaborate feature concepts), you hope he's actually just got a bunch of responses cached and you're getting a random one each time from that endpoint... and that freely sent content doesn't actually hit OpenAI's APIs.
jihadjihad · a year ago
It's like the memes where people in the future will just grunt and gesticulate at the computer instead.
ChaitanyaSai · a year ago
Loved those! How are those created?
Xenoamorphous · a year ago
I used that button in Outlook once and the summary was longer than the original email
Etheryte · a year ago
"k"
crote · a year ago
I think a big problem is that the most useful AI agents essentially go unnoticed.

The email labeling assistant is a great example of this. Most mail services can already do most of this, so the best-case scenario is using AI to translate your human speech into a suggestion for whatever format the service's rules engine uses. Very helpful, not flashy: you set it up once and forget about it.

Being able to automatically interpret the "Reschedule" email and suggest a diff for an event in your calendar is extremely useful, as it'd reduce it to a single click - but it won't be flashy. Ideally you wouldn't even notice there's a LLM behind it, there's just a "confirm reschedule button" which magically appears next to the email when appropriate.

Automatically archiving sales offers? That's a spam filter. A really good one, mind you, but hardly something to put on the frontpage of today's newsletters.

It can all provide quite a bit of value, but it's simply not sexy enough! You can't add a flashy wizard staff & sparkles icon to it and charge $20 / month for that. In practice you might be getting a car, but it's going to look like a horseless carriage to the average user. They want Magic Wizard Stuff, not invest hours into learning prompt programming.

petekoomen · a year ago
> Most mail services can already do most of this

I'll believe this when I stop spending so much time deleting email I don't want to read.

phito · a year ago
And dumpster diving in my spam folder for actually important emails
sanderjd · a year ago
Yeah but I'm looking forward to the point where this is not longer about trying to be flashy and sexy, but just quietly using a new technology for useful things that it's good at. I think things are headed that direction pretty quickly now though! Which is great.
crote · a year ago
Honestly? I think the AI bubble will need to burst first. Making the rescheduling of appointments and dozens of tasks like that slightly more convenient isn't a billion-dollar business.

I don't have a lot of doubt that it is technically doable, but it's not going to be economically viable when it has to pay back hundreds of billions of dollars of investments into training models and buying shiny hardware. The industry first needs to get rid of that burden, which means writing off the training costs and running inference on heavily-discounted supernumerary hardware.

oceanplexian · a year ago
A lot of people assume that AI naturally produces this predictable style writing but as someone who has dabbled in training a number of fine tunes that's absolutely not the case.

You can improve things with prompting but can also fine tune them to be completely human. The fun part is it doesn't just apply to text, you can also do it with Image Gen like Boring Reality (https://civitai.com/models/310571/boring-reality) (Warning: there is a lot of NSFW content on Civit if you click around).

My pet theory is the BigCo's are walking a tightrope of model safety and are intentionally incorporating some uncanny valley into their products, since if people really knew that AI could "talk like Pete" they would get uneasy. The cognitive dissonance doesn't kick in when a bot talks like a drone from HR instead of a real person.

palsecam · a year ago
> My pet theory is the BigCo's are walking a tightrope of model safety and are intentionally incorporating some uncanny valley into their products, since if people really knew that AI could "talk like Pete" they would get uneasy. The cognitive dissonance doesn't kick in when a bot talks like a drone from HR instead of a real person.

FTR, Bruce Schneier (famed cryptologist) is advocating for such an approach:

We have a simple proposal: all talking AIs and robots should use a ring modulator. In the mid-twentieth century, before it was easy to create actual robotic-sounding speech synthetically, ring modulators were used to make actors’ voices sound robotic. Over the last few decades, we have become accustomed to robotic voices, simply because text-to-speech systems were good enough to produce intelligible speech that was not human-like in its sound. Now we can use that same technology to make robotic speech that is indistinguishable from human sound robotic again.https://www.schneier.com/blog/archives/2025/02/ais-and-robot...

MichaelDickens · a year ago
Reminds me of the robot voice from The Incredibles[1]. It had an obviously-robotic cadence where it would pause between every word. Text-to-speech at the time already knew how to make words flow into each other, but I thought the voice from The Incredibles sounded much nicer than the contemporaneous text-to-speech bots, while also still sounding robotic.

[1] https://www.youtube.com/watch?v=_dxV4BvyV2w

momojo · a year ago
Like adding the 'propane smell' to propane.
nyanpasu64 · a year ago
That doesn't sound like ring modulation in a musical sense (IIRC it has a modulator above 30 Hz, or inverts the signal instead of attenuating?), so much as crackling, cutting in and out, or an overdone tremolo effect. I checked in Audacity and the signal only gets cut out, not inverted.
Semaphor · a year ago
Interestingly, it's just kinda hiding the normal AI issues, but they are all still there. I think people know about those "normal" looking pictures, but your example has many AI issues, especially with hands and background
GuinansEyebrows · a year ago
> but can also fine tune them to be completely human

what does this mean? that it will insert idiosyncratic modifications (typos, idioms etc)?

a2128 · a year ago
If you play around with base models, they will insert typos, slang, they will generate curse words and pointless internet flamewars
pmarreck · a year ago
Loved the fact that the interactive demos were live.

You could even skip the custom system prompt entirely and just have it analyze a randomized but statistically-significant portion of the corpus of your outgoing emails and their style, and have it replicate that in drafts.

You wouldn't even need a UI for this! You could sell a service that you simply authenticated to your inbox and it could do all this from the backend.

It would likely end up being close enough to the mark that the uncanny valley might get skipped and you would mostly just be approving emails after reviewing them.

Similar to reviewing AI-generated code.

The question is, is this what we want? I've already caught myself asking ChatGPT to counterargue as me (but with less inflammatory wording) and it's done an excellent job which I've then (more or less) copy-pasted into social-media responses. That's just one step away from having them automatically appear, just waiting for my approval to post.

Is AI just turning everyone into a "work reviewer" instead of a "work doer"?

petekoomen · a year ago
honestly you could try this yourself today. Grab a few emails, paste them into chatgpt, and ask it to write a system prompt that will write emails that mimic your style. Might be fun to see how it describes your style.

to address your larger point, I think AI-generated drafts written in my voice will be helpful for mundane, transaction emails, but not for important messages. Even simple questions like "what do you feel like doing for dinner tonight" could only be answered by me, and that's fine. If an AI can manage my inbox while I focus on the handful of messages that really need my time and attention that would be a huge win in my book.

segh · a year ago
The system prompt can include examples. That is often a good idea.
crote · a year ago
It all depends on how you use it, doesn't it?

A lot of work is inherently repetitive, or involves critical but burdensome details. I'm not going to manually write dozens of lines of code when I can do `bin/rails generate scaffold User name:string`, or manually convert decimal to binary when I can access a calculator within half a second. All the important labor is in writing the prompt, reviewing the output, and altering it as desired. The act of generating the boilerplate itself is busywork. Using a LLM instead of a fixed-functionality wizard doesn't change this.

The new thing is that the generator is essentially unbounded and silently degrades when you go beyond its limits. If you want to learn how to use AI, you have to learn when not to use it.

Using AI for social media is distinct from this. Arguing with random people on the internet has never been a good idea and has always been a massive waste of time. Automating it with AI just makes this more obvious. The only way to have a proper discussion is going to be face-to-face, I'm afraid.

emaro · a year ago
About writing a counterargument for social media: I kinda get it, but what's the end game of this? People reading generated responses others (may have) approved? Do we want that? I think I don't.
mvieira38 · a year ago
It's what we want, though, isn't it? AI should make our lives easier, and it's much easier (and more productive) to review work already done than to do it yourself. Now, if that is a good development morally/spiritually for the future of mankind is another question... Some would argue industrialization was bad in that respect and I'm not even sure I fully disagree
ai_ · a year ago
No? Not everyone's dream is being a manager. I like writing code, it's fun! Telling someone else to go write code for me so that I can read it later? Not fun, avoid it if possible (sometimes it's unavoidable, we don't have unlimited time).
selkin · a year ago
> and it's much easier (and more productive) to review work already done than to do it yourself

This isn't the tautology you imagine it to be.

Consider the example given here of having AI write one line draft response to emails. To validate such response, you have to: (1) read the original email, (2) understand it, (3) decide what you want to communicate in your reply, then (4) validate that the suggested draft communicates the same.

If the AI gave a correct answer, you saved yourself from typing one sentence, which you probably already formulated in your head in step (3). A minor help, at best.

But if the AI was wrong, you now have to write that reply yourself.

To get positive expected utility from the above scenario, you'd need the probability of the AI to be correct extremely high, and even then, the savings would be small.

A task that requires more effort to turn ideas into deliverables would have better expectation, but complex tasks often have results that are not simple nor easy to check, so the savings may not be as meaningful as you naively assume.

__float · a year ago
The live demos were neat! I was playing around with "The Pete System Prompt", and one of the times, it signed the email literally "Thanks, [Your Name]" (even though Pete was still right there in the prompt).

Just a reminder that these things still need significant oversight or very targeted applications, I suppose.

segh · a year ago
The live demos are using a very cheap and not very smart model. Do not update your opinion on AI capabilities based on the poor performance of gpt-4o-mini
bluGill · a year ago
What is the point? The effort to write the email is equal to the effort to ask the AI to write the email for you. Only when the AI turns your unprofessional style into something professional is any effort saved - but the "professional" sounding style is most of the time wrong and should get dumped into junk.
aldous · a year ago
Yeah, I'm with you on this one. Surely in most instances it is easier to just bash out the email plus you get the added bonus of exercising your own mind: vocabulary, typing skills, articulating concepts, defining appropriate etiquette. As the years role by I aiming to be more conscious and diligent with my own writing and communication, not less. If one extrapolates on the use of AI for such basic communication, is there a risk some of us lose our ability to meaningfully think for ourselves? The information space of the present day already feels like it is devolving; shorter and shorter content, lack of nuance, reductive messaging. Sling AI in as a mediator for one to one communication too and it feels perilous for social cohesion.
throwaway2037 · a year ago
I cannot remember which blogging platform shows you the "most highlighted phrase", but this would be mine:

    > The email I'd have written is actually shorter than the original prompt, which means I spent more time asking Gemini for help than I would have if I'd just written the draft myself. Remarkably, the Gmail team has shipped a product that perfectly captures the experience of managing an underperforming employee.
This paragraph makes me think of the old Joel Spolsky blog post that he probably wrote 20+ years ago about his time in the Israeli Defence Forces, explaining to readers how showing is more impactful than telling. I feel like this paragraph is similar. When you have a low performer, you wonder to yourself, in the beginning, why does it seem like I spend more time explaining the task than the low performer spends to complete it!?

adr1an · a year ago
Medium
pchristensen · a year ago
Kindle.
kristjank · a year ago
I tread carefully with anyone that by default augments their (however utilitarian or conventionally bland) messages with language models passing them as their own. Prompting the agent to be as concise as you are, or as extensive, takes just as much time in the former case, and lacks the underlying specificity of your experience/knowledge in the latter.

If these were some magically private models that have insight into my past technical explanations or the specifics of my work, this would be a much easier bargain to accept, but usually, nothing that has been written in an email by Gemini could not have been conceived of by a secretary in the 1970s. It lacks control over the expression of your thoughts. It's impersonal, it separates you from expressing your thoughts clearly, and it separates your recipient from having a chance to understand you the person thinking instead of you the construct that generated a response based on your past data and a short prompt. And also, I don't trust some misandric f*ck not to sell my data before piping it into my dataset.

I guess what I'm trying to say is: when messaging personally, summarizing short messages is unnecessary, expanding on short messages generates little more than semantic noise, and everything in between those use cases is a spectrum deceived by the lack of specificity that agents usually present. Changing the underlying vague notions of context is not only a strangely contortionist way of making a square peg fit an umbrella-shaped hole, it pushes around the boundaries of information transfer in a way that is vaguely stylistic, but devoid of any meaning, removed fluff or added value.

petekoomen · a year ago
Agreed! As i mentioned in the piece I don't think LLMs are very useful for original writing because instructing an agent to write anything from scratch inevitably takes more time than writing it yourself.

Most of the time I spend managing my inbox is not spent on original writing, however. It's spent on mundane tasks like filtering, prioritizing, scheduling back-and-forths, introductions etc. I think an agent could help me with a lot of that, and I dream of a world in which I can spend less time on email and finally be one of those "inbox zero" people.

Retric · a year ago
The counter argument is some people are terrible at writing. Millions of people sit at the bottom of any given bell curve.

I’d never trust a summery from a current generation LLM for something as critical as my inbox. Some hypothetical drastically improved future AI, sure.

derektank · a year ago
For the case of writing emails, I tend to agree though I think creative writing is an exception. Pairing with an LLM really helps overcome the blank page / writer's block problem because it's often easier to identify what you don't want and then revise all the flaws you see.
elieskilled · a year ago
On that topic I’m the founder of inbox zero: https://getinboxzero.com

May help you get half way there

rahimnathwani · a year ago

  instructing an agent to write anything from scratch inevitably takes more time than writing it yourself
But you can reuse your instructions with zero additional effort. I have some instructions that I wrote for a 'Project' in Claude (and now a 'Gem' in Gemini). The instructions give writing guidelines for a children's article about a topic. So I just write 'write an article about cross-pollination' and a minute later I have an article I can hand to my son.

Even if I had the subject matter knowledge, it would take me much longer to write an article with the type of style and examples that I want.

(Because you said 'from scratch', I deliberately didn't choose an example that used web search or tools.)

jon_richards · a year ago
Writing an email with AI and having the recipient summarize it with AI is basically all the fun of jpeg compression, but more bandwidth instead of less.

https://m.youtube.com/watch?v=jmaUIyvy8E8

jonplackett · a year ago
Why can’t the LLM just learn your writing style from your previous emails to that person?

Or a your more general style for new people.

It seems like Google at least should have a TONNE of context to use for this.

Like in his example emails about being asked to meet - it should be checking the calendar for you and putting in if you can / can’t or suggesting an alt time you’re free.

If it can’t actually send emails without permission there’s less harm with giving an LLM more info to work with - and it doesn’t need to get it perfect. You can always edit.

If it deals with the 80% of replies that don’t matter much then you have 5X more time to spend on the 20% that do matter.

unoti · a year ago
> Why can’t the LLM just learn your writing style from your previous emails to that person?

It totally could. For one thing you could fine tune the model, but I don't think I'd recommend that. For this specific use case, imagine an addition to the prompt that says """To help you with additional context and writing style, here snippets of recent emails Pete wrote to {recipient}: --- {recent_email_snippets} """

samrolken · a year ago
They are saving this for some future release I would guess. A “personalization”-focused update wave/marketing blitz/privacy Overton window shift.
foxglacier · a year ago
There's a whole lot of people who struggle to write professionally or when there's any sort of conflict (even telling your boss you won't come to work). It can be crippling trying to find the right wording and certainly take far longer than writing a prompt. AI is incredible for these people. They were never going to express their true feelings anyway and were just struggling to write "properly" or in a way that doesn't lead to misunderstandings. If you can just smash out good emails without a second thought, you wouldn't need it.
skeptrune · a year ago
>As I mentioned above, however, a better System Prompt still won't save me much time on writing emails from scratch.

>The thing that LLMs are great at is reading text and transforming it, and that's what I'd like to use an agent for.

Interestingly, the OP agrees with you here and noted in the post that the LLMs are better at transforming data than creating it.

kristjank · a year ago
I reread those paragraphs. I find the transformative effect of the email missing from the whole discussion. The end result of the inbox examples is to change some internal information in the mind of the recipient. Agent working within the context of the email has very little to contribute because it does not know the OP's schedule, dinner plans, whether he has time for the walk and talk or if he broke his ankle last week... I'd be personally afraid to have something rummaging in my social interface that can send (and let's be honest, idiots will CtrlA+autoreply their whole inboxes) invites, timetables, love messages etc. in my name. It has too many lemmas that need to be fulfilled before it can be assumed competent, and none of those are very well demonstrated. It's cold fusion technology. Feasible, should be nice if it worked, but it would really be a disappointment if someone were to use it in its current state.
jimbokun · a year ago
A lot of people would love to have a 1970s secretary capable of responding to many mundane requests without any guidance.
bluGill · a year ago
I have a large part of that though. The computer (outlook today) just schedules meetings rooms for me ensuring there are not multiple different meetings in it at the same time. I can schedule my own flights.

When I first started working the company rolled out the first version of meeting scheduling (it wasn't outlook), and all the other engineers loved it - finally they could figure out how to schedule our own meetings instead of having the secretary do it. Apparently the old system was some mainframe based things other programmers couldn't figure out (I never worked with it so I can't comment on how it was). Likewise scheduling a plane ticket involved calling travel agents and spending a lot of time on hold.

If you are a senior executive you still have a secretary. However by the 1970s the secretary for most of us would be department secretary that handled 20-40 people not just our needs, and thus wasn't in tune with all those details. However most of us don't have any needs that are not better handled by a computer today.

kristjank · a year ago
I would too, but I would have to trust AI at least as much as a 1970s secretary not to mess up basic facts about myself or needlessly embellish/summarize my conversations with known correspondents. Comparing agents and past office cliches was not to imply agents do it and it's stupid; I'm implying agents claim to do it, but don't.
AlienRobot · a year ago
So AI is SaaS (Secretary as a Service)
calf · a year ago
AI for writing or research is useful like a dice roll. Terence Tao famously showed how talking to an LLM gave him an idea/approach to a proof that he hadn't immediately thought of (but probably he would have considered it eventually). The other day I wrote an unusal, four-word neologism that I'm pretty sure no one has ever seen, and the AI immediately drew the correct connection to more standard terminology and arguments used, so I did not even have to expand/explain and write it out myself.

I don't know but I am considering the possibility that even for everyday tasks, this kind of exploratory shortcut can be a simple convenience. Furthermore, it is precisely the lack of context that enables LLMs to make these non-human, non-specific connective leaps, their weakness also being their strength. In this sense, they bode as a new kind of discursive common-ground--if human conversants are saying things that an LLM can easily catch then LLMs could even serve as the lowest-common-denominator for laying out arguments, disagreements, talking past each other, etc. But that's in principle, and in practice that is too idealistic, as long as these are built and owned as capitalist IPs.

AndrewHart · a year ago
Aside from saving time, I'm bad at writing. Especially emails. I often open ChatGPT, paste in the whole email chain, write out the bullets of the points I want to make and ask it to draft a response which frames it well.
ori_b · a year ago
I'd prefer to get the bullet points. There's no need to waste time reading autogenerated filler.
Swizec · a year ago
> write out the bullets of the points I want to make

Just send those bullet points. Everyone will thank you

worik · a year ago
My boss does that I am sure

One of their dreadful behaviors, among many

My advice is to stop doing this for the sake of your colleagues

hooverd · a year ago
Hopefully you're specifying that your email is written with ChatGPT so other parties can paste it back into ChatGPT and get bullet points back instead of wasting their time reading the slop.
ripe · a year ago
Why not just send the bullet points? Kinder to your audience than sending them AI slop.

Dead Comment