This is cool, but I wish it were integrated into tools already used for coding and writing rather than having it be a separate app.
This also demonstrates the type of things Google could do with Gemini integrated into Google Docs if they step up their game a bit.
Honestly I’m scratching my head on OpenAI’s desire to double down on building out their consumer B2C use cases rather than truly focussing on being the infrastructure/API provider for other services to plug into. If I had to make a prediction, I think OpenAI will end up being either an infrastructure provider OR a SaaS, but not both, in the long-term (5-10 yrs from now).
When they are focusing on just being an API provider then they will be in a market with (long term) razor thin margins and high competition - most likely unable to build a deep moat. But if you can shape customers habits to always input "chatgpt.com" into the browser whenever they want to use AI then that's a very powerful moat. Those customers will also most likely be on a subscription basis, meaning much more flexibility in pricing and more rent for openAI (people using it less then what OpenAI calculates for subscription costs).
> the type of things Google could do with Gemini integrated into Google Docs
Google already does have this in Google Docs (and all their products)? You can ask it questions about the current doc, select a paragraph and ask click on "rewrite", things like that. Has helped me get over writer's block at least a couple of times. Similarly for making slides etc. (It requires the paid subscription if you want to use it from a personal account.)
I wouldn't be surprised to see Apple add something like this to Pages and some of their other apps. Their approach to AI, from what we've seen so far, has been about integrating it into existing apps and experiences, rather than making a separate AI app. I have to imagine this is the way forward, and these stand alone apps are basically tech demos for what is possible, rather than end-state for how it should be consumed by the masses.
I agree with you on where OpenAI will/should sit in 5-10 years. However, I don't think them building the occasional tool like this is unwarranted, as it helps them show the direction companies could/should head with integration into other tools. Before Microsoft made hardware full time, they would occasionally produce something (or partner with brands) to show a new feature Windows supports as a way to tell the OEMs out there, "this is what we want you to do and the direction we'd like the PC to head." The UMPC[0] was one attempt at this which didn't take off. Intel also did something like this with the NUC[1]. I view what OpenAI is doing as a similar concept, but applied to software.
Every app with a significant installed user base is adding AI features.
OP is lamenting that Cursor and OpenAI chose to create new apps instead of integrating with (someone else’s) existing apps. But this is a result of a need to be always fully unblocked.
Also, owning the app opens up greater financial potential down the line…
How many people use Pages these days? I don't think Apple even mentions the product in their WWDC these days. My guess is that most people either use Microsoft suite as required by their employer or use cloud based knowledge base/notes tools like Notion/Quip/Obsidian/Confluence etc. I doubt Apple thinks it worthwhile to invest in these products.
75% of OpenAI's revenue is coming from their consumer business - the better question is the long term viability of their public API.
But if they believe they're going to reach AGI, it makes no sense to pigeonhole themselves to the interface of ChatGPT. Seems like a pretty sensible decision to maintain both.
75%? Thats astonishing to me. Where are you able to see those details?
It wouldn't surprise me if not a lot of enterprises are going through OpenAI's enterprise agreements - most already have a relationship with Microsoft in one capacity or another so going through Azure just seems like the lowest friction way to get access. If how many millions we spend on tokens through Azure to OpenAI is any indication of what other orgs are doing, I would expect consumer's $20/month to be a drop in the bucket.
The most amazing thing with notebooklm is that is can turn your docs into a very high quality podcast of two people discussing the content of your docs.
ChatGPT itself is them copying their own API users, this is just them building out more features already built by users. My guess is they know they don't have a long term edge in models alone, so they are going to rely on expanding ChatGPT for better margins and to keep getting training data from users. They obviously want to control the platform, not integrate with other platforms
Their API is unusable due to rate limits. Myself and my wife have both had ideas, started using it, and found other approaches after hitting rate limits. I tried funding more money in the account to increase the rate limits and it did not work. I imagine they see poor growth there because of this.
It's pretty trivial to get increased limits, I've used the API for a few consulting projects and got to tier 4 in a month. At that point you can burn near $200 a day and 2 million tokens per minute.
You only need 45 days to get tier 5 and if you have that many customers after 45 days you should just apply to YC lol.
Maybe you checked over a year ago, which was the wild wild West at the time, they didn't even have the tier limits.
You need to use it for some time to get into their higher tiers of usage. I used to also have this problem and it annoyed me greatly, but once I got to usage tier 4 it never happened again (except for o1-preview but that just wastes tokens IMO).
If I'm reading this right; it's been in VSCode as Copilot Chat for a fair bit now. I use it often, when they added context (provide extra files to reference or even the entire @workspace if it's small enough), absolute gamechanger.
LLM as a service is much easier to replicate than physical data centers and there's a much lower potential user base than consumers, so I'd imagine they're swimming upstream into B2C land in order to justify the valuation
Aren't we talking about, say, GitHub Copilot? That's integrated into Visual Studio/VSCode. I just started using it again as they've done some small upgrades, and the results can often be phenomenal. Like, I will visualize an entire block of code in my mind, and I'll type the first couple of characters and the entire block will just appear. I'm literally that predictable.
Copilot is only using GPT3.5 for most of the results though, seemingly. I'd be more excited if they would update the API they're using.
> Honestly I’m scratching my head on OpenAI’s desire to double down on building out their consumer B2C use cases rather than truly focussing on being the infrastructure/API provider for other services to plug into
I think it's because LLMs (and to some extent other modalities) tend to be "winner takes all." OpenAI doesn't have a long term moat, their data and architecture is not wildly better than xAI, Google, MS, Meta, etc.
If they don't secure their position as #1 Chatbot I think they will eventually become #2, then #3, etc.
I suspect they are building their B2C products because it gives them better data to train on. It's a lot harder to control the quality of data when you have no idea how API inputs were produced, what the UI is like, or who the users are. You don't know the provenance of the data, or the context. Or even if multiple unrelated client products are being commingled through the same key.
If you control the UI, you have none of those problems.
> demonstrates the type of things Google could do with Gemini integrated into Google Docs
Or Microsoft!
> think OpenAI will end up being either an infrastructure provider OR a SaaS, but not both
Microsoft cut off OpenAI's ability to execute on the former by making Azure their exclusive cloud partner. Being an infrastructure provider with zero metal is doable, but it leaves obvious room for a competitor to optimise.
To be honest I think they’re having less success than it appears with their B2B offerings. A lot of cloud providers services like AWS have their own things they sell through those channels and I think a lot of businesses are finding those solutions to be cheaper and “good enough”
> but I wish it were integrated into tools already used for coding
Unless I'm missing something about Canvas, gh CoPilot Chat (which is basically ChatGPT?) integrates inline into IntelliJ. Start a chat from line numbers and it provides a diff before applying or refining.
Yea, I'm wondering the same. Is there any good resource to look up whether copilot follows the ChatGPT updates? I would be renewing my subscription, but it does not feel like it has improved similarly to how the new models have...
I think this is already built into Microsoft's Office365 "CoPilot" (which I assume is a ChatGPT frontend. You can ask the AI to make changes to your Office documents.
But my subscription at $20/mo is a fraction of my API usage at $5/day (about $100/mo).
You can sell a lot more GPT services through a higher bandwidth channel — and OpenAI doesn’t give me a way to reach the same bandwidth through their user interface.
I only use Gemini in Colab perhaps 5% of the times I use Colab, yet it is nice to have.
I use Gemini, OpenAI, Claude, smaller models in Grok, and run small models locally using Ollama. I am getting to the point where I am thinking I would be better off choosing one (or two.)
I'm really happy to see ChatGPT doing this. The idea of a canvas made me really enjoy using Claude as I felt it to be the (so far) most "appropriate interface" to AI Chatbots as you are often doing two unique things in an AI chat:
- holding in your mind a "thing" (i.e. some code)
- talking about a "thing" (i.e. walking through the code)
The same applies to non-code tasks as well. The ability to segregate the actual "meat" from the discussion is an excellent interface improvement for chatbots.
Why do you ask? I did use "4o with canvas" shortly after writing the above. To be clear, my original comment was not about the actual OpenAI implementation of a "canvas style chatbot", but rather that I have found the canvas-chat-UX to be the most effective way to interact with a chatbot. I am basing this on many hours with both ChatGPT (non-canvas until today) and Claude/Anthropic (who has had this feature). As of this writing, I prefer Claude both because of (a) the content of its output and (b) the canvas style, which allows my brain to easily parse what is the topic vs. discussion about the topic.
Not the op, but I just tried it. I agree with his point that this is a huge step up in having the discussion and then a separate canvas where the work takes place, and then iterations are done to the canvas.
I LOVE the UX animation effect ChatGPT added to show the canvas being updated (even if it really is just for show).
I don't really don't understand the appeal iterating on code in the browser away from the rest of the code base outside of cool demos. You really need to give the model context about your data model, code conventions, library choices, etc and be able to exercise it with tests to get to an acceptable result.
I use ChatGPT/Claude in bed and when I otherwise don’t have a computer handy like going on walks. When developing isolated code it works just fine without having to specify a lot of detail. For example, I’m working on a Qt QML app backed by Rust but use the mobile interfaces to code up C++ components that are independent of the Rust logic and can be wired together from QML/Javascript. Since the LLMs already have lots of data on how that API works, I don’t need to feed it much info about my project. It actually helps enforce code boundaries.
That's funny, I do the exact same thing. If I'm out and about or otherwise not at a computer, and I have an idea or solution to a bug I wanted to solve, I'll fire up an LLM on my phone (usually Claude these days) and basically brainstorm together the pros and cons, with code sometimes as well.
it's useful for my tampermonkey scripts that dont even exist in source control anywhere, and i have a friend that works in physics and has to write a ton of python code, but in really esoteric software platforms that are definitely not IDEs. they'd be copying and pasting code from a tool somewhere anyway.
Using Canvas (OpenAI) to design graphics on the HTML5 Canvas for my project in Canvas (the Learning Management System). I’m planning to print my work via CanvasPop and advertise it using Facebook Canvas ads.
Well as most of us know, Naming Things is one of the last 2 hard problems in computer science ;) [0] Tangentially related: I just found out yesterday that Apple has a piece of software called "Apple Devices" [2] which acts as a successor to iTunes for backup, sync, transfers, etc. Using traditional search engines to find information about my use case was almost as painful as searching for good information about working with an IBM i server...
Not to mention the HTML Canvas, it's a really bad name, it's even a bit worrying a company about intelligence is this bad at naming, I mean CanvasGPT was right there for crying out loud.
Awesome improvements, but compared to Claude Artifacts, it lacks the html/js "Preview" where you can run the code and check/validate the result without leaving the browser. This is a killer feature
Mixed feelings about this: clearly this is meant to match one of the killer features of Claude. I like using Claude, and I'm also a big supporter of Anthropic - not just because it's an underdog, but due to its responsible and ethical corporate governance model[1], which stands in stark contrast to OpenAI. It's worrying to see ChatGPT close one of the gaps between it and Claude.
I wonder how Paul Graham thinks of Sam Altman basically copying Cursor and potentially every upstream AI company out of YC, maybe as soon as they launch on demo day.
It's just a company that promised AGI would somehow come from developing LLM-based products, rapidly scrambling to keep up with other LLM-based products, to distract from the fact that it's becoming increasingly apparent that AGI is not coming anytime soon.
Cursor was one of the first AI editors I used, but recently Aider has completely replaced the AI assisted coding for me. I still use cursor but just as an editor, all LLM work is done with aider in the shell.
Like Amazon cloning the best selling products, bringing them in house, and then closing the accounts of competitors.
Met a guy who got brought in by Amazon after they hit 8 figures in sales, wined and dined, then months later Amazon launched competing product and locked them out of their accounts, cost them 9 figures.
As much as I want to like Claude, it sucks in comparison to ChatGPT in every way I've tested, and I'm going to use the better product. As a consumer, the governance model only results in an inferior product that produces way more refusals for basic tasks.
I code and document code and imho Claude is superior, try to tell Gpt to draw a mermaid chart to explain a code flow... the mermaid generated will have syntax errors half of the time.
Code output from is Claude pretty good. It seems to hallucinate less than o1 for me. It's been a struggle to get o1 to stop referencing non-existent methods and functions.
I think you can enable Artifacts, which are similar to OpenAI Canvas. Recently, Anthropic also added the ability to select elements within the created Artifact and adjust them (e.g., adjust length, improve code), similar to what Canvas can do.
I have some bad experience about it. Asked it to help generate python code to make a vpn server with extra layers, but it refused. What in the dictatorship is this? ChatGPT on the other hand did it with no problems. Seems like Claude has a lot more censorship and restrictions for what I tested it.
Attempting to do any form of security work using LLM is nigh impossible without a few steps of nudging it out of its “while user is asking me to do bad things: say no” loop.
After a year of heavy LLM use I’ve found the utility limits, my usage has peaked, and I’m developing very restrictive use cases.
Beyond functioning as an interactive O’Reilly manual, LLM only save time if you never read the code they produce. Which is a short term win, but things will blow up eventually, as with all code, and now you’ve got a bigger problem than you started with.
OpenAI started the same, so we'll see. One thing I dislike is that Claude is even more "over safeguarded" then ChatGPT. It disallows even kind of reasonable questions about Ritalin bioavailability in different ways of administration.
> They're not wasting hours trying to "figure out" a solution
I am pretty sure that we don't have AGI that would figure our solutions to our problems (coding or not) on its own yet. And from experience, you would need to solve the problems at least conceptually before using LLM and try to get something useful out of that.
That's probably the most FOMO driven landing page I've seen so far! I'll bite:
Your competition is using AI to build their product right now.
They're not writing code by hand.
They're not Googling code answers on StackOverflow.
They're not wasting hours trying to "figure out" a solution.
Extraordinary claims require some investigation!
Let's see how far I get without "Googling answers".
Spoiler: If the competition uses such AI that, good for me!
(Note I don't want to pick on Parrot here; from my first 5 minutes with the tool, it's /fun/, has a nice UI, a good idea for iteration, and somebody has to build such tools to make them work eventually even if they don't work great today yet. I wish it lots of success. But I want to also point out examples of the practical limitations of all such tools, and state that I highly doubt that they today can _in any way_ replace a human programmer who knows what they are doing, and has the awesome skill to use Google, StackOverflow, and "figure out" things.)
Prompt:
I need a Haskell function that hashes a given file (FilePath)
with the blake3 hash function.
It is important that it works on large files (e.g. 100 GB)
without freezing the Haskell runtime.
Output:
import qualified Crypto.Hash.BLAKE3 as BLAKE3
import qualified Data.ByteString as BS
import System.IO (withFile, IOMode(ReadMode))
hashFileBLAKE3 :: FilePath -> IO BS.ByteString
hashFileBLAKE3 filePath = do
let chunkSize = 8192 -- 8KB chunks
withFile filePath ReadMode $ \handle -> do
let loop ctx = do
chunk <- BS.hGet handle chunkSize
if BS.null chunk
then return $ BLAKE3.finalize ctx
else loop $! BLAKE3.update ctx chunk
loop BLAKE3.init
Explanation:
This Haskell function, `hashFileBLAKE3`, calculates the BLAKE3 hash of a
given file. It's designed to work efficiently with large files without
freezing the Haskell runtime. Here's a step-by-step explanation of how it
works:
1. We import necessary modules: `Crypto.Hash.BLAKE3` for the BLAKE3 hash
function, `Data.ByteString` for efficient byte handling, and `System.IO`
for file operations.
2. The function takes a `FilePath` as input and returns an `IO ByteString`
containing the hash.
3. We define a `chunkSize` of 8192 bytes (8KB) for reading the file in
manageable chunks.
4. The `withFile` function is used to safely open and close the file,
ensuring proper resource management.
5. Inside the file handling function, we define a recursive `loop` that
reads the file in chunks:
- It reads a chunk of the file using `BS.hGet`.
- If the chunk is empty (end of file), it finalizes and returns the
hash.
- Otherwise, it updates the hash context with the new chunk and
continues the loop.
6. The loop starts with `BLAKE3.init` to create the initial hash context.
7. By using lazy evaluation (`$!`), we ensure that the hash context is
strictly evaluated in each iteration, preventing memory buildup.
This implementation is memory-efficient and can handle very large files
without issues, as it processes the file in small chunks rather than
loading it entirely into memory.
My remark:
* This solution is already pretty good.
* The explanation is also useful for people new to Haskell
or this type of task.
* Some explanations are subtly wrong, e.g. `$!` is the
_opposite_ of lazy evaluation, and the sentence
is contradicting itself.
* It mainly got the imported module name wrong
(should be `import BLAKE3` from
https://hackage.haskell.org/package/blake3).
Let's see if it fixes it when I tell it what I'd observe
when trying to compile the code.
A simple Google search for "haskell blake3" should point out the correct name.
Corrective prompt:
The mentioned `Crypto.Hash.BLAKE3` module doesn't exist as a Haskell library.
Output:
import Crypto.Hash (hash, Blake2b_512, Digest)
import qualified Data.ByteString as BS
import qualified Data.ByteString.Lazy as LBS
import System.IO (withFile, IOMode(ReadMode))
hashFileBLAKE2b :: FilePath -> IO (Digest Blake2b_512)
hashFileBLAKE2b filePath = do
let chunkSize = 8192 -- 8KB chunks
withFile filePath ReadMode $ \handle -> do
let loop acc = do
chunk <- BS.hGet handle chunkSize
if BS.null chunk
then return $ hash (LBS.fromChunks acc)
else loop (chunk : acc)
loop []
My observation:
* I asked for BLAKE3, but now it implemented BLAKE2B.
* The solution is very wrong.
It splits the file into _reversed_ chunks and then hashes them.
The error is subtle and will be hard to debug.
So I'm quite sure I wouldn't allow anybody to build my "product" with such tools yet.
But it certainly is fun and there's definitely some usefulness in there, especially for learning general concepts that are well understood.
I guess in the end it's this: Programmers don't need to "roughly get it right" (which is what current tools do well) when it comes to products. They need to get it /precisely/ right.
They all obey the same masters, be it the big tech companies providing subsidized cloud, VC, or the stock market (post-IPO).
Trying to delude oneself that company A is superior morally to company B without a very clear distinction between incentive structures (eg A makes money from causing pollution, B sells widgets for cleaning up pollution), which is not the case with these companies, is magical thinking.
Ironically for these products, the amount of obviously llm generated bot comments astroturfing on HN and making it to the top of these corporate shill posts really aren't attracting me to using any of them. Read a bunch of glassdoor or google map reviews and then read these comments, clear as the water in a glacial lake; if the comments aren't fake, maybe it's just the people working for them that are? If you produce enough fake slop to be indistinguishable from a machine slop generator, does the difference even matter anymore? Are you still human?
If I understand your point, counterargument is that hn comments are made by people who are almost certainly in top 20% by intellect from general population.
Compared to some subreddits and YouTube, comments here are made by geniuses.
Comments on HN are made by all sorts of people, some are geniuses that contribute authentically to threads that interest them, and some others operate or invest in glorified template generators that they use to write propaganda about their own products in order to fake popularity and consensus. If one believes everyone else is a genius that uses whichever product, then whoever's not on that train is already obsolete. The nature and style of top comments on LLM related (last years news) threads often reads in such a predictible and distinguishable way from other arbitrary threads that it seems a bit... robotic. Same thing on Reddit.
This also demonstrates the type of things Google could do with Gemini integrated into Google Docs if they step up their game a bit.
Honestly I’m scratching my head on OpenAI’s desire to double down on building out their consumer B2C use cases rather than truly focussing on being the infrastructure/API provider for other services to plug into. If I had to make a prediction, I think OpenAI will end up being either an infrastructure provider OR a SaaS, but not both, in the long-term (5-10 yrs from now).
We are getting front row seats to an object lesson in “absolute power corrupts absolutely”, and I am relieved they have a host of strong competitors.
* Or which ever variant the average user might try to type in
Google already does have this in Google Docs (and all their products)? You can ask it questions about the current doc, select a paragraph and ask click on "rewrite", things like that. Has helped me get over writer's block at least a couple of times. Similarly for making slides etc. (It requires the paid subscription if you want to use it from a personal account.)
https://support.google.com/docs/answer/13951448 shows some of it for Docs, and https://support.google.com/mail/answer/13447104 is the one for various Workspace products.
I agree with you on where OpenAI will/should sit in 5-10 years. However, I don't think them building the occasional tool like this is unwarranted, as it helps them show the direction companies could/should head with integration into other tools. Before Microsoft made hardware full time, they would occasionally produce something (or partner with brands) to show a new feature Windows supports as a way to tell the OEMs out there, "this is what we want you to do and the direction we'd like the PC to head." The UMPC[0] was one attempt at this which didn't take off. Intel also did something like this with the NUC[1]. I view what OpenAI is doing as a similar concept, but applied to software.
[0] https://en.wikipedia.org/wiki/Ultra-mobile_PC
[1] https://en.wikipedia.org/wiki/Next_Unit_of_Computing
OP is lamenting that Cursor and OpenAI chose to create new apps instead of integrating with (someone else’s) existing apps. But this is a result of a need to be always fully unblocked.
Also, owning the app opens up greater financial potential down the line…
But if they believe they're going to reach AGI, it makes no sense to pigeonhole themselves to the interface of ChatGPT. Seems like a pretty sensible decision to maintain both.
(1) https://www.tanayj.com/p/openai-and-anthropic-revenue-breakd...
It wouldn't surprise me if not a lot of enterprises are going through OpenAI's enterprise agreements - most already have a relationship with Microsoft in one capacity or another so going through Azure just seems like the lowest friction way to get access. If how many millions we spend on tokens through Azure to OpenAI is any indication of what other orgs are doing, I would expect consumer's $20/month to be a drop in the bucket.
I've been using it for about a year.
This is exactly what Google’s NotebookLM does. It’s (currently) free and it reads your Google Docs and does RAG on them.
https://notebooklm.google/
Same as it ever was.
You only need 45 days to get tier 5 and if you have that many customers after 45 days you should just apply to YC lol.
Maybe you checked over a year ago, which was the wild wild West at the time, they didn't even have the tier limits.
I’m firmly in the camp that their rate limits are entirely reasonable.
Copilot is only using GPT3.5 for most of the results though, seemingly. I'd be more excited if they would update the API they're using.
I think it's because LLMs (and to some extent other modalities) tend to be "winner takes all." OpenAI doesn't have a long term moat, their data and architecture is not wildly better than xAI, Google, MS, Meta, etc.
If they don't secure their position as #1 Chatbot I think they will eventually become #2, then #3, etc.
But can they do it at all? It's not like they are like early Google vs other search engines.
If you control the UI, you have none of those problems.
Or Microsoft!
> think OpenAI will end up being either an infrastructure provider OR a SaaS, but not both
Microsoft cut off OpenAI's ability to execute on the former by making Azure their exclusive cloud partner. Being an infrastructure provider with zero metal is doable, but it leaves obvious room for a competitor to optimise.
Take a look at cursor.com
Unless I'm missing something about Canvas, gh CoPilot Chat (which is basically ChatGPT?) integrates inline into IntelliJ. Start a chat from line numbers and it provides a diff before applying or refining.
Yea, I'm wondering the same. Is there any good resource to look up whether copilot follows the ChatGPT updates? I would be renewing my subscription, but it does not feel like it has improved similarly to how the new models have...
Professionals instead don't love to change the tools once they got used to it for small incremental gains.
You can sell a lot more GPT services through a higher bandwidth channel — and OpenAI doesn’t give me a way to reach the same bandwidth through their user interface.
I use Gemini, OpenAI, Claude, smaller models in Grok, and run small models locally using Ollama. I am getting to the point where I am thinking I would be better off choosing one (or two.)
- holding in your mind a "thing" (i.e. some code)
- talking about a "thing" (i.e. walking through the code)
The same applies to non-code tasks as well. The ability to segregate the actual "meat" from the discussion is an excellent interface improvement for chatbots.
I LOVE the UX animation effect ChatGPT added to show the canvas being updated (even if it really is just for show).
Here's my user test so you know I actually used it. My jaw begins to drop around minute 7: https://news.pub/?try=https://www.youtube.com/embed/jx9LVsry...
Otherwise it’s mostly Cursor and Aider.
Oh I need to grab all the products in the database and calculate how many projects they were a part of.
I'm already using ChatGPT to do this because it turns what used to be a half day task into a 1 hour one.
This will presumably speed it up more.
[0] https://martinfowler.com/bliki/TwoHardThings.html
[II] https://support.apple.com/guide/devices-windows/welcome/wind...
"Name it clay" -- artistic CMO
"Won't people think they will have to get their hands dirty?" -- CEO
"Right. Name it sculpt. It has a sense of je ne sais quoi about it." -- hipster CMO
"No one can spell sculpt, and that French does not mean what you think it means." -- CFO
"Got it! Name it canvas! It's a blank canvas! It can be anything!" -- CMO
"Perfect!" -- Entire executive team who is now sick of CMO's naming excercise.
- Siv
- Forge
- Crucible
- Quench
- Workbench
- Calipers
- Lathe
[1] https://www.anthropic.com/news/the-long-term-benefit-trust
I wonder how Paul Graham thinks of Sam Altman basically copying Cursor and potentially every upstream AI company out of YC, maybe as soon as they launch on demo day.
Is it a retribution arc?
If OpenAI can copy Cursor, so can everyone else.
Met a guy who got brought in by Amazon after they hit 8 figures in sales, wined and dined, then months later Amazon launched competing product and locked them out of their accounts, cost them 9 figures.
You mean downstream.
Claude's been far and away superior on coding tasks. What have you been testing for?
Projector make it even better. But I could imagine it depends on the specific needs one has.
ChatGPT will just start to pretend like some perfect library that doesn't exist exists.
Deleted Comment
where does Claude have a canvas like interface?
I'm only seeing https://claude.ai/chat and I would love to know.
[0] https://support.anthropic.com/en/articles/9487310-what-are-a...
Canvas appears to be different in that it allows inline editing and also prompting on a selection. So not the same as Claude.
Deleted Comment
Deleted Comment
After a year of heavy LLM use I’ve found the utility limits, my usage has peaked, and I’m developing very restrictive use cases.
Beyond functioning as an interactive O’Reilly manual, LLM only save time if you never read the code they produce. Which is a short term win, but things will blow up eventually, as with all code, and now you’ve got a bigger problem than you started with.
[1] https://codewithparrot.com
> They're not wasting hours trying to "figure out" a solution
I am pretty sure that we don't have AGI that would figure our solutions to our problems (coding or not) on its own yet. And from experience, you would need to solve the problems at least conceptually before using LLM and try to get something useful out of that.
Let's see how far I get without "Googling answers".
Spoiler: If the competition uses such AI that, good for me!
(Note I don't want to pick on Parrot here; from my first 5 minutes with the tool, it's /fun/, has a nice UI, a good idea for iteration, and somebody has to build such tools to make them work eventually even if they don't work great today yet. I wish it lots of success. But I want to also point out examples of the practical limitations of all such tools, and state that I highly doubt that they today can _in any way_ replace a human programmer who knows what they are doing, and has the awesome skill to use Google, StackOverflow, and "figure out" things.)
Prompt:
Output: Explanation: My remark: Corrective prompt: Output: My observation: So I'm quite sure I wouldn't allow anybody to build my "product" with such tools yet.But it certainly is fun and there's definitely some usefulness in there, especially for learning general concepts that are well understood.
I guess in the end it's this: Programmers don't need to "roughly get it right" (which is what current tools do well) when it comes to products. They need to get it /precisely/ right.
Trying to delude oneself that company A is superior morally to company B without a very clear distinction between incentive structures (eg A makes money from causing pollution, B sells widgets for cleaning up pollution), which is not the case with these companies, is magical thinking.
Dead Comment