Readit News logoReadit News
mark_l_watson · 10 months ago
After just spending 15 minutes trying to get something useful accomplished, anything useful at all, with latest beta Apple Intelligence with a M1 iPad Pro (16G RAM), this article appealed to me!

I have been running the 32B parameters qwen2.5-coder model on my 32G M2 Mac and and it is a huge help with coding.

The llama3.3-vision model does a great job processing screen shots. Small models like smollm2:latest can process a lot of text locally, very fast.

Open source front ends like Open WebUI are improving rapidly.

All the tools are lining up for do it yourself local AI.

The only commercial vendor right now that I think is doing a fairly good job at an integrated AI workflow is Google. Last month I had all my email directed to my gmail account, and the Gemini Advanced web app did a really good job integrating email, calendar, and google docs. Job well done. That said, I am back to using ProtonMail and trying to build local AIs for my workflows.

I am writing a book on the topic of local, personal, and private AIs.

tracerbulletx · 10 months ago
I wrote a script to queue and manage running llama vision on all my images and writing the results to an sqlite db used by my Media Viewer, and now I can do text or vector search on it. It's cool to not have to rely on Apple or Google to index my images and obfuscate how they're doing it from me. Next I'm going to work on a pipeline for doing more complex things like multiple frames in a video, doing multiple passes with llama vision or other models to separate out the OCR, description, and object, people recognition. Eventually I want to feed all of this in here https://lowkeyviewer.com/ and have the ability to manually curate the automated classifications and text.
Eisenstein · 10 months ago
I'm curious why you find descriptions of images useful for searching. I developed a similar flow and ended up embedding keywords into the image metadata instead. It makes them easily searchable and not tied to any databases, and it is faster (dealing with tens of thousands of images personally).

* https://github.com/jabberjabberjabber/LLavaImageTagger

mark_l_watson · 10 months ago
nice!
mark_l_watson · 10 months ago
Another thought: OpenAI has done a good enough job productizing ChatGPT with advanced voice mode and now also integrated web search. I don’t know if I would trust OpenAI with access to my Apple iCloud data, Google data, my private GitHub repositories, etc., but given their history of effective productization, they could be a multi-OS/platform contender.

Still, I would really prefer everything running under my own control.

alexander2002 · 10 months ago
Who can trust a company whose name contradicts its presence.
zerop · 10 months ago
Have you tried RAG on Open WebUI. How does it do in asking questions from source docs?
mark_l_watson · 10 months ago
Not yet. It has ‘Knowledge sources’ that you can set up, and I think that supplies data for built in RAG - but I am not sure until I try it.
bboygravity · 10 months ago
can llama 3.3 vision do things like "there's a textbox/form field at location 1000, 800 with label "address"" ?

I did a quick and dirty prototype with Claud for this, but it returned everything with an offset and/or scaled.

Would be a killer app to be able to auto-fill any form using OCR.

MaxLeiter · 10 months ago
Were you using claude’s computer mode? It can do this
honestAbe22 · 10 months ago
Open source frontend??? Wtf. Show your code. This post is bs. AI coding has not advanced software in any meaningful way and will make coders dumber
JSDevOps · 10 months ago
Is anyone instantly suspicious when they introduce themselves these days an "AI Developer"
master-lincoln · 10 months ago
I would interpret that as somebody working on ML algorithms and architectures, not somebody developing a product that uses some form of AI at runtime...
JohnFen · 10 months ago
I work on such a team, and we don't refer to ourselves as "AI engineers". We're devs working on developing deep learning systems (but not genAI).
spmurrayzzz · 10 months ago
I haven't seen this sort of work called AI Developer yet, but I may have missed the trend shift. Isn't the convention for this still to use the title of Machine Learning Engineer (MLE)?
pjmlp · 10 months ago
Yes, just like "Data Engineer" for knowing how to use Tableau or doing OLAP queries.
Ldragon27 · 10 months ago
Are you saying that true Data Engineers typically do more than just use Tableau or run OLAP queries, or do you see the title 'Data Engineer' itself as a bit of a red flag these days? I’m pretty early in my career and was leaning toward Data Engineering, but hearing stuff like this makes me wonder if going for SWE might be smarter.
bubbleRefuge · 10 months ago
having data engineer in my title has led to recruiters calling about DBA type roles. ugh.
WorldWideWebb · 10 months ago
I feel the same way about “UX Designer/Engineer”. Seems to mean someone who can put a wireframe together but has no design chops. Any good designer that has experience crafting user interfaces should be skilled at thinking through in-depth about what the user is doing when using the product and how to successfully guide them through that process.
Taylor_OD · 10 months ago
Eh. Titles dont really mean anything. Do we want to start the age old software engineer vs software developer title debate again?

Let someone call themselves whatever they want. If they can do the job they were hired for then... who cares?

valval · 10 months ago
To be fair, the market dictates the job titles.
wrs · 10 months ago
I’m instantly suspicious when anyone uses the term “AI”.
kmmlng · 10 months ago
I used to be in the same camp, but have since abandoned this position. The world has changed. Very serious people use the term "AI" nowadays. I don't find that it's a high-precision signal for identifying charlatans anymore (though it is high-recall).
ocular-rockular · 10 months ago
You would be surprised how many people have no idea what "machine learning" means (not even technical definition but just as in the field). I'm working on a PhD in an adjacent field and basically have to tell people I work in AI for them to have any semblance of what I do.
bdangubic · 10 months ago
in 2024 thats as crazy as being suspicious of anyone that uses the term “internet”
0points · 10 months ago
Instant red flag. Like "Scrum master" used to be back in the day.
marcosdumay · 10 months ago
> back in the day

Well, most are called "project manager" now. But it would still be a giant red flag, just like the project manager job title or even worse, using PM so you don't know exactly what it means.

amelius · 10 months ago
"AI engineer" is the new "Web developer".
moffkalast · 10 months ago
I prefer the term "Scrumlord" /s

Deleted Comment

woodson · 10 months ago
I say “I work in machine learning, what kids these days call AI” ;).

Deleted Comment

noch · 10 months ago
> Is anyone instantly suspicious when they introduce themselves these days an "AI Developer"

I'm only suspicious if they don't simultaneously and eagerly show me their Github so that I can see what they've accomplished.

llm_nerd · 10 months ago
Of the great developers I have worked with in real life, across a large number of projects and workplaces, very few have any Github presence. Most don't even have LinkedIn. They usually don't have any online presence at all: No blog with regular updates. No Twitter presence full of hot takes.

Sometimes this industry is a lot like the "finance" industry: People struggling for credibility talk about it constantly, everywhere. They flex and bloviate and look for surrogates for accomplishments wherever they can be found. Peacocking on github, writing yet another tutorial on what tokens are and how embeddings work, etc.

That obviously doesn't mean in all cases, and there are loads of stellar talents that have a strong online presence. But by itself it is close to meaningless, and my experience is that it is usually a negative indicator.

PreInternet01 · 10 months ago
"after years of working in DevOps, MLOps, and now GenAI"

You truly know how to align yourself with hype cycles?

snovv_crash · 10 months ago
They missed out on drones and blockchain.
red-iron-pine · 10 months ago
drones may still be a thing, and weren't nearly as hyped as blockchain.

my barometer for penetration is how often the non-tech people talk about it, e.g. goofball uncle didn't buy a drone, but he went hard on BTC. if he's still holding he probably made money recently, too.

taway1874 · 10 months ago
... and Cloud! Don't forget The Klaaooud ...
satvikpendem · 10 months ago
Well, resume driven development does work, it seems.
cle · 10 months ago
Of course it does, your resume is a critical part of selling your skills & services to employers. Want to close faster and for more $$$? Demonstrate your value prop in the terms they know and care about.
JohnFen · 10 months ago
I don't want to be an "AI engineer" in the way the article means. There's nothing about that sort of job that I find interesting or exciting.

I hope there will still be room for devs in the future.

croes · 10 months ago
Is that really AI engineering or Software engineering with AI?

If a model goes sideways how do you fix that? Could you find and fix flaws in the base model?

sigmar · 10 months ago
Agree that the use of "AI engineers" is confusing. Think this blog should use the term "engineering software with AI-integration" which is different from "AI engineering" (creating/designing AI models) and different from "engineering with AI" (using AI to assist in engineering)
crimsoneer · 10 months ago
The term AI engineer is now pretty well recognised in the field (https://www.latent.space/p/ai-engineer), and is very much not the same as an AI researcher (which would be involved in training and building new models). I'd expect an AI engineer to be primarily a software developer, but with an excellent understanding of how to implement, use and evaluate LLMs in a production environment, including skills like evaluation and fine-tuning. This is not some dataset you can just bundle in software developer.
billmalarky · 10 months ago
You find issues when they surface during your actual use case (and by "smoke testing" around your real-world use case). You can often "fix" issues in the base model with additional training (supervised fine-tuning, reinforcement learning w/ DPO, etc).

There's a lot of tooling out there making this accessible to someone with a solid full-stack engineering background.

Training an LLM from scratch is a different beast, but that knowledge honestly isn't too practical for everyday engineers given even if you had the knowledge you wouldn't necessarily have the resources necessary to train a competitive model. Of course you could command a high salary working for the orgs who do have these resources! One caveat is there are orgs doing serious post-training even with unsupervised techniques to take a base model and reeaaaaaally bake in domain-specific knowledge/context. Honestly I wonder if even that is unaccessible to pull off. You get a lot of wiggle-room and margin for error when post-training a well-built base model because of transfer learning.

liampulles · 10 months ago
I wonder if either could be really be called engineering.
sourcepluck · 10 months ago
I feel like I see this comment fairly often these days, but nonetheless, perhaps we need to keep making it - the AI generated image there is so poor, and so off-putting. Does anyone like them? I am turned off whenever I see someone has used one on a post, with very few exceptions.

Is it just me? Why are people using them? I feel like objectively they look like fake garbage, but obviously that must be my subjective biases, because people keep using them.

RodgerTheGreat · 10 months ago
Some people have no taste, and lack the mental tools to recognize the flaws and shortcomings of GANN output. People who enthuse about the astoundingly enthralling literary skills of LLMs tend to be the kind of person who hasn't read many books. These are sad cases: an undeveloped palate confusing green food coloring and xylitol for a bite of an apple.

Some people can recognize these shortcomings and simply don't care. They are fundamentally nihilists for whom quantity itself is the only important quality.

Either way, these hero images are a convenient cue to stop reading: nothing of value will be found below.

nuancebydefault · 10 months ago
> these hero images are a convenient cue to stop reading.

If you don't like such content. But I would say don't judge a book by its cover.

vunderba · 10 months ago
That picture doesn't even have anything to do with the contents of the post either.

Reminds me of the image attached to Karpathy's (one of the founding members of openAI) twitter post on founding an education AI lab:

https://x.com/karpathy/status/1813263734707790301

I just don't understand how he didn't take 10 seconds to review the image before attaching it. If the image is emblematic of the power of AI, I wouldn't have a lot of faith in the aforementioned company.

If you're going to use GenAI (stable diffusion, flux) to generate an image, at least take the time to learn some basic photobashing skills, inpainting, etc.

aeturnum · 10 months ago
A trend I see in the "frothiest" parts of the AI world is an inattention to details and overwhelming excitement about things just over the horizon. LLMs are clearly a huge deal and will be changing a lot of things and also there are a bunch of folks wielding it like a blunt object without the discernment to notice that they're slightly off. I'm looking forward to the next couple of decades but I'm worried about the next five years.
rurp · 10 months ago
You aren't exaggerating! There are some creepy arms in that image, along with the other weirdness. I'm surprised Karpathy of all people used such a poor quality image for such a post.
sourcepluck · 10 months ago
Yeah, wow. And it's a lovely post otherwise, too!
GPerson · 10 months ago
An insidious feature of machine generated content is that it will dominate not because it’s better, but because it’s cheaper.
namaria · 10 months ago
The Spam sketch from Monty Python comes to mind.
nuancebydefault · 10 months ago
I don't find the image poor, but somehow I see immediately that it is generated because of the stylistic style. And that simply triggers the 'fake' flag in the back of my head, which has this bad subjective connotation. But objectively I believe it is a very nice picture.
nuancebydefault · 10 months ago
Edit: i looked at it more closely and the way the people are sitting, their direction, is totally unnatural.
colonelspace · 10 months ago
You mean to tell me this doesn't convincingly look like people working on their laptops on trestle tables in the forest at dusk?

Last time I worked on my laptop on a trestle table in the forest at dusk it looked almost exactly like this.

Deleted Comment

auggierose · 10 months ago
I think AI images can be very nice, I like to use them myself. I don't use images I don't personally like very much. So if you don't like them, it is not because, AI, it is because your taste and my taste don't match. Or maybe you would like them, if you didn't have a bias against AI. What I love about AI images is that you can often generate very much the thing you want. The only better alternative would be to hire an actual human to do that work, and the difference in price here is huge, of course.

It is like standing in front of a Zara, and wondering why people are in that shop, and not in the Versace shop across town. Surely, if you cannot afford Versace, you rather walk naked?

hpen · 10 months ago
Does AI engineer == API Engineer?
bormaj · 10 months ago
The P is silent
sdesol · 10 months ago
If the P is silent, how can you "Prompt"?
tejtm · 10 months ago
as it is in "swimming"
ein0p · 10 months ago
An AI engineer with some experience today can easily pull down 700K-1M TC a year at a bigtech. They must be unaware that the "barriers are coming down fast". In reality it's a full time job to just _keep up with research_. And another full time job to try and do something meaningful with it. So yeah, you can all be AI engineers, but don't expect an easy ride.
actusual · 10 months ago
I run an ML team in fintech, and am currently hiring. If a resumè came across my desk with this "skill set" I'd laugh my ass off. My job and my team's jobs are extremely stressful because we ship models that impact people's finances. If we mess up our customers lose their goddamn minds.

Most of the ML candidates I see now are all "working with LLMs". Most of the ML engineers I know in the industry who are actually shipping valuable models, are not.

Cool, you made a chatbot that annoys your users.

Let me know when you've shipped a fraud model that requires four 9's, 100ms latency, with 50,000 calls an hour, 80% recall and 50% precision.

btdmaster · 10 months ago
What does 50% precision mean in this case? I know 50% accuracy might mean P(fraud_predicted | fraud) = 50%, but I don't understand what you mean by precision?
ein0p · 10 months ago
I'd hope you're shipping _anti_-fraud models. Fraud models are abundant in fintech. On a side note, anyone who _really_ knows LLMs will be able to do your primitive fintech models with their eyes closed. You don't need a ton of skill to build a logistic regression model that runs at 14 qps. :-)