Readit News logoReadit News
captainregex commented on Llama-Scan: Convert PDFs to Text W Local LLMs   github.com/ngafar/llama-s... · Posted by u/nawazgafar
deepsquirrelnet · 6 days ago
Give the nanonets-ocr-s model a try. It’s a fine tune of Qwen 2.5 vl which I’ve had good success with for markdown and latex with image captioning. It uses a simple tagging scheme for page numbers, captions and tables.
captainregex · 6 days ago
I desperately wanted Qwen vl to work but it just unleashes rambling hallucinations off basic screencaps. going to try nanonet!
captainregex commented on GPT-OSS-120B runs on just 8GB VRAM & 64GB+ system RAM   old.reddit.com/r/LocalLLa... · Posted by u/zigzag312
tyfon · 13 days ago
I have a 5950x with 128 gb ram and a 12 gb 3060 gpu. The speed of generating tokens is excellent, the killer is that when the context grows even a little processing of it is super slow. Hopefully someone smart will optimize this, but as it is now I keep using other models like qwen, mistral and gemma.
captainregex · 13 days ago
What are you aiming to do with these models that isn’t chat/text manipulation?
captainregex commented on Linear sent me down a local-first rabbit hole   bytemash.net/posts/i-went... · Posted by u/jcusch
incorrecthorse · 16 days ago
> For the uninitiated, Linear is a project management tool that feels impossibly fast. Click an issue, it opens instantly. Update a status and watch in a second browser, it updates almost as fast as the source. No loading states, no page refreshes - just instant, interactions.

How garbage the web has become for a low-latency click action being qualified as "impossibly fast". This is ridiculous.

captainregex · 16 days ago
one of my day to day responsibilities involves using a portal tied to MSFT dynamics on the back end and it is the laggiest and most terrible experience ever. we used to have java apps that ran locally and then moved to this in the name of cloud migration and it feels like it was designed by someone whose product knowledge was limited to the first 2/5 lessons in a free Coursera (RIP) module
captainregex commented on Linear sent me down a local-first rabbit hole   bytemash.net/posts/i-went... · Posted by u/jcusch
sturza · 16 days ago
Local-first buys you instant UX by moving state to the client, and then makes everything else a little harder
captainregex · 16 days ago
this is such a clean and articulate way of putting it. The discussion around here the last few days about local and the role it is going to play has been phenomenal and really genuine
captainregex commented on US reportedly forcing TSMC to buy 49% stake in Intel to secure tariff relief   notebookcheck.net/Despera... · Posted by u/voxadam
cherryteastain · 19 days ago
Tariffs kinda make sense when you have a deficit in a widely available item. Big trade deficit with Bangladesh? Sure you can buy cheap textiles from Thailand or Vietnam or something.

Unfortunately this approach does not work when you lack a viable domestic alternative and you're up against a monopoly.

What will the US do if TSMC does not blink? Not buy TSMC made chips? Obviously that is impossible, so the logical conclusion is that American consumers will end up paying the tariffs.

captainregex · 18 days ago
more likely than not I think this ends with a vague promise, a loudly declared victory, and a quiet defanging of the promise or just outright ignoring it in the future
captainregex commented on Ollama Turbo   ollama.com/turbo... · Posted by u/amram_art
janalsncm · 19 days ago
The problem is that OSS is free to use but it is not free to create or maintain. If you want it to remain free to use and also up to date, Ollama will need someone to address issues on GitHub. Usually people want to be paid money for that.
captainregex · 19 days ago
money is great! I like money! but if this is their version of buy me a coffee I think there’s room to run elsewhere for their skillset/area of expertise
captainregex commented on Ollama Turbo   ollama.com/turbo... · Posted by u/amram_art
captainregex · 19 days ago
I am so so so confused as to why Ollama of all companies did this other than an emblematic stab at making money-perhaps to appease someone putting pressure on them to do so. Their stuff does a wonderful job of enabling local for those who want it. So many things to explore there but instead they stand up yet another cloud thing? Love Ollama and hope it stays awesome
captainregex commented on Open models by OpenAI   openai.com/open-models/... · Posted by u/lackoftactics
wizee · 19 days ago
Privacy, both personal and for corporate data protection is a major reason. Unlimited usage, allowing offline use, supporting open source, not worrying about a good model being taken down/discontinued or changed, and the freedom to use uncensored models or model fine tunes are other benefits (though this OpenAI model is super-censored - “safe”).

I don’t have much experience with local vision models, but for text questions the latest local models are quite good. I’ve been using Qwen 3 Coder 30B-A3B a lot to analyze code locally and it has been great. While not as good as the latest big cloud models, it’s roughly on par with SOTA cloud models from late last year in my usage. I also run Qwen 3 235B-A22B 2507 Instruct on my home server, and it’s great, roughly on par with Claude 4 Sonnet in my usage (but slow of course running on my DDR4-equipped server with no GPU).

captainregex · 19 days ago
I do think Devs are one of the genuine users of local into the future. No price hikes or random caps dropped in the middle of the night and in many instances I think local agentic coding is going to be faster than the cloud. It’s a great use case
captainregex commented on Open models by OpenAI   openai.com/open-models/... · Posted by u/lackoftactics
JKCalhoun · 19 days ago
I do it because 1) I am fascinated that I can and 2) at some point the online models will be enshitified — and I can then permanently fall back on my last good local version.
captainregex · 19 days ago
love the first and am sad you’re going to be right about the second
captainregex commented on Open models by OpenAI   openai.com/open-models/... · Posted by u/lackoftactics
canvascritic · 19 days ago
Healthcare organizations that can't (easily) send data over the wire while remaining in compliance

Organizations operating in high stakes environments

Organizations with restrictive IT policies

To name just a few -- well, the first two are special cases of the last one

RE your hallucination concerns: the issue is overly broad ambitions. Local LLMs are not general purpose -- if what you want is local ChatGPT, you will have a bad time. You should have a highly focused use case, like "classify this free text as A or B" or "clean this up to conform to this standard": this is the sweet spot for a local model

captainregex · 19 days ago
Aren’t there HIPPA compliant clouds? I thought Azure had an offer to that effect and I imagine that’s the type of place they’re doing a lot of things now. I’ve landed roughly where you have though- text stuff is fine but don’t ask it to interact with files/data you can’t copy paste into the box. If a user doesn’t care to go through the trouble to preserve privacy, and I think it’s fair to say a lot of people claim to care but their behavior doesn’t change, then I just don’t see it being a thing people bother with. Maybe something to use offline while on a plane? but even then I guess United will have Starlink soon so plane connectivity is gonna get better

u/captainregex

KarmaCake day60August 15, 2022View Original