Readit News logoReadit News
matthewolfe commented on Ex-Waymo engineers launch Bedrock Robotics to automate construction   techcrunch.com/2025/07/16... · Posted by u/boulos
ecshafer · a month ago
I think that more than physics the bottleneck for this is political (at least in the US). All of the local large projects around me are expensive because of massive amounts of red tape (environmental studies, zoning, planning), and political patronage systems. After the kick backs, political donations, promises to only work 8 hours a day, only use union labor, hire x police officers for y hours in overtime security positions a month, use xyz contractor etc. a small cost seems to be the actual labor and materials. Hell these robots if they work will be made illlegal.
matthewolfe · a month ago
I believe SchemeFlow [0] is working on solving some of these problem, particularly with the insane reporting requirements. But of course, that still leaves the unions...

[0] https://www.schemeflow.com/

matthewolfe commented on Show HN: Cactus – Ollama for Smartphones   github.com/cactus-compute... · Posted by u/HenryNdubuaku
matthewolfe · 2 months ago
For argument's sake, suppose we live in a world where many high-quality models can be run on-device. Is there any concern from companies/model developers about exposing their proprietary weights to the end user? It's generally not difficult to intercept traffic (weights) sent to and app, or just reverse the app itself.
matthewolfe commented on Show HN: TokenDagger – A tokenizer faster than OpenAI's Tiktoken   github.com/M4THYOU/TokenD... · Posted by u/matthewolfe
superlopuh · 2 months ago
Can someone familiar with performance of LLMs please tell me how important this is to the overall perf? I'm interested in looking into optimizing tokenizers, and have not yet run the measurements. I would have assumed that the cost is generally dominated by matmuls but am encouraged by the reception of this post in the comments.
matthewolfe · 2 months ago
To echo the other replies, the tokenizer is definitely not the bottleneck. It just happens to be the first step in inference, so it's what I did first.
matthewolfe commented on Show HN: TokenDagger – A tokenizer faster than OpenAI's Tiktoken   github.com/M4THYOU/TokenD... · Posted by u/matthewolfe
semiinfinitely · 2 months ago
I'm relieved to see that its not written in rust
matthewolfe · 2 months ago
haha, I thought about it.
matthewolfe commented on Show HN: TokenDagger – A tokenizer faster than OpenAI's Tiktoken   github.com/M4THYOU/TokenD... · Posted by u/matthewolfe
diggan · 2 months ago
Heh, seems people I've been learning from been biased away from beauty, as I know that as "Make It Work, Make It Right, Make It Fast".
matthewolfe · 2 months ago
Fair chance I'm remembering it wrong :D
matthewolfe commented on Show HN: TokenDagger – A tokenizer faster than OpenAI's Tiktoken   github.com/M4THYOU/TokenD... · Posted by u/matthewolfe
pama · 2 months ago
Cool. Would it be possible to eliminate that little vocab format conversion requirement for the vocab I see in the test against tiktoken? It would be nice to have a fully compatible drop in replacement without having to think about details. It also would be nice to have examples that work the other way around: initialize tiktoken as you normally would, including any specialized extension of standard tokenizers, and then use that initialized tokenizer to initialize a new tokendagger and test identity of results.
matthewolfe · 2 months ago
Alright, 0.1.1 should now be a true drop-in replacement. I'll write up some examples soon.

u/matthewolfe

KarmaCake day126February 27, 2021View Original