RIP to a legend
- Buy a small robot kit from Amazon or a local reseller. Yahboom make some good robot toy car kits. Hugging Face have the open source SO-ARM101 that plenty of companies manufacture and sell now. Expect to spend about $250 USD including a Jetson Nano for a good kit, up to $1000 USD if you want some more sensors
- If you can’t afford a real robot, play around with simulators like Isaac Sim and Mujoco
- Check out LeRobot, excellent framework for ML robotics from Hugging Face
- Learn the basics of ROS (pubsub), even if you don’t end up using it, a lot of the industry jargon and design patterns come from ROS so it helps to understand it. Think of ROS like Ruby on Rails, it’s a heavyweight batteries-included framework with lots of opinions.
- ROS does have some nice libraries for manipulation (MoveIt) and navigation (Nav2) using more classical (non-ML) methods
- Leverage AI tools such as ChatGPT and Cursor when you get stuck, it’s a lot faster than Googling when you’re just getting started and don’t even know the right term to search for.
- (Shameless plug) Check out two tools I’m working on: mcap.dev for logging and foxglove.dev for visualization
Don't bother with a Jetson Nano, you don't need that to get started, and by the time you need that you'll know a lot already. You can just drive the robot from your laptop!
Getting to training your own VLA fine-tuned model is a super quick and easy process. You can see examples of other people completing the tutorial and uploading their training/evaluation datasets here (shameless plug for my thing): https://app.destroyrobots.com
I wouldn't bother much with ROS at first tbh. It'll bog you down, and startups are moving toward using other approaches that are more developer friendly, like Rust-based embedded.
You can go far with a robot connected to USB though!
All that aside, the article doesn't really make much of an argument as to why 3 billion current users shouldn't be worth lots of money to someone wanting to try to monetize (even if the author doesn't see a good monetization opportunity). It, instead, focuses on why the Google integrations Chrome had are what made it popular. One of the biggest differences between Google selling Chrome and any old chromium fork is precisely that the "other" browsers no longer have to try to compete with Google's own browser to get users to monetize.
I worked at Mozilla when this deal was struck. The deal with Yahoo did require Yahoo be the default for Firefox, I'm not sure what you mean by "absence of any requirement"?
Mozilla broke that contract with Yahoo (there was a clause allowing them to do so without repercussion and keep the money, if they deemed it better for the users, wild contract) less than 3 years later because users hated Yahoo so much, and went back to Google.
Google is dominant because it just _is_ the best search engine.
> One of the biggest differences between Google selling Chrome and any old chromium fork is precisely that the "other" browsers no longer have to try to compete with Google's own browser to get users to monetize.
Isn't that literally anti-competitive? The DoJ is saying Google search is dominant partially because of Chrome pushing users to Google.
You're saying Chrome is dominant because users like it too much, and other browsers can't compete? Tough, that's the users' choice, though.
Deleted Comment
For example, so that you don't crush a human when doing massage (but still need to press hard), or apply the right amount of force (and finesse?) to skin a fish fillet without cutting the skin itself.
Practically in the near term, it's hard to sample from failure examples with videos on Youtube, such as when food spills out of the pot accidentally. Studying simple tasks through the happy path makes it hard to get the robot to figure out how to do something until it succeeds, which can appear even in relatively simple jobs like shuffling garbage.
With that said, I suppose a robot can be made to practice in real life after learning something from vision.
I'm not sure that's necessarily true for a lot of tasks.
A good way to measure this in your head is this:
"If you were given remote control of two robot arms, and just one camera to look through, how many different tasks do you think you could complete successfully?"
When you start thinking about it, you realize there are a lot of things you could do with just the arms and one camera, because you as a human have really good intuition about the world.
It therefore follows that robots should be able to learn with just RGB images too! Counterexamples would be things like grabbing an egg without crushing, perhaps. Though I suspect that could also be done with just vision.