Readit News logoReadit News
Posted by u/jashmota 3 months ago
Launch HN: Flywheel (YC S25) – Waymo for Excavators
Hey HN, We're Jash and Mahimana, cofounders of Flywheel AI (https://useflywheel.ai). We’re building a remote teleop and autonomous stack for excavators.

Here's a video: https://www.youtube.com/watch?v=zCNmNm3lQGk.

Interfacing with existing excavators for enabling remote teleop (or autonomy) is hard. Unlike cars which use drive-by-wire technology, most of the millions of excavators are fully hydraulic machines. The joysticks are connected to a pilot hydraulic circuit, which proportionally moves the cylinders in the main hydraulic circuit which ultimately moves the excavator joints. This means excavators mostly do not have an electronic component to control the joints. We solve this by mechanically actuating the joysticks and pedals inside the excavators.

We do this with retrofits which work on any excavator model/make, enabling us to augment existing machines. By enabling remote teleoperation, we are able to increase site safety, productivity and also cost efficiency.

Teleoperation by the operators enables us to prepare training data for autonomy. In robotics, training data comprises observation and action. While images and videos are abundant on the internet, egocentric (PoV) observation and action data is extremely scarce, and it is this scarcity that is holding back scaling robot learning policies.

Flywheel solves this by preparing the training data coming from our remote teleop-enabled excavators which we have already deployed. And we do this with very minimal hardware setup and resources.

During our time in YC, we did 25-30 iterations of sensor stack and placement permutations/combinations, and model hyperparams variations. We called this “evolution of the physical form of our retrofit”. Eventually, we landed on our current evolution and have successfully been able to train some levels of autonomy with only a few hours of training data.

The big takeaway was how much more important data is than optimizing hyperparams of the model. So today, we’re open sourcing 100hrs of excavator dataset that we collected using Flywheel systems on real construction sites. This is in partnership with Frodobots.ai.

Dataset: https://huggingface.co/datasets/FlywheelAI/excavator-dataset

Machine/retrofit details:

  Volvo EC380 (38 ton excavator)
  4xcamera (25fps)
  25 hz expert operator’s action data
The dataset contains observation data from 4 cameras and operator's expert action data which can be used to train imitation learning models to run an excavator autonomously for the workflows in those demonstrations, like digging and dumping. We were able to train a small autonomy model for bucket pick and place on Kubota U17 from just 6-7 hours of data collected during YC.

We’re just getting started. We have good amounts of variations in daylight, weather, tasks, and would be adding more hours of data and also converting to lerobot format soon. We’re doing this so people like you and me can try out training models on real world data which is very, very hard to get.

So please checkout the dataset here and feel free to download and use however you like. We would love for people to do things with it! I’ll be around in the thread and look forward to comments and feedback from the community!

seabrookmx · 3 months ago
> The joysticks are connected to a pilot hydraulic circuit, which proportionally moves the cylinders in the main hydraulic circuit which ultimately moves the excavator joints

I've actually spent a decent amount of time running an excavator, as my Dad owns a construction / road building company. It was a great summer job!

An important note about the pilot hydraulics is that they _provide feedback to the operator_. I would encourage any system that moves these controls on behalf of a remote human operator or AI to add strain gauges or some other way to measure this force feedback so that this data isn't lost.

The handful of "drive by wire" pieces of equipment that my Dad or other skilled operators in my family have ran were universally panned, because the operators are isolated from this feedback and have a harder time telling when the machine is struggling or when their inputs are not sufficiently smooth. In the automotive world, skilled drivers have similar complaints about fully electronic steering or braking systems, as opposed to traditional vacuum or hydraulic boosting approaches where your foot still has a direct hydraulic connection to the brake pads.

jashmota · 3 months ago
You're right! This is exactly why we like to do mechanical actuation - we are able to achieve bilateral telepresence, which essentially gives the torque (haptic) feedback over the internet! So on small excavators, you can absolutely feel the resistance. We also stream the engine audio, which tells you how hard the hydraulic pump is working. Operators like our system for these reasons :)

I'd like to get a chance to talk to you and your Dad to get feedback. How do I reach you? My email is contact at useflywheel dot ai

IgorPartola · 3 months ago
Not my industry at all.

I am curious if something like this is an opportunity for a whole new type of controls and feedback. Since the operator doesn’t have to be in the excavator physically they could take on any position: standing, sitting, lying down, etc. Instead of sending haptic feedback to the joystick it could be sent to a vibrating wrist band. You could hook up the equivalent of a Nintendo Power Glove to have the scoop operated by the operator simulating scooping action. Turning the excavator can be controlled by the operator turning their head and moving it around can be done by the operator walking on an infinite treadmill. Motor strain can be done via color of light or temperature rather than sound. You could have a VR helmet that can also show you a birds eye view from a companion drone, overlay blueprints, show power and water lines, measure depth, etc. I don’t know if it is possible but maybe you could even measure soil composition somehow to show large rocks, soft soil that is dangerous to drive over, inclination angles where the excavator is about to drive, etc.

I imagine skilled operators prefer familiar controls but perhaps there are interesting improvements unlocked by exploring alternatives. It might also fundamentally change how accessible it is for non-professionals to use these machines. I rented an excavator from Home Depot a few years ago to dig a foundation and the learning curve was not shallow. I wonder if a more “natural” interface would help keep people safer.

roamerz · 3 months ago
I have a 20 ton Takeuchi and I don’t recall any feedback in the controls at all. The feedback I use is from the seat and sounds of the machine - well besides of course visual of course.

I cannot imagine this being useful to me unless the virtual operators cab closely mimicked an actual machine. It would have to have audio from the machine and be on a platform that tilted relative to the real thing. It would also need 270 degrees of monitors with a virtual mirror to see behind. On the front monitor, minimally, would need the to see vertically up and down too.

I also imagine all of this would be more useful to seasoned operators who can do most things on excavators in their sleep (definitely not me lol)

jashmota · 3 months ago
The way I think about this - we should not have multi screens. Human field of vision is 60 degrees for central and about 120 degrees binocular. The bucket of the excavator is way narrower than this which means actual task doesn't require wide vision.

So if we are able to have really good autonomous safety layers to ensure safe movements, and dynamically resize remote teleop windows, you'd make the operator more efficient. So while we stream 360 degree view, we get creative in how we show it.

That's on the vision side. We also stream engine audio, and do haptic feedback.

Takeuchi are interesting! Rare ones to have blades even on bigger sizes - is that why you got one?

opwieurposiu · 3 months ago
Yes, and you also get feedback from your butt as the machine tips and wobbles, particularly on smaller machines. Hearing the engine straining helps also. Often you can not clearly see what you are digging, this feedback lets you know if you are running into a rock or something.

One big advantage would be cameras mounted on the boom and rear view cameras, as many machines have obstructed views.

jashmota · 3 months ago
We're indeed streaming the audio and have haptic feedback. My hypothesis is the seat vibration isn't as helpful. It's sub optimal and operators would be far more productive without it. We would do a paper when we have enough data on this. We're also putting more cameras but streaming a lot of them at once is tricky.
jeffbee · 3 months ago
My car with its drive-by-wire brakes has a brake feedback simulator that gives the driver the kind of feeling associated with power-boosted hydraulic brakes. This is by far the most expensive single component in the car. Arguably these are just expensive accommodations for human flaws. A self-driving car wouldn't need them. Can't the self-driving system act directly on data like pressure, flow, and displacement?
01HNNWZ0MV43FF · 3 months ago
Maybe it doesn't matter for a car because feeling the car's motion tells you most of what you need to know. A car is not meant to touch anything but the road, in normal conditions. I think steering is the only case where force feedback is very important for a car - In the winters up here, I can feel the steering go loose when I hit a patch of ice.

I imagine an excavator, meant to touch and dig through things, and lift things, benefits from force feedback for the same reason VR would.

Have you played those VR sword games? BeatSaber works great because you're cutting through abstract blobs that offer no resistance. But the medieval sword-slashing games feel weird because your sword can't impact your opponent.

I saw a video recently of a quadcopter lifting heavy objects. When it's overloaded, it can't maneuver because all its spare power is spent generating lift to maintain altitude. If the controls had force feedback, the copter's computer could tell you "I'm overloaded, I can't move" by putting maximum resistance on the sticks.

jashmota · 3 months ago
That's indeed what we're trying to test to the extreme - to see how far we could go with just vision. We haven't done heavy excavation workflows yet, but we have some early success with some excavation workflows with just vision input and joystick action output (even without joint angle feedback!). We're betting on having really huge data with compact observation input and experiment to see if it holds water. If not we can always dial it down and add more sensors/feedback.
cyberax · 3 months ago
There are no drive-by-wire brakes in the US or Europe for regular cars. Your car's actuator moves the piston that is mechanically linked to your pedal.

So even if the electric system fails completely, you can still actuate the brakes.

toomuchtodo · 3 months ago
Would you be able to replicate this with the heavy equipment and movements needed to plug orphaned oil wells? In Texas alone, it's a TAM of ~$38B, and ~$150B for the entire US.

The Looming Disaster Under America's Biggest Oil Field [video] - https://news.ycombinator.com/item?id=45361022 - September 2025

Texas has thousands of abandoned oil and gas wells. Who is responsible for cleaning them up? - https://www.texastribune.org/2025/05/08/texas-orphan-wells-e... - May 8th, 2025

The Rising Cost of the Oil Industry’s Slow Death - https://www.propublica.org/article/the-rising-cost-of-the-oi... - February 22nd, 2024

Well plugging SOP:

https://www.epa.gov/natural-gas-star-program/well-plugging

https://www.osha.gov/etools/oil-and-gas/abandoning-well

jashmota · 3 months ago
It is possible. One good thing about our retrofit is it works on any hydraulic equipment, and not just excavators, although we are focussed on excavators for now. We are able to retrofit any hydraulic machinery because we actuate the joystics connected to the valves, which are quite standardized and usually come from handful of companies like Bosch. Will definitely 'dig' more into this though.
Onavo · 3 months ago
Your service will need to be cheaper than just lobbying the current federal government.
westurner · 3 months ago
Could a mini tunnel boring machine plug a well, from the side?
ceejayoz · 3 months ago
algo_trader · 3 months ago
> TAM of ~$38B, and ~$150B for the entire US.

Sigh. No. Actual TAM is $0 which is why these wells are orphaned

For contrast, maritime shipping IMO (my current obsession) has set a $100-$400 carbon price. So now the efuel market a realistic $10B ZNZ pot to aim at.

abraae · 3 months ago
What a fantastic business. I've watched lots of digger (as we call them here) drivers in action and I'm in awe of what these machines can do and how incredibly skilled their operators are. We have several large rock walls on our property and watching these guys delicately picking and placing huge rocks to key them in just blows my mind.

There was a YouTube video recently of an AI-assisted digger making a wall out of the concrete rubble from a demolition.

I believe the applications for really smart excavators must be huge. Sounds like this might be a step on that voyage.

jashmota · 3 months ago
You're spot on! The experienced operators are quite skilled at moving these machines. The number of operators who could do this work are decreasing and it's also pretty dangerous which explains the reduction of new people joining in, especially with better ways to make money today. We're getting started with remote teleop and moving towards full autonomy!
yreg · 3 months ago
Robot Excavator Builds Massive Stone Wall Without Human Assistance: https://www.youtube.com/watch?v=Belu_1pUPVk
jasongill · 3 months ago
This looks very interesting. I understand that hydraulic retrofit is compelling, but you may want to consider CANbus-based retrofits as well (similar to how CommaAI works, where it simply sends messages to the CANbus network to control steering or speed just like ADAS features do). Modern equipment is moving away from hydraulic controls toward a fully digital cockpit and being able to plug in and teleop (for less than the cost of the manufacturers expensive robotics suite) would be amazing.

I have a Cat 289D skid steer I would happily contribute to the effort if you guys move into the compact equipment space (compact being a relative term, as my machine is only 6 tons compared to your 38 ton machine)

jashmota · 3 months ago
Thanks! Yes we're testing both CAN spoofing and retrofits. CAN spoofing really limits compatibility to a few machines/brands. Would love to get in touch with you: my email is contact ai useflywheel dot ai
jasongill · 3 months ago
I disagree - your AI model should be decoupled from your control logic; so if the model determines that you need to move the boom, your control logic takes that action - either via hydraulic retrofit, canbus control, or just faking it in a simulator, and the AI model observes and reacts to the change as a result of that action.

So in reality, CANbus control should be no different than hydraulic control to your system - just another style of I/O

mschuster91 · 3 months ago
I'd never thought anyone even attempting such a feat, interesting that you went all the way to mechanically actuate levers. Nice job.

But... I assume you're intending to run on extremely large sites such as highway construction, open pit mines and the likes primarily? Because my experience (if a bit dated cough) is running small 750kg baby excavators under sometimes extreme space constraints - digging trenches for telco in urban and rural areas, which often enough meant having to work with 5-10 cm distance to walls, lighting posts or other infrastructure, and directly next to workers shoving soil into the bucket. Will you add stuff like 360° camera vision, LIDAR etc. to make that safe and help a remote/AI operator, or are you planning on large sites with less danger potential only?

trhway · 3 months ago
When i first time 35 years ago in Siberia saw an excavator driving itself toward me without driver, i was struck like by lightning and first moment attributed it to hangover. Next moment though i saw the driver walking along the road looking for mushrooms and understood that the deep tracks in the unpaved road are like rails so that the excavator just moves along these tracks. Couple years later (this time another part of the country - a low populated part of the St.Petersburg region) my friends who were vacationing in a remote village there saw one day a large tractor passing on its own through that village, and next day guys came asking whether anybody saw a tractor without driver - the guys were drinking and details of how the tractor got away weren't clear though ...
jashmota · 3 months ago
I can relate to how you felt. Even though I developed this myself, when I saw it running remotely with no one sitting inside for the first time, it felt magical and weird. This is going to save lives and look magical and that's too much fun to put off and do something else!
rapjr9 · 3 months ago
One potential application might be fighting wildfires. Seems like it would be really useful if one firefighter could remotely monitor a dozen autonomous bulldozers that were given general instructions. Would have to acquire a different training data set I'd guess, but the same approach seems applicable, get the data from teleoperated bulldozers used to fight fires. Getting the bulldozers into the right area would require some transport, like a truck or heavy lift helicopter, though maybe mini bulldozers would also work.

You might be interested in the work of Peter Corke also, he's automated horizontal mine shaft loaders and huge drag line shovels in his research:

https://www.youtube.com/watch?v=YUb9_Ysd2Hw

I think he used a different approach than you do, using visual servoing to get feedback and data from a camera. Maybe there's some value in combining both approaches, learn to control a machine from an operator, and also keep track of what is being moved with a camera to add another layer of control.

aorloff · 3 months ago
I would create a bespoke training set from the very best operator you can find, after significant interviewing, and then only use that to train your model (eg. partner with a local firm that rents at Herc or United Rentals a lot and has a skilled operator and train on their jobs while you rent them free equipment).

The range of skillset on these things is large, there are 2 dominant (fairly swappable - meaning most machines do both) different operating modes (excavator / backhoe) and I see that operators have a specific one they are best with.

Honestly, when I see a real pro using one of these machines I think this is one area AI is not going to win at soon - in the real world there's a whole support crew working with the excavator operator

jashmota · 3 months ago
This is a great idea - I would dig into this. I disagree on AI not winning here. Support crew mostly work on eliminating blindspots for operator. Yes you can eliminate blindspots with cameras and a 7inch TFT screen, but those are small, don't work great in direct sunlight. The current solution not being 10x better is why they still fallback to having those crew members. I genuinely believe we will change that.
aorloff · 3 months ago
So if you just built the blind spot detection for the rig and offered that it probably would be enough product