• Correlates syslogs with mcap/bag file anomalies automatically
• Flags when a hardware failure might have begun (not just when it manifests)
• Surfaces probable root causes instead of leaving teams to manually chase timestamps
From your experience across 50+ clients, which do you think is the bigger timesink: data triage across multiple logs/files or interpreting what the signals actually mean once you’ve found them?
Maybe there could be value in signal interpretation for purely software engineers but I reckon it would be hard for such team to build robots.
* Combing through the syslogs to find issues is an absolute nightmare, even more so if you are told that the machine broke at some point last night
* Even if you find the error, it's not necessarily when something broke; it could have happened way before, but you just discovered it because the system hit a state that required it
* If combing through syslog is hard, try rummaging through multiple mcap files by hand to see where a fault happened
* The hardware failing silently is a big PITA - this is especially true for things that read analog signals (think PLCs)
Many of the above issues can be solved with the right architecture or tooling, but often the teams I joined didn't have it, and lacked the capacity to develop it.
At Foxglove, we make it easy to aggregate and visualize the data and have some helper features (e.g., events, data loaders) that can speed up workflows. However, I would say that having good architecture, procedures, and an aligned team goes a long way in smoothing out troubleshooting, regardless of the tools.
Location: Prague, Czech Republic
Remote: Yes
Willing to relocate: No
Technologies: ROS, Python, C++, Robotics, Px4, Ardupilot, Pixhawk
Résumé/CV: https://www.linkedin.com/in/mateuszsadowski/
Email: mat[at]msadowski.ch
I have 11 years of experience making software for robots. For the past 7 years, I've been working as a consultant on various projects. I have experience building software for hardware platforms, from autonomous mobile robots through drones to heavy-metal industrial robots. These days, I have a slight preference for consulting projects but can consider a full-time position if there is a good match.
Do you think there are specific triage workflows where even a small automation (say, correlating error timestamps across syslog and bag files) would save meaningful time?
* I was setting up Ouster lidar to use gos time, don’t remember the details now but it was reporting the time ~32 seconds in the past (probably some leap seconds setting?)
* I had a ROS node misbehaving in some weird ways - it turned out there was a service call to insert something into db and for some reason the db started taking 5+ minutes to complete which wasn’t really appropriate for a blocking call
I think the timing is one thing that needs to be consistently done right on every platform. The other issues I came across were very application specific.