Hi HN, sharing a product that I built since I didn't want to manual test our app.
The approach used is a multi-agent system that uses an augmented vision based approach for interacting with the UI, rather than relying on the DOM. The goal is to mimic how your users would naturally use the UI, rather than following a predefined path.
I built this because I got tired of doing manual QA for my last SaaS product. I couldn't trust automated UI tests (Selenium/Playwright) because they were flaky, broke whenever the DOM changed, and missed obvious usability bugs.
Happy to answer any questions about it here, and would love your feedback.
Thanks!