I would prefer a cli tool with partial gather support. Something that I could easily setup to run on a cheap instance somewhere and have it scrape all my data continuously at set intervals, and then give me the data in the most readable format possible through an easy access path. I've been thinking of making something like that, but with https://github.com/microsoft/graphrag at the center of it. A continuously rebuilt GraphRAG of all your data.
I seriously don't see any disadvantage in picking Firefox over Chrome. I still have Chrome around if any website requires it specifically, but I haven't launched it in ages.
There were a few Chrome extensions that weren't there on Firefox [1] [2] but I fixed that _easily_ by getting the crx file, unpacking it, then adding the https://github.com/mozilla/webextension-polyfill to the extension to make it cross-browser.
It's easy enough to make an extension work on both Firefox and Chrome, I've done it myself with SideHN (https://github.com/alin23/sidehn), but I guess Firefox is not really in the mind of Chrome extension devs.
[1] https://chromewebstore.google.com/detail/anchor-headings/lgg...
[2] https://chromewebstore.google.com/detail/xpath-helper/hgimno...
Is that actually true?