But in general we found the best course of action is simply label everything. Because our customers will want those answers and rag won’t really work at the scale of “all podcasts the last 6 months. What is the trend of sentiment Hillary Clinton and what about the top topics and entities mentioned nearby”. So we take a more “brute force” approach :-)
There is also more generally Mention, Brand24, Meltwater, etc, in the media monitoring space. But all are generally weak in the audio media space.
When we stop looking at audio media like a newspaper article, things get more interesting.
I used to work for a media monitoring company but was instantly struck by how old fashion everything was. And the total reliance on boolean searches meant only experts could find relevant information. This still appears to be the case for most players in the industry.
So I'm building a platform that finds what is important before you look for it. Novel entity linking, and sentiment analysis plus speaker tracking. It has come a long way from the proof of concept. Focused on audio media at the moment as it is the hardest to index compared to news articles in my opinion. And the hypothesis is audio media such as podcasts can contain so many juicy insights.
Next steps are converting pilot customers to paying customers, testing more markets (based in a tiny market now), raise a small pre-seed (bootstrapped at the moment), and quickly evolve the product based on feedback.
You can show up to 4 different times besides your local time.
This is really handy to see the time in UTC and other timezones where some of your team mates might exist. This saves having to Google for timezone converters.
I found it to reliably produce JSON correctly, but I've found 3.5 to be a poor performer at things like entity extraction and following directions compared to other fast models such as claude-instant (though that does not have function calling).
I can definitely argue that mirrorless is always better for most users! Thinner bodies, modern lens selection (so faster autofocus and wider aperture, etc), real-time image previews, better low-light preview (in both rear and viewfinder screens), video recording capabilities better match professional video cameras (by removing mirror complexity) which better matches the hybrid needs of the modern camera buyer.
Most of the work of accepting and holding concurrent TCP sockets done by the OS, not the language runtime. One can easily tune Linux kernel to 1M concurrent sockets.
The real issues: memory usage per concurrent socket (idle or active), and ability to do something useful with all these active connections, e.g. send pings every 30s, or broadcast a message to all of them.
I'm not sure NodeJS/C++ based system from this post will allow sending pings every 30s to 1M websockets, let alone to do something useful with them, beside some low-traffic or infrequent notifications (Of course one always need to perform realistic loadtests in order to answer these kinds of questions).
Erlang/Elixir/BEAM have a relatively large memory usage per active socket, but it allows doing something useful with them under an easy to use programming model (read: no callback hell).
It’s been a journey but getting close to launching our first version to pilot customers in August. We use an enormous amounts of AI tokens every month to extract data not possible with any traditional player in this media monitoring space. Benchmarking competitors, tracking impactful discussions, and receiving actionable brand insights.
If you are currently using one of the big media monitoring companies, I’d love to chat!
https://www.gossipinsights.com/en/top-companies/us/