Again, remember that grey market payouts are tranched, so you could get 3x more than Google would pay, or you could get 0.5x, and for much more work.
The great thing is I have enough data (100k+) to fine-tune and run a meaningful classification report over. The data is very diverse, and while the labels aren't totally evenly distributed, I can deal with the imbalance with a few tricks.
Can't wait to swap it out for this and see the changes in the scores. Will report back
Since Claude Code is cli based, I reviewed my cli toolset: Migrated from iTerm2 to Ghostty and Tmux, from Cursor to NeoVim (my God is it good!).
Just had a 14h workday with this tooling. It’s so good that I complete the work of weeks and months within days! Absolutely beast.
At this point I am thinking IDEs do not reflect the changing reality of software development. They are designed for navigating project folders, writing / changing files. But I don’t review files that much anymore. I rather write prompts, watch Claude Code create a plan, implement it, even write meaningful commit messages.
Yes I can navigate the project with neovim, yes I can make commits in git and in lazygit, but my task is best spent in designing, planning, prompting, reviewing and testing.
Except that is _not_ true, there is an entire thread of people saying they are unintuitive and hard to understand!
Now imagine saying that sentence to a person outside tech
I’ll also add. I don’t have a good mental model for what a passkey is or how it works. And again, like most users if I don’t really understand what’s going on I’m just not gonna bother with it. For all the complexity that it takes to implement secure login with a username and password, most of it is hidden from the user, with passkeys it feels like they’re shoving all the complexity front and center, but not explaining any of it.
They are woefully designed and implemented, wish we just cut our losses with them and stopped pushing them.
Tuck them away in settings, not on the default login path.
What's a good roll your own solution? DB storage doesn't need to be dynamic like with DynamoDB. At max 1TB - maybe double in the future.
Could this be done on a mid size VPS (32GB RAM) hosting Apache Spark etc - or better to have a couple?
P.S. total beginner in this space, hence the (naive) question.
Computationally speaking - again depends on what your company does - Collect a lot of data? You need a lot of storage.
Train ML Models, you will need GPUs - and you need to think about how to utilise those GPUs.
Or...you could pay databricks, log in and start working.
I worked at a company who tried to roll their own, and they wasted about a year to do it, and it was flaky as hell and fell apart. Self hosting makes sense if you have the people to manage it, but the vast majority of medium sized companies will have engineers who think they can manage this, try it, fail and move on to another company.