It doesn't surprise me that people aren't seeking out book reviews on the AP website or app—I don't think AP is particularly associated with reviews, maybe deliberately because they've historically been read in local papers that don't emphasize the AP sourcing, so people wanting reviews from a national source probably go to NYT, WaPo, WSJ, the New Yorker, etc. first.
This is a tangent, but I wonder if they feel that they are just creating LLM training data and that few readers (even of Sunday papers) will actually read their reviews.
If the query uses the index, then the on the fly tsvector rechecks are only on the matches and the benchmark queries have LIMIT 10, so few rechecks right?
Edit: yes but the query predicates have conditions on 2 gin indexes, so I guess the planner chooses to recheck all the matches for one index first even though it could avoid lots of work by rechecking row-wise
Also, 10M input token context is insane!
EDIT: https://huggingface.co/meta-llama/Llama-3.1-405B is BF16 so yes, it seems training in FP8 is new.
At one time I thought I read that there was a project to embed Kuzu into DuckDB, but bringing a vector store natively into kuzu sounds even better.
htmz is a minimalist HTML microframework for creating interactive and modular web user interfaces with the familiar simplicity of plain HTML.