Neoverse V3 is the server version of the Cortex-X4 core which has been used in a large number of smartphones.
The Neoverse V3 and Cortex-X4 cores are very similar in size and performance with the Intel E-cores Skymont and Darkmont (the E-cores of Arrow Lake and of the future Panther Lake).
Intel will launch next year a server CPU with Darkmont cores (Clearwater Forest), which will have cores similar to this AWS Graviton5, but for now Intel only has the Sierra Forest server CPUs with E-cores (belonging to the Xeon 6 series), which use much weaker CPU cores than those of the new Graviton5 (i.e. cores equivalent with the Crestmont E-cores of the old Meteor Lake).
AMD Zen 5 CPUs are significantly better for computationally-intensive workloads, but for general-purpose applications without great computational demands the cores of Graviton5 and also Intel Skymont/Darkmont have greater performance per die area and power consumption, therefore lower cost.
They used a c5.4xlarge that has peak 10Gbps bandwidth, which at a constant 100% saturation would take in the ballpark of 9 minutes to load those 650GB from S3, making those 9 minutes your best case scenario for pulling the data (without even considering writing it back!)
Minute differences in how these query engines schedule IO would have drastic effects in the benchmark outcomes, and I doubt the query engine itself was constantly fed during this workload, especially when evaluating DuckDB and Polars.
The irony of workloads like this is that it might be cheaper to pay for a gigantic instance to run the query and finish it quicker, than to pay for a cheaper instance taking several times longer.
BUT the author did say this is the simple stupid naive take, in which case DuckDB and Polars really shined.
Cloud’s big promise was speed to market and price, and let’s be honest, price is no longer there compared to a decent operation.
The one thing where clouds remain kings is speed for small teams. Any large enough company should probably Ask themselves whether running their own operation using ias would be a better choice.
Because on prem is inelastic, we are at sub 10% peak utilization of compute resources. If we added in the likely higher cloud utilization rate we are talking of 30%+ savings from on prem.
Maybe it's just the kind of work I'm doing, a lot of web development with html/scss, and Google has crawled the internet so they have more data to work with.
I reckon different models are better at different kinds of work, but Gemini is pretty excellent at UI/UX web development, in my experience
Very excited to see what 3.0 is like