But I’m routing for you!
The small minority that keeps a country by the balls is not the unions but the owning class. The 2008 crash that put the whole world in a decade recession is collateral damage.
What makes life better for everyone is competition. Canada's stagnation can be be summed up in a single phrase - lack of competition. Generally, the US has been a free-for-all when it comes to competition and hence its populace enjoys some of the best living standards.
I'll also relate my experience traveling the subway in Asia vs. Manhattan. Asian transit seems like space-age compared to what we have in the West. I think UBI won't save us as the income must come from somewhere. Hiking taxes kills incentives. The better way is to have more freedom/efficiencies in my humble opinion.
I get that that's unfortunate and perhaps a very serious problem for people working in that industry but what choice do we really have? I'm reminded of something a teacher said to me in high school about how one day many of my classmates and I would have jobs that didn't exist when he was a kid. Isn't that how this works? As time goes on some jobs go away and new jobs come about and there is some pain in the interim? I'm all for figuring out some way to ease that pain for the people in the transitional period but I don't know who's really responsible for that.
Another thing that seems troubling is how a small group of people can hold a majority of the country by the bXlls. Given how this is an election year, I can see this turning into huge fiasco. The rest of the economy is collateral damage.
In designing software, there's often a trade off between (i) generality / configurability, and (ii) performance.
llama.cpp is built for inference, not for training or model architecture research. It seems reasonable to optimize for performance, which is what ~100% of llama.cpp users care about.
llama
mpt
gptneox
gptj
gpt2
bloom
falcon
rwkvI'm also a bit confused by the quantization aspect. This is a pretty complex topic. GGML seems to use 16bit as per the article. If was pushing it to 8bit, I reckin I'd see no size improvement the GGML file? The article says they encode quantization versions in that file. Where are they defined?
I like to play around with smaller models and regular app code in Common Lisp or Racket, and Mistral 7b is very good for that. Mixing and matching old fashioned coding with the NLP, limited world knowledge, and data manipulation capabilities of LLMs.