Countless books with irrevocably broken references - https://www.google.com/search?q=%22://goo.gl%22&sca_upv=1&sc...
And for what? The cost of keeping a few TB online and a little bit of CPU power?
An absolute act of cultural vandalism.
Overcast link to relevant chapter: https://overcast.fm/+BOOFexNLJ8/02:33
Original episode link: https://shows.arrowloop.com/@abstractions/episodes/001-the-r...
However such projects require a lot of time and effort and it’s not clear if this project can be forked and kept alive.
Those README changes only served to provide greater transparency to would-be users.
Ulterior motives, indeed.
It’s amazing to see others build on top of open-source projects. Forks like RamaLama are exactly what open source is all about. Developers with different design philosophies can still collaborate in the open for everyone’s benefit.
Some folks on the Ollama team have contributed directly to the OCI spec, so naturally we started with tools we know best. But we made a conscious decision to deviate because AI models are massive in size - on the order of gigabytes - and we needed performance optimizations that the existing approaches didn’t offer.
We have not forked llama.cpp, We are a project written in Go, so naturally we’ve made our own server side serving in server.go. Now, we are beginning to hit performance, reliability and model support problems. This is why we have begun transition to Ollama’s new engine that will utilize multiple engine designs. Ollama is now naturally responsible for the portability between different engines.
I did see the complaint about Ollama not using Jinja templates. Ollama is written in Go. I’m listening but it seems to me that it makes perfect sense to support Go templates.
We are only a couple of people, and building in the open. If this sounds like vendor lock-in, I'm not sure what vendor lock-in is?
You can check the source code: https://github.com/ollama/ollama
Those rejected README changes only served to provide greater transparency to would-be users, and here we are a year and a half later with woefully inadequate movement on that front.
I am very glad folks are working on alternatives.
I have been using a shell function to do this, and it works wonderfully well: https://github.com/phiresky/ripgrep-all/wiki/fzf-Integration
The built-in rga-fzf command appeared in v0.10 and ostensibly obviates the need for the above shell function, but the built-in command produces errors for me on MacOS: https://github.com/phiresky/ripgrep-all/issues/240
It used to be that you could count on a sleeping Mac to stay asleep until you explicitly woke it up by opening the lid, pressing a key on the keyboard, or pressing the trackpad (or separate trackpad buttons — yes, those used to exist). Perhaps more importantly, you could count on the sleep function to have barely any effect on the Mac battery.
I don’t recall when, but at some point over the aforementioned three decades, Apple started changing the terms of this sleep contract.
It seems Apple decided that some functions should still be available when the Mac is “sleeping”, with no way to restore the previous behavior. As a result, random wake-ups are the new normal, replete with unexpected battery drainage.
I have seen modern MacBooks go to “sleep” with an 80% charge at night, only to be rendered dead with a 0% battery level by morning. Does something this extreme happen often? No. But it never happened before. And moderate battery loss while sleeping? That happens very often on today’s MacBooks.
Your mileage may very. And Apple probably continues to implement better computer sleep behavior than competing vendors. But I would argue that people who think MacBook sleep behavior is excellent… have never experienced how it behaved in the past. That behavior was superb.
But sadly, I imagine most people have never experienced that excellence.
Most of Pelican’s code was written by other people, and yet I have spent almost zero time debugging that code, much less my own. After taking advantage of Pelican’s rich plugin ecosystem and adding a handful of useful plugins, I continue to be amazed by how much time this publishing system saves me, and how little time I must spend to keep everything running smoothly.
What it would take to accomplish this by writing HTML by hand instead… I simply can’t fathom it. But once again, that’s just one person’s experience, and YMMV.
Thank you for helping me learn Japanese! :)
Can you please explain what do you mean by actively supporting JMDict? I hope I didn't make an attribution mistake, or misunderstood something. My understanding is that I can use those files in my project as long as I follow the license guidelines.
It makes me really happy that so many people are interested in it. :)
I didn't think this many people would be interested. I'll write a guide for Jellyfin, then add Italian, French and Dutch languages tomorrow.
And as someone who now also speaks Italian, I am even more pleased to see that Italian support will be added tomorrow.
It is wonderful to see such a useful tool released as an open-source, self-hosted project. (^_^)
[1] EDICT: http://edrdg.org/jmdict/edict_doc_2009.html
[2] JMdict: https://en.wikipedia.org/wiki/JMdict
I have not yet tried either one, but here is the other project for those who want to compare and contrast them:
https://github.com/manaflow-ai/cmux