Readit News logoReadit News
realPtolemy commented on A love letter to the CSV format   github.com/medialab/xan/b... · Posted by u/Yomguithereal
realPtolemy · 5 months ago
I love CSV
realPtolemy commented on I learned Haskell in just 15 years   duckrabbit.tech/articles/... · Posted by u/aranchelk
munchler · a year ago
Cute. All kidding aside, though, functional programming is worth the effort to learn, and it doesn't actually take 15 years. The payoff is at the end of the article:

"It’s quite natural to program in Haskell by building a declarative model of your domain data, writing pure functions over that data, and interacting with the real world at the program’s boundaries. That’s my favorite way to work, Haskell or not."

Haskell can be intimidating, though, so I would recommend F# for most beginners. It supports OOP and doesn't require every single function to be pure, so the learning curve is less intense, but you end up absorbing the same lesson as above.

realPtolemy · a year ago
Or Elixir! Quite easy to grasp as well.
realPtolemy commented on Extracting concepts from GPT-4   openai.com/index/extracti... · Posted by u/davidbarker
lanceflt · a year ago
https://github.com/openai/sparse_autoencoder

They actually open sourced it, for GPT-2 which is an open model.

realPtolemy · a year ago
Thanks, I must have read through the document to hastily.
realPtolemy commented on Extracting concepts from GPT-4   openai.com/index/extracti... · Posted by u/davidbarker
thegrim33 · a year ago
From the article:

"We currently don't understand how to make sense of the neural activity within language models."

"Unlike with most human creations, we don’t really understand the inner workings of neural networks."

"The [..] networks are not well understood and cannot be easily decomposed into identifiable parts"

"[..] the neural activations inside a language model activate with unpredictable patterns, seemingly representing many concepts simultaneously"

"Learning a large number of sparse features is challenging, and past work has not been shown to scale well."

etc., etc., etc.

People say we don't (currently) know why they output what they output, because .. as the article clearly states, we don't.

realPtolemy · a year ago
Could there also be a “legal hedging” reason for why you would release a paper like this?

By reaffirming that “we don’t know how this works, nobody does” it’s easier to avoid being charged with copyright infringement from various actors/data sources that have sued them.

realPtolemy commented on Extracting concepts from GPT-4   openai.com/index/extracti... · Posted by u/davidbarker
andreyk · a year ago
Exciting to see this so soon after Anthropic's "Mapping the Mind of a Large Language Model" (under 3 weeks). I find these efforts really exciting; it is still common to hear people say "we have no idea how LLMs / Deep Learning works", but that is really a gross generalization as stuff like this shows.

Wonder if this was a bit rushed out in response to Anthropic's release (as well as the departure of Jan Leike from OpenAI)... the paper link doesn't even go to Arxiv, and the analysis is not nearly as deep. Though who knows, might be unrelated.

realPtolemy · a year ago
Indeed, and the very last section about how they’ve now “open sourced” this research is also a bit vague. They’ve shared their research methodology and findings… But isn’t that obligatory when writing a public paper?
realPtolemy commented on Programming languages endorsed for server-side use at Meta (2022)   engineering.fb.com/2022/0... · Posted by u/ra7
realPtolemy · 2 years ago
I’m curious to know why they don’t support Elixir and Haskell to a greater extent.
realPtolemy commented on Microsoft Phi-2 model changes licence to MIT   huggingface.co/microsoft/... · Posted by u/regularfry
regularfry · 2 years ago
It's 2.7B, not 1.1. In my experience it goes off the rails and starts generating nonsense after a few paragraphs, but I haven't dug too much into tweaking the kv cache params to see if that's controllable. It also needs a fair bit of prompt massaging to get it to do what you want. So no, not GPT3.5, but it's comfortably better than anything else in its size class.
realPtolemy · 2 years ago
How is it compared to 7B LLaMA quantized to run on a raspberry pi?
realPtolemy commented on Microsoft Phi-2 model changes licence to MIT   huggingface.co/microsoft/... · Posted by u/regularfry
borissk · 2 years ago
Don't think this is the biggest danger. In a few years if they continue to improve at the current speed these models can become really dangerous. E.g. an organization like ISIS can feed one some books and papers on chemistry and ask it "I have such and such ingredients available, what is the deadliest chemical weapon of mass destruction i can create". Or use it to write the DNA for a deadly virus. Or a computer virus. Or use one to contact millions of say Muslim young men and try to radicalize them.
realPtolemy · 2 years ago
They can already do that with some simple googling.
realPtolemy commented on Reviving decade-old Macs with antiX and MX Linux (2022)   sts10.github.io/2022/12/1... · Posted by u/indigodaddy
pndy · 2 years ago
For my macbook pro 13" late 2011 I've pick manjaro with gnome and with OpenCore Legacy Patcher I managed to squeeze macos 13.5 on in.

While manjaro makes the machine relatively useful for the current day - for really simple tasks that is, the 13.5 misses much especially with low amount of ram. But worse thing is that this device comes with Intel HD Graphics 3000 and some stuff just simply won't load like Maps because this GPU isn't supported by Metal API.

In the past I tried Lubuntu on eMac G4 with 800mhz PPC - it was a rather painful experience. So I replaced it with 10.4 and with help of Macintosh Garden slapped additionally os 9.2.2 for for some dualboot retro fun

realPtolemy · 2 years ago
I’m running Manjaro on a late 2012 MacBook Pro, it’s running flawlessly.

The only initial issue I had was that I had to configurate the broadcom WiFi driver manually. Super easy after some googling for the more technically interested person… But nothing I would expect the general person to solve without frustration.

8gb RAM, 125gb SSD. 1xBattery replacement.

It does not run IDE’s like ST32 (based on Eclipse) and OneNote via Firefox smoothly simultaneously. But everything else so far has worked flawlessly. I use it every day and have done so since the beginning of this fall.

Edit: Call me a script kid all you want, for this particular laptop it’s just more convenient for me to run Manjaro rather than Arch right of the bat. But I guess I could eventually set up a shared partition for storage and an individual partition for each OS.

u/realPtolemy

KarmaCake day43May 20, 2023View Original