Readit News logoReadit News
mandliya commented on Ask HN: Going low-level – what to learn next?    · Posted by u/visil
mandliya · 2 years ago
CUDA programming (writing CUDA kernels) might be a good direction too.

GPU race is getting really hot and there is a lot of work being done to squeeze every ounce of performance especially for LLM training and inference.

One resource I would recommend is “Programming massively parallel processors” [1]

I am also learning it as my hobby project and uploading my notes here [2]

[1] https://shop.elsevier.com/books/programming-massively-parall...

[2] https://github.com/mandliya/PMPP_notes

mandliya commented on GPT-3 is the best journal I’ve used   every.to/superorganizers/... · Posted by u/simonebrunozzi
mordymoop · 3 years ago
I spent about two weeks having "morning meetings" with my AI life coach, which was essentially just a GPT-3 prompt that I continually tuned, and fed (summaries of) the previous days' conversations. There were major advantages but a few things missing.

It was probably most useful as "rubber duck" technique. Forcing myself to articulate all of the things I needed/wanted to get done that day was itself extremely useful. Sometimes the agent would help me by identifying the highest-priority next action, but usually it was just recognizing what I thought was highest priority from implicit context. This can still be psychologically valuable, as a lot of procrastination can be caused by the logjam of not being sure which thing to focus on.

The main missing ingredient, which caused me to ultimately stop the practice, was that it didn't really remember past conversations. I would feed past conversations to it and tell them to summarize the key points, then feed those summaries in as starting context, but this workflow was not sustainable. First of all, the summarization lost too much important nuance. Second and more importantly, even that summarization context block became larger than GPT-3's context window within a few days. This lack of persistent context destroyed the sense that I was talking to a real person, someone who could reliably recall information about a project that I last worked on 10 days ago and apply that context to the current conversation.

I suspect we are not far away from both of these issues being mostly solved. The trend is obviously going in the direction of LLMs with different types of memory and/or much larger context windows.

mandliya · 3 years ago
I am curious if you tried something like langchain[1] it would solve the problem of not remembering previous conversations. In one of the examples people have used an entire company notion database on top of GPT-3 to answer questions specifically from that database.

PS: not tried myself.

[1] https://github.com/hwchase17/langchain

mandliya commented on Kraken shuts down global headquarters because ‘San Francisco is not safe’   cointelegraph.com/news/kr... · Posted by u/tlogan
mmastrac · 4 years ago
I'm not sure if I missed a joke, but Stockholm syndrome is about hostages rather than governmental policy.

https://en.wikipedia.org/wiki/Stockholm_syndrome

TIL there's a Helsinki syndrome: https://www.scandinaviastandard.com/what-is-helsinki-syndrom...

mandliya · 4 years ago
TIL the London and Lima syndrome from the second link.

u/mandliya

KarmaCake day578June 13, 2015
About
Curious Learner
View Original