Readit News logoReadit News
ignorantguy commented on GPT-5.4   openai.com/index/introduc... · Posted by u/mudkipdev
ignorantguy · 8 days ago
it shows a 404 as of now.
ignorantguy commented on Thin desires are eating life   joanwestenberg.com/thin-d... · Posted by u/mitchbob
ignorantguy · 3 months ago
Did you read the article? its definitely not bs.
ignorantguy commented on My dream thermostat (2024)   leontrolski.github.io/the... · Posted by u/leontrolski
behnamoh · 10 months ago
Am I the only one who naively clicked on all the buttons hoping to see them flick? :)
ignorantguy · 10 months ago
I did the same thing!!
ignorantguy commented on QwQ-32B: Embracing the Power of Reinforcement Learning   qwenlm.github.io/blog/qwq... · Posted by u/nwjsmith
antirez · a year ago
Note the massive context length (130k tokens). Also because it would be kinda pointless to generate a long CoT without enough context to contain it and the reply.

EDIT: Here we are. My first prompt created a CoT so long that it catastrophically forgot the task (but I don't believe I was near 130k -- using ollama with fp16 model). I asked one of my test questions with a coding question totally unrelated to what it says:

<QwQ output> But the problem is in this question. Wait perhaps I'm getting ahead of myself.

Wait the user hasn't actually provided a specific task yet. Let me check again.

The initial instruction says:

"Please act as an AI agent that can perform tasks... When responding, first output a YAML data structure with your proposed action, then wait for feedback before proceeding."

But perhaps this is part of a system prompt? Wait the user input here seems to be just "You will be given a problem. Please reason step by step..." followed by a possible task? </QwQ>

Note: Ollama "/show info" shows that the context size set is correct.

ignorantguy · a year ago
Yeah it did the same in my case too. it did all the work in the <think> tokens. but did not spit out the actual answer. I was not even close to 100K tokens
ignorantguy commented on Ask HN: What should I read/watch/code next to understand ChatGPT technicals?    · Posted by u/humblebeing
ignorantguy · a year ago
I just got this book Build a Large Language Model (From Scratch) by Sebastian Raschka (Author) Might help you dive into the internals.
ignorantguy commented on     · Posted by u/NovaSkye
ignorantguy · a year ago
dont do this. it will ask you a bunch of questions and you have to enter your email address and name to get the results. (which subscribes to their newsletter)
ignorantguy commented on Own Vercel alternative in 8 less cost. Do self host in one command   titancli.com/... · Posted by u/ivanskripnik
ignorantguy · a year ago
doesn't cdk provide the same functionality? what are the additional benefits?
ignorantguy commented on     · Posted by u/dorendra
ignorantguy · a year ago
goes straight to amazon link for the book. seems like an advertisement.
ignorantguy commented on I have an idea but I need a sanity check    · Posted by u/peterweyand38
ignorantguy · a year ago
sure. definitely
ignorantguy commented on     · Posted by u/gnicholas
huseyinkeles · 2 years ago
TLDR: They are planning to not to include the Apple logo stickers in the boxes.
ignorantguy · 2 years ago
thank you for that.

u/ignorantguy

KarmaCake day162May 16, 2016
About
[ my public key: https://keybase.io/ignorantguy; my proof: https://keybase.io/ignorantguy/sigs/Hl2Xzc0zWXWLxEq812Y2pikrGixC9ZV99XiKFuFtL1w ]
View Original