Readit News logoReadit News
snake_doc commented on GPT-5.2   openai.com/index/introduc... · Posted by u/atgctg
snake_doc · 5 days ago
> Models were run with maximum available reasoning effort in our API (xhigh for GPT‑5.2 Thinking & Pro, and high for GPT‑5.1 Thinking), except for the professional evals, where GPT‑5.2 Thinking was run with reasoning effort heavy, the maximum available in ChatGPT Pro. Benchmarks were conducted in a research environment, which may provide slightly different output from production ChatGPT in some cases.

Feels like a Llama 4 type release. Benchmarks are not apples to apples. Reasoning effort is across the board higher, thus uses more compute to achieve an higher score on benchmarks.

Also notes that some may not be producible.

Also, vision benchmarks all use Python tool harness, and they exclude scores that are low without the harness.

snake_doc commented on IBM CEO says there is 'no way' spending on AI data centers will pay off   businessinsider.com/ibm-c... · Posted by u/nabla9
throwaway31131 · 14 days ago
100GW per year is not going to happen.

The largest plant in the world is the Three Gorges Dam in China at 22GW and it’s off the scales huge. We’re not building the equivalent of four of those every year.

Unless the plan is to power it off Sam Altman’s hot air. That could work. :)

https://en.wikipedia.org/wiki/List_of_largest_power_stations

snake_doc · 14 days ago
China added ~90GW of utility solar per year in last 2 years. There's ~400-500GW solar+wind under construction there.

It is possible, just may be not in the U.S.

Note: given renewables can't provide base load, capacity factor is 10-30% (lower for solar, higher for wind), so actual energy generation will vary...

snake_doc commented on Modular Manifolds   thinkingmachines.ai/blog/... · Posted by u/babelfish
aghilmort · 3 months ago
Interesting. Modular manifolds are precisely what hypertokens use for prompt compiling.

Specifically, we linearize the emergent KVQ operations of an arbitrary prompt in any arbitrary model by way of interleaving error-correcting code (ECC).

ECC tokens are out-of-band tokens, e.g., Unicode's Private Use Area (PUA), interleaved with raw context tokens. This construction induces an in-context associate memory.

Any sort of interleaved labeling basis, e.g., A1, quick brown fox, A2, jumped lazy dog, induces a similar effect to for chaining recall & reasoning more reliably.

This trick works because PUA tokens are generally untrained hence their initial embedding is still random Gaussian w.h.p. Similar effects can be achieved by simply using token combos unlikely to exist and are often in practice more effective since PUA tokens like emojis or Mandarin characters are often 2,3, or 4 tokens after tokenization vs. codeword combos like zy-qu-qwerty every k content tokens, where can be variable.

Building attention architecture using modular manifolds in white / gray-box models like this new work shows vs. prompt-based black box injection is a natural next step, and so can at least anecdotally validate what they're building ahead of next paper or two.

Which is all to say, absolutely great to see others building in this way!

snake_doc · 3 months ago
Wot? Is this what AI generated non-sense has come to? This is totally unrelated.
snake_doc commented on Modular Manifolds   thinkingmachines.ai/blog/... · Posted by u/babelfish
whimsicalism · 3 months ago
this is a bad example to claim the bitter lesson applies to, it’s about the fundamentals of optimization techniques not about tying to hand-crafted things for the solution space.
snake_doc · 3 months ago
Aren’t they all optimization techniques at the end of the day? Now you’re just debating semantics
snake_doc commented on Modular Manifolds   thinkingmachines.ai/blog/... · Posted by u/babelfish
jasonjmcghee · 3 months ago
The learning rates they demonstrate are crazy - though the standard when talking about CIFAR-10 is 94% accuracy iirc. Showing ~60% accuracy is weird.

Has DAWNBench been done with manifold Muon (with a more appropriate architecture)?

snake_doc commented on Trump to impose $100k fee for H-1B worker visas, White House says   reuters.com/business/medi... · Posted by u/mriguy
mschuster91 · 3 months ago
It does go to the government and not to Trump's personal wallet (like the memecoins and lavish gift), it's just a tax that's just not being called a tax, and frankly it's a good idea. The current abuse of H1B doesn't work out positively for anyone but the companies making a boatload of money on exploiting people.
snake_doc · 3 months ago
Oh? And taxes can’t be used to buy influence and votes? How naive… Money is fungible… one pocket into another

Exhibit 1: Tariff revenues to bail out American farmers: https://www.ft.com/content/0267b431-2ec9-4ca4-9d5c-5abf61f2b...

snake_doc commented on Trump to impose $100k fee for H-1B worker visas, White House says   reuters.com/business/medi... · Posted by u/mriguy
breadwinner · 3 months ago
snake_doc · 3 months ago
SCMP is owned by Alibaba, which is subject to the purview of the Chinese Central Government [1].

[1]: https://www.cecc.gov/agencies-responsible-for-censorship-in-...

snake_doc commented on Trump to impose $100k fee for H-1B worker visas, White House says   reuters.com/business/medi... · Posted by u/mriguy
snake_doc · 3 months ago
Mafia behavior continues… (not my observation, but the Texas senator’s Ted Cruz[1]).

$100k is a big pizzo (protection fee)!

[1]: https://www.bloomberg.com/news/articles/2025-09-19/ted-cruz-...

> “That’s right outta ‘Goodfellas,’ that’s right out of a mafioso going into a bar saying, ‘Nice bar you have here, it’d be a shame if something happened to it,’” Cruz said, using the iconic New York accent associated with the Mafia.

u/snake_doc

KarmaCake day673April 18, 2022View Original