And this isn’t how LLMs are used in practice! Actual agents do a thinking/reasoning cycle after each tool-use call. And I guarantee even these 6-month-old models could do significantly better if a researcher followed best practices.
> you’d have to either memorize the entire answer before speaking or come up with a simple pattern you could do while reciting that takes significantly less brainpower
This part i dont understand. Why would coming up with an algorithm (e.g. a simple pattern) and reciting it be impossible? The paper doesnt mention the models coming up with the algorithm at all AFAIK. If the model was able to come up with the pattern required to solve the puzzles and then also execute (e.g. recite) the pattern, then that'd show understanding. However the models didn't. So if the model can answer the same question for small inputs, but not for big inputs, then doesnt that imply the model is not finding a pattern for solving the answer but is more likely pulling from memory? Like, if the model could tell you fibbonaci numbers when n=5 but not when n=10, that'd imply the numbers are memorized and the pattern for generation of numbers is not understood.
https://www.nhk.or.jp/kishou-saigai/tsunami/
https://www3.nhk.or.jp/news/live/ (live, Japanese)
https://www3.nhk.or.jp/nhkworld/en/live/ (live, English)
The east coast is also where the vast majority of Japan's population lives, and was previously hit by the 2011 tsunami (Fukushima and all that). We're about to find out the hard way what lessons they have learned.
Update: First detected wave in Nemuro, Hokkaido (northernmost Japan) was only 30cm. There may be more. Waves of 3-4m have apparently already hit Kamchatka in Russia.
Update 2: We're almost an hour in and highest waves to actually hit Japan remain only 40 cm. It looks unlikely that this will cause major damage.