Readit News logoReadit News
wafngar commented on Terrain Diffusion: A Diffusion-Based Successor to Perlin Noise   arxiv.org/abs/2512.08309... · Posted by u/kelseyfrog
noodletheworld · 3 months ago
Mm. This paper makes it hard to understand what they've done.

For example:

> MultiDiffusion remains confined to bounded domains: all windows must lie within a fixed finite canvas, limiting its applicability to unbounded worlds or continuously streamed environments.

> We introduce InfiniteDiffusion, an extension of MultiDiffusion that lifts this constraint. By reformulating the sampling process to operate over an effectively infinite domain, InfiniteDiffusion supports seamless, consistent generation at scale.

…but:

> The hierarchy begins with a coarse planetary model, which generates the basic structure of the world from a rough, procedural or user-provided layout. The next stage is the core latent diffusion model, which transforms that structure into realistic 46km tiles in latent space. Finally, a consistency decoder expands these latents into a high-fidelity elevation map.

So, the novel thing here is slightly better seemless diffusion image gen.

…but, we generate using a heirsrchy based on a procedural layout.

So basocally, tldr: take perlin noise, resize it, and then image-2-image use it as a seed to generate detailed tiles?

People have already been doing this.

Its not novel.

The novel part here is making the detailed tiles slightly nicer.

Eh. :shrug:

The paper obfuscates this, quite annoyingly.

Its unclear to me why you cant just use multi diffusion for this, given your top level input is already bounded (eg. User input) and not infinite.

wafngar · 3 months ago
Very common in ML research these days - claim novelty / cite prior work in an obfuscated way and so on.
wafngar commented on Magistral — the first reasoning model by Mistral AI   mistral.ai/news/magistral... · Posted by u/meetpateltech
pu_pe · 9 months ago
Benchmarks suggest this model loses to Deepseek-R1 in every one-shot comparison. Considering they were likely not even pitting it against the newer R1 version (no mention of that in the article) and at more than double the cost, this looks like the best AI company in the EU is struggling to keep up with the state-of-the-art.
wafngar · 9 months ago
But they have built a fully “independent” pipeline. Deepseek and others probably trained in gpt4, o1 or whatever data.
wafngar commented on Nvidia adds native Python support to CUDA   thenewstack.io/nvidia-fin... · Posted by u/apples2apples
disgruntledphd2 · a year ago
This just makes it much, much easier for people to build numeric stuff on GPU, which is great.

I'm totally with you that it's better that this took so long, so we have things like PyTorch abstracting most of this away, but I'm looking forward to (in my non-existent free time :/ ) playing with this.

wafngar · a year ago
Why not use torch.compile()?
wafngar commented on Project Aardvark: reimagining AI weather prediction   turing.ac.uk/blog/project... · Posted by u/bentobean
scellus · a year ago
No, they say end-to-end, meaning they use raw obsevations. Most or all other medium-range models start with ERA5.

There's a paper from Norway that tried end-to-end, but their results were not spectacular. That's the aim of many though, including ECMWF. Note that ECMWF already has their AIFS in production, so AI weather prediction is pretty mainstream nowadays.

Google has a local nowcast model that uses raw observations, in production, but that's a different genre of forecasting than the medium-range models of Aardvark.

wafngar · a year ago
They train with ERA5 and observations.
wafngar commented on The FFT Strikes Back: An Efficient Alternative to Self-Attention   arxiv.org/abs/2502.18394... · Posted by u/iNic
wafngar · a year ago
Should be a relevant reference: https://arxiv.org/abs/2111.13587

Adaptive Fourier Neural Operators: Efficient Token Mixers for Transformers John Guibas, Morteza Mardani, Zongyi Li, Andrew Tao, Anima Anandkumar, Bryan Catanzaro

wafngar commented on EU to mobilize 200B Euros to invest in AI   euronews.com/next/2025/02... · Posted by u/FinnLobsien
jycr753 · a year ago
I believe it's because they have this massive monster in Brussels that needs something to do (chiringuitos) to extract the maximum amount of money—corruption.
wafngar · a year ago
Brussels is quite small. Probably less than Washington.
wafngar commented on The future of Deep Learning frameworks   neel04.github.io/my-websi... · Posted by u/lairv
n7g · 2 years ago
> with torch.compile the main advantage of jax is disappearing.

Interesting take - I agree here somewhat.

But also, wouldn't you think a framework that has been from the ground-up designed around a specific, mature compiler stack be better able to integrate compilers in a more stable fashion than just shoe-horning static compilers into a very dynamic framework? ;)

wafngar · 2 years ago
Depends. PyTorch on the other hand has a large user base and well defined and tested api. So should be doable; and is already progressing and rapid speed..
wafngar commented on The future of Deep Learning frameworks   neel04.github.io/my-websi... · Posted by u/lairv
patrickkidger · 2 years ago
I think a lot of the commenters here are being rather unfair.

PyTorch has better adoption / network effects. JAX has stronger underlying abstractions.

I use both. I like both :)

wafngar · 2 years ago
Probably unfair as a reaction to the unfair statements in the blog…
wafngar commented on The future of Deep Learning frameworks   neel04.github.io/my-websi... · Posted by u/lairv
wafngar · 2 years ago
PyTorch is developed by multiple companies / stake holders while jax is google only with internal tooling they don’t share with the world. This alone is a major reason not to use jax. Also I think it is more the other way around: with torch.compile the main advantage of jax is disappearing.

u/wafngar

KarmaCake day29December 12, 2020View Original