Readit News logoReadit News
ssivark commented on Over fifty new hallucinations in ICLR 2026 submissions   gptzero.me/news/iclr-2026... · Posted by u/puttycat
stainablesteel · 8 days ago
this brings us to a cultural divide, westerners would see this as a personal scar, as they consider the integrity of the publishing sphere at large to be held up by the integrity of individuals

i clicked on 4 of those papers, and the pattern i saw was middle-eastern, indian, and chinese names

these are cultures where they think this kind of behavior is actually acceptable, they would assume it's the fault of the journal for accepting the paper. they don't see the loss of reputation to be a personal scar because they instead attribute blame to the game.

some people would say it's racist to understand this, but in my opinion when i was working with people from these cultures there was just no other way to learn to cooperate with them than to understand them, it's an incredibly confusing experience to be working with them until you understand the various differences between your own culture and theirs

ssivark · 8 days ago
PSA: Please note that the names are hallucinated author lists part of the hallucinated citations, and not names of offending authors.

AFAIK the submissions are still blinded and we don't know who the authors are. We will, surely, soon -- since ICLR maintains all submissions in public record for posterity, even if "withdrawn". They are unblinded after the review period finishes.

ssivark commented on Zig quits GitHub, says Microsoft's AI obsession has ruined the service   theregister.com/2025/12/0... · Posted by u/Brajeshwar
sunnyday_002 · 13 days ago
One of my favourite GitHub features is the ability to do a code search over the whole of GitHub, not sure GitLab has the same when I use to use it?
ssivark · 13 days ago
Code search over all of Gitlab (even if available) wouldn't help much when many of the interesting repos might be on Github. To be truly useful, it would need to index repos across many different forges. But there's a tension in presenting that to users if you're afraid that they might exit your ecosystem to go to another forge.
ssivark commented on India orders smartphone makers to preload state-owned cyber safety app   reuters.com/sustainabilit... · Posted by u/jmsflknr
Animats · 15 days ago
What does this app actually do, in detail? Anyone know?
ssivark · 15 days ago
This seems to be the app: https://www.sancharsaathi.gov.in/

Looks like it's quire popular/established already, with over 10 million downloads. Basically a "portal" for basic digital safety/hygiene related services.

Quoting Perplexity regarding what facilities the app offers:

1. Chakshu: Report suspicious calls, SMS, or WhatsApp for scams like impersonation, fake investments, or KYC frauds.

2. Block Lost/Stolen Phones: Trace and block devices across all telecom networks using IMEI; track if reactivated.

3. Check Connections in Your Name: View and disconnect unauthorized numbers linked to your ID.

4. Verify Device Genuineness: Confirm if a phone (new or used) is authentic before purchase.

ssivark commented on Migrating to Positron, a next-generation data science IDE for Python and R   posit.co/blog/positron-mi... · Posted by u/ionychal
akst · 18 days ago
I guess in my mind I was thinking use cases other than AI. Like statistical or hierarchical scientific models, simulations or ETL work. I also don't know if some of the econometricians I know with a less technical background would even know how to get setup with AWS, and I feel more boardly there's enough folks doing data work in a none tech field who know how to use Python or R or Matlab to do their modelling but likely isn't comfortable with cloud infrastructure, but might have an apple laptop with apple silicon that could improve their iteration loop. Folks in AI are probably more comfortable with a cloud solution.

There are aspects of data science which is iterative and you're repeatedly running similar computations with different inputs, I think there's some value in shaving off time between iterations.

In my case I have a temporal geospatial dataset with 20+ million properties for each month over several years each with various attributes, it's in a nonprofessional setting and the main motivator for most of my decisions is "because I can and I think it would be fun and I have a decent enough GPU". While I could probably chuck it on a cluster, I'd like to avoid if I can help it and an optimisation done on my local machine would still pay off if I did end up setting up a cluster. There's quite a bit of ETL preprocessing work before I load it into the database, I think are portions that might be doable on the GPU. But it's more so the computations I'd like to do on the dataset before generating visualisations in which I think I could reduce the iteration wait time for processing for plots, ideally to the point I can make iterations more interactive. There's enough linear operations you could get some wins with a GPU implementation.

I am keen to see how far I'll get, but worst case scenario I learn a lot, and I'm sure those learnings will be transferrable to other GPU experiments.

ssivark · 18 days ago
TBC, I too did not really mean "AI" (as in LLMs and the like) which is often hosted/served with a very convenient interface. I do include more bespoke statistical / mathematical models -- be it hierarchical, Bayesian, whatever.

Since AWS/etc are quite complicated, there are now a swarm of startups trying to make it easier to take baby steps into the cloud (eg. Modal, Runpod, etc) and make it very easy for the user to get a small slice of that GPU pie. These services have drastically simpler server-side APIs and usage patterns, including "serverless" GPUs from Modal, where you can just "submit jobs" from a Python API without really having to manage containers. On the client side, you have LLM coding agents that are the next evolution in UI frontends -- and they're beginning to make it much much easier to write bespoke code to interact with these backends. To make it abundantly clear what target audience I'm referring to: I imagine they are still mostly using sklearn (or equivalents in other languages) and gradient boosting with Jupyter notebooks, still somewhat mystified by modern deep learning and stuff. Or maybe those who are more mathematically sophisticated but not software engg sophisticated (eg: statisticians / econometricians)

To inspire you with a really concrete example: since Modal has a well documented API, it should be quite straight-forward ("soon", if not already) for any data scientist to use one of the CLI coding agents and

1. Implement a decent GPU-friendly version of whatever model they want to try (as long as it's not far from the training distribution i.e. not some esoteric idea which is nothing like prior art)

2. Whip up a quick system to interface with some service provider, wrap that model up and submit a job, and fetch (and even interpret) results.

----

In case you haven't tried one of these new-fangled coding agents, I strongly encourage you to try one out (even if it's just something on the free tier eg. gemini-cli). In case you have and they aren't quite good enough to solve your problem, tough luck for now... I anticipate their usability will improve substantially every few months.

ssivark commented on Migrating to Positron, a next-generation data science IDE for Python and R   posit.co/blog/positron-mi... · Posted by u/ionychal
akst · 18 days ago
I know "next-generation" is just SEO slop, but I'm going to hyper fixate on this for a moment (so feel free to ignore if you're actually interested in Positron).

I think the future of data science will likely be something else, with the advent of WebGPU[1] (which isn't just a web technology) and the current quality/availability of GPUs in end user devices, and a lot of data computation clearly standing to benefit from this.

The real next generation of data science tools will likely involve tools that are GPU first and try to keep as much work in the GPU as possible. I definitely think we'll see some new languages eventually emerge to abstract much of the overhead of batching work but also forces people to explicitly consider when they write code that simply won't run on the GPU, like sequential operations that are nonlinear, nonassociative/noncommutative (like highly sequential operations like processing an ordered block of text).

I think WebGPU is going to make this a lot easier.

That said I'd imagine for larger compute workloads people are going to continue to stick with large CUDA clusters as they have more functionality and handle a larger variety of workloads. But on end user devices there's an opportunity to create tools that allow data scientists to more trivially do this kind of work when they compute their models, process their datasets.

[1] Other compute APIs existed in the past, but WebGPU might be one of the most successful attempt to provide a portable (and more accessible) way to write general GPU compute code. I've seen people say WebGPU is hard, but having given it ago (without libraries) I don't think this is all that true, compared to OpenGL there are no longer specialised APIs to load data into uniforms everything is just a buffer. I wonder if this has more to do with non JS bindings for use outside the browser/node or the fact you're forced to consider memory layout of anything your loading into the GPU from the start (something that can be abstracted and generalised), just in my experience after my first attempt at writing a compute shader it's fairly simple IMO. Like stuff that always complicated in rendering like text is still complicated, but at least its not a state based API like web/opengl.

ssivark · 18 days ago
Interesting question. I don't know much about WebGPU, but I'd posit (heh!) that the GPU on the client devices doesn't matter too much since folks will likely be working over the network anyways (cloud-based IDE, coding agent connected to cloud-hosted LLM, etc) and we also have innovations like Modal which allow serverless lambdas for GPUs.

As long as silicon is scarce it would make sense to hoard it and rent it out (pricing as a means of managing scarcity); if we end up in a scenario where silicon is plentiful, then everyone would have powerful local GPUs, using local AI models, etc.

ssivark commented on Gemini CLI tips and tricks for agentic coding   github.com/addyosmani/gem... · Posted by u/ayoisaiah
novoreorx · 19 days ago
I really wish there were a de facto state-of-the-art coding agent that is LLM-agnostic, so that LLM providers wouldn't bother reinventing their own wheels like Codex and Gemini-CLI. They should be pluggable providers, not independent programs. In this way, the CLI would focus on refining the agentic logic and would grow faster than ever before.

Currently Claude Code is the best, but I don't think Anthropic would pivot it into what I described. Maybe we still need to wait for the next groundbreaking open-source coding agent to come out.

ssivark · 19 days ago
Opencode (from SST; not the thing that got rebranded as Crush) seems to be just that. I've had a very good experience with it for the last couple of days; having previously used gemini-cli quite a bit. Opencode also has/hosts a couple of "free" models options right now, which are quite decent IMO.

https://github.com/sst/opencode

There are many many similar alternatives, so here's a random sampling: Crush, Aider, Amp Code, Emacs with gptel/acp-shell, Editor Code Assistant (which aims for an editor-agnostic backend that plugs into different editors)

Finally... there is quite a lot of scope for co-designing the affordances / primitives supported by the coding agent and the LLM backing it (especially in LLM post-training). So factorizing these two into completely independent pieces currently seems unlikely to give the most powerful capabilities.

ssivark commented on Migrating the main Zig repository from GitHub to Codeberg   ziglang.org/news/migratin... · Posted by u/todsacerdoti
petersellers · 19 days ago
> I agree that the writing in the blog post is more colorful than precise, but sanitizing every bit of expression dulls the internet. Humans invented language for a reason.

Where do you draw the line, then? Is a racist screed acceptable to you as long as the following paragraph references technical issues correctly?

The language in the blog post is insulting. Imagine how you would feel if you were the person who wrote this code, and now you are being called a monkey in front of thousands of people on the internet. Certainly you've made mistakes in your code before...or are you saying you've always written flawless code without errors?

These codes of conduct always seemed a bit superfluous to me, but after reading comments like these I can totally see why they are necessary.

ssivark · 19 days ago
Would you perhaps have preferred if they referred to it as "unprofessional" or "sloppy" instead alluding of monkeys?

To me all those mean the same thing, except the latter is more flavorful and makes my eyes less likely to glaze over.

> Imagine how you would feel if you were the person who wrote this code, and now you are being called a monkey in front of thousands of people on the internet.

Er.. so? Why should anyone be allowed into a position of responsibility where their code impacts millions of people if they can't handle the tiniest bit of strong feedback? It was, after all, a pretty egregious bug.

> Certainly you've made mistakes in your code before...or are you saying you've always written flawless code without errors?

I've definitely made mistakes, and also accept that my output might have on occasion been "monkey-esque". I don't see what's insulting about that; we are all human/animal.

ssivark commented on Migrating the main Zig repository from GitHub to Codeberg   ziglang.org/news/migratin... · Posted by u/todsacerdoti
johnfn · 19 days ago
> it’s abundantly clear that the talented folks who used to work on the product have moved on to bigger and better things, with the remaining losers eager to inflict some kind of bloated, buggy JavaScript framework on us in the name of progress.

> More importantly, Actions is created by monkeys

This writing really does not reflect well on Zig. If you have technical issues with Github, fine: cite them. But leave ad hominems like "losers" and "monkeys" out of it.

ssivark · 19 days ago
> created by monkeys

I don't particularly care for either Zig or Github, but...

they do precisely cite the technical issues. That snippet links to a Github discussion comment https://github.com/actions/runner/issues/3792#issuecomment-3...

(reproduced below)

"The bug in this "safe sleep" script is obvious from looking at it: if the process is not scheduled for the one-second interval in which the loop would return (due to $SECONDS having the correct value), then it simply spins forever. That can easily happen on a CI machine under extreme load. When this happens, it's pretty bad: it completely breaks a runner until manual intervention. On Zig's CI runner machines, we observed multiple of these processes which had been running for hundreds of hours, silently taking down two runner services for weeks."

"I don't understand how we got here. Even ignoring the pretty clear bug, what makes this Bash script "safer" than calling into the POSIX standard sleep utility? It doesn't seem to solve any problem; meanwhile, it's less portable and needlessly eats CPU time by busy-waiting."

"The sloppy coding which is evident here, as well as the inaction on core Actions bugs (in line with the decay in quality of almost every part of GitHub's product), is forcing the Zig project to strongly consider moving away from GitHub Actions entirely. With this bug, and many others (severe workflow scheduling issues resulting in dozens of timeouts; logs randomly becoming inaccessible; random job cancellations without details; perpetually "pending" jobs), we can no longer trust that Actions can be used to implement reliable CI infrastructure. I personally would seriously encourage other projects, particularly any using self-hosted runners, to look carefully at the stability of Actions and ask themselves whether it is a solution worth sticking with long-term when compared with alternatives."

----

I agree that the writing in the blog post is more colorful than precise, but sanitizing every bit of expression dulls the internet. Humans invented language for a reason.

ssivark commented on Freer Monads, More Extensible Effects (2015) [pdf]   okmij.org/ftp/Haskell/ext... · Posted by u/todsacerdoti
ssivark · a month ago
I would appreciate a simpler conceptual explanation for someone not steeped in the Haskell / functional programming world :-)
ssivark commented on Ask HN: How are Markov chains so different from tiny LLMs?    · Posted by u/JPLeRouzic
Sohcahtoa82 · a month ago
A Markov Chain trained by only a single article of text will very likely just regurgitate entire sentences straight from the source material. There just isn't enough variation in sentences.

But then, Markov Chains fall apart when the source material is very large. Try training a chain based on Wikipedia. You'll find that the resulting output becomes incoherent garbage. Increasing the context length may increase coherence, but at the cost of turning into just simple regurgitation.

In addition to the "attention" mechanism that another commenter mentioned, it's important to note that Markov Chains are discrete in their next token prediction while an LLM is more fuzzy. LLMs have latent space where the meaning of a word basically exists as a vector. LLMs will generate token sequences that didn't exist in the source material, whereas Markov Chains will ONLY generate sequences that existed in the source.

This is why it's impossible to create a digital assistant, or really anything useful, via Markov Chain. The fact that they only generate sequences that existed in the source mean that it will never come up with anything creative.

ssivark · a month ago
Uhhh... the above comment has a bunch of loose assertions that are not quite true, but with a enough truthiness that makes them hard to refute. So I'll point to my other comment for a more nuanced comparison of Markov models with tiny LLMs: https://news.ycombinator.com/item?id=45996794

u/ssivark

KarmaCake day6015January 3, 2012
About
http://sivark.me/

email: siva dot rk dot sw at Google's email service

(Opinions my own; not representing any entity)

View Original