Readit News logoReadit News
mishu2 commented on Ask HN: What are you working on? (January 2026)    · Posted by u/david927
mishu2 · a month ago
Over the holidays I built a simple website which lets children (of all ages) easily draw something and then bring the sketch to life using AI and a prompt.

https://funsketch.kigun.org/

Only shared it via Show HN so far, and am still regularly getting some creative submissions. Will be sharing it at an art festival later this year so kids can have a more active role when visiting.

mishu2 commented on Show HN: Fun sketch – Bring your sketches to life   funsketch.kigun.org/... · Posted by u/mishu2
mishu2 · 2 months ago
Thank you everyone for taking a look. The website had around 1,200 visitors and received over 90 sketches over the past 24 hours, and I'm happy to say I could approve almost all of them (all except 2 'rocket' sketches which were starting to look a bit dubious, but all in good fun).

The results look very interesting to me, and I think as a next step I will look into adjusting the prompts (both positive and negative). On the one hand, I'd like to keep the sketch/drawing look rather than going full photorealism as some of the videos do, on the other hand I don't want to restrict the users' creativity too much.

I plan to simplify this further and optimize it for tablets. Some friends said they'd like to have their kids try it out it and I like the idea, as by having to sketch it keeps the user in a more active role.

Please let me know if you have any other suggestions.

mishu2 commented on Show HN: Fun sketch – Bring your sketches to life   funsketch.kigun.org/... · Posted by u/mishu2
VierScar · 2 months ago
Could do something funky like convert it to grayscale, add a 4th "colour" channel and put the grayscale image it that
mishu2 · 2 months ago
I'm actually trying to reduce the 'funkyness', initially the idea was to start from a child's sketch and bring it to life (so kids can safely use it as part of an exhibit at an art festival) :)

There's a world of possibilities though, I hadn't even thought of combining color channels.

mishu2 commented on Show HN: Fun sketch – Bring your sketches to life   funsketch.kigun.org/... · Posted by u/mishu2
fouc · 2 months ago
Oh I was wondering why some of the hallucinations introduced Chinese text/visuals, I'm guessing that might be due to the negative prompt.
mishu2 · 2 months ago
I think the main reason is that the model has a lot of training material with Chinese text in it (I'm assuming, since the research group who released it is from China), but having the negative prompt in Chinese might also play a role.

What I've found interesting so far is that sometimes the image plays a big part in the final video, but other times it gets discarded almost immediately after the first few frames. It really depends on the prompt, so prompt engineering is (at least for this model) even more important than I expected. I'm now thinking of adding a 'system' positive prompt and appending the user prompt to it.

mishu2 commented on TurboDiffusion: 100–200× Acceleration for Video Diffusion Models   github.com/thu-ml/TurboDi... · Posted by u/meander_water
ash_091 · 2 months ago
So with 108,000 (60 X 1,800) Bob Ross PPUs (parallel painting units) we should be able to achieve a stable 60FPS!
mishu2 · 2 months ago
Once you set up a pipeline, sure. They'd need a lot of bandwidth to ensure the combined output makes any kind of sense, not unlike the GPU I guess.

Otherwise it's similar to the way nine women can make a baby in a month. :)

mishu2 commented on Show HN: Fun sketch – Bring your sketches to life   funsketch.kigun.org/... · Posted by u/mishu2
fouc · 2 months ago
Pretty good, I've noticed the animation tends to veer off / hallucinate quite a lot near the end. It is clear that the model is not maintaining any awareness of the first image. I wonder if there's a way to keep the original model in the context, or add original image back in at the half way mark.
mishu2 · 2 months ago
Thank you. I've noticed that too, and also that it has a tendency to introduce garbled text when not given a prompt (or a short one).

This is using the default parameters for the ComfyUI workflow (including a negative prompt written in Chinese), so there is a lot of room for adjustments.

mishu2 commented on Show HN: Fun sketch – Bring your sketches to life   funsketch.kigun.org/... · Posted by u/mishu2
forgingahead · 2 months ago
Nice work! What hardware are you running ComfyUI on, specifically the Wan2.2 workflow? It must be expensive having a GPU running for a hobby project like this?
mishu2 · 2 months ago
Thank you. I'm running it locally on a 4090 (24 GB).

I was running into OOM issues with Wan 2.2 before, but I found the latest version of ComfyUI can now run it (using about 17 GB of VRAM for what I assume is a quantized model?).

mishu2 commented on TurboDiffusion: 100–200× Acceleration for Video Diffusion Models   github.com/thu-ml/TurboDi... · Posted by u/meander_water
mishu2 · 2 months ago
Having the ability to do real-time video generation on a single workstation GPU is mind blowing.

I'm currently hosting a video generation website, also on a single GPU (with a queue), which is also something I didn't even think possible a few years ago (my show HN from earlier today, coincidentally: https://news.ycombinator.com/item?id=46388819). Interesting times.

mishu2 commented on Ask HN: What are you working on? (April 2025)    · Posted by u/david927
mishu2 · 10 months ago
Started working on a case discussion platform for students almost two years ago. Mostly for dentistry and medicine, but it's template-based so works well for other purposes (e.g. teachers, social workers, etc.). It's going well and is being used by three universities right now.

On the way, I developed lightweight image editor and 3D model viewer components, which I've open sourced [1].

[1]: https://github.com/kigun-org/

u/mishu2

KarmaCake day247June 30, 2019
About
website: mihaitarce.com
View Original