As an enthusiast sometimes I feel that I get stuck in a corner and could use some overview. A newsforum about AI engineering could help. E.g. most articles about practical things (think civitai guides), some articles dive deeper into how this tech works (think 3b1b). Do you know a good one?
I only google things and am not e-social, so may miss something obvious like “civitai’s good though, why” or “there’s tech jesus but for AI on YT”.
Interested: AI tech, SD inference and training, local LLM models and prompting/settings, VLMs, TTSs, other models, scripts, experiments, popsci.
Disinterested: CEO said, politics, winter, alignment, eats us all, not a human, etc.
I can provide my lora training and automation experience, for the most part (around 100 sd 1.5 loras, few versions each).
https://simonwillison.net/
Dead Comment
https://www.youtube.com/@aiexplained-official
They also have a free newsletter and more behind a Patreon subscription.
https://signaltonoise.beehiiv.com/p/the-3rd-era-of-ai-langua...
https://www.patreon.com/AIExplained
Gwern is the place for me. His deep dive on meta-learning is always interesting.
For deep learning in particular, I will add Neel Nanda's interpretability work: https://www.neelnanda.io/mechanistic-interpretability
https://ym2132.github.io
I try to go for those things you're looking for. It's hard to find good resources nowadays with real people behind them.
I hope you enjoy the posts. Feel free to reach out about anything on there
Your content is amazing and with some more polish I feel like it would really shine! (Some sentences not flowing quite right is a little confusing for me, reading the GAN deep dive)
Noted I am quite new to writing like this, where exactly was it confusing? I'll definitely take this on board and try to increase the clarity in my writing.
Totally not suspicious.
But if you’re in a few discords and a bunch of subreddits, you’re doing it right.
The most interesting stuff happens in GitHub PR’s, but you have to know where to look. Kohya’s misnamed SD3 branch has a ton of good flux hints, for example. It’s also where furkan gets pretty much all his content, before it gets paywalled.
Unfortunately, unless you participate full-time it’s hard to follow along. But if you really dig in and learn to modify your tooling (Comfy, kohya etc), you’ll start to come across some really impressive people who are all self-taught, and very accessible.
It’s totally possible to work your way up to the frontier with a few months of hacking. (And disposable income for GPU time.)
And the overlap between image AI’s and LLM’s is actually pretty great since they’re all transformers under the hood.
Civit, in my experience, is a good source for weights but most of the guides are written by people without much actual experience.
If you haven’t already, use tensorflow or wandb to get an intuitive understanding of your training parameters. It’s very easy to connect your tools to these services. This is by far the most helpful thing I’ve done, and something I really regret not doing sooner.
Image
- Terminus Research Group, from bghira of SimpleTuner/diffusers https://discord.gg/cSmvcU9Me9
- AI Toolkit https://github.com/ostris/ai-toolkit https://discord.gg/VXmU2f5WEU
- Stable Diffusion https://discord.gg/stablediffusion
LLM
- LLamaIndex https://www.llamaindex.ai https://discord.com/invite/eN6D2HQ4aX
- Nous research https://discord.gg/nousresearch
- LangChain https://discord.gg/hMrfPpUk
Platforms
- Replicate https://discord.gg/replicate
- Fal https://discord.gg/fal-ai
Who/what is Furkan?
I think you meant TensorBoard?
You can find 99% of companies and reasearchers. Just follow them. If you need some names, just ask!
https://www.reddit.com/r/singularity
https://paperswithcode.com
https://www.reddit.com/r/StableDiffusion
https://www.reddit.com/r/ClaudeAI