Readit News logoReadit News
sumo43 commented on Tongyi DeepResearch – open-source 30B MoE Model that rivals OpenAI DeepResearch   tongyi-agent.github.io/bl... · Posted by u/meander_water
aliljet · 2 months ago
Sunday morning, and I find myself wondering how the engineering tinkerer is supposed to best self-host these models? I'd love to load this up on the old 2080ti with 128gb of vram and play, even slowly. I'm curious what the current recommendation on that path looks like.

Constraints are the fun part here. I know this isn't the 8x Blackwell Lamborghini, that's the point. :)

sumo43 · 2 months ago
sumo43 commented on Tongyi DeepResearch – open-source 30B MoE Model that rivals OpenAI DeepResearch   tongyi-agent.github.io/bl... · Posted by u/meander_water
sumo43 · 2 months ago
I made a 4B Qwen3 distill of this model (and a synthetic dataset created with it) a while back. Both can be found here: https://huggingface.co/flashresearch
sumo43 commented on Π0: Our First Generalist Policy   physicalintelligence.comp... · Posted by u/krishpop
golol · a year ago
Isn't this a real big deal? They manage to train a foundation model that connects the physical understanding gained from unsupervised training on images and videos with the real physical understanding necessary to control a robot. They have impressive videos to show for it. The approach seems indeed scalable and generalist. I've thought for a while that the keystone missing for household androids is connecting the understanding of large multimodal language models with physical hardware. To me this looks like exactly that. It actually makes me optimistic that we will see household robots within a decade. Now I wonder why Tesla, Figure etc. are messing around so much with Teleoperation if this indeed works. Maybe I don't understand what's going on.
sumo43 · a year ago
I think the fine tuned policies are still very brittle, but I agree that this is super promising. It's also one of the most open (the model is still closed) research blogposts we've seen from any private embodied AI lab
sumo43 commented on Octo: An Open-Source Generalist Robot Policy   octo-models.github.io/... · Posted by u/jasondavies
sumo43 · 2 years ago
seems like an improvement on the aloha approach? You still need to finetune it on roughly the same amount of OOD examples. Contrast this with google's approach over 2023, which was training large vision-language models with the goal of generalizing on OOD.
sumo43 commented on Ask HN: Who wants to be hired? (May 2024)    · Posted by u/whoishiring
sumo43 · 2 years ago
Location: US

Remote: Yes

Willing to relocate: Yes (US)

Technologies: Python, PyTorch, HuggingFace, C++

Résumé/CV: https://drive.google.com/file/d/1qY-m1tKz4_QpHgxaGryC2vk-DGs...

Email: sumo43@proton.me

ML Engineer & Research Scientist. Ex AI grant startup, hedge fund. I've previously worked on inference for LLMs and vision models & have experience with data curation and multinode training. Looking for summer internships or part time positions

sumo43 commented on DBRX: A new open LLM   databricks.com/blog/intro... · Posted by u/jasondavies
n2d4 · 2 years ago
There's at least an argument to be made that this is because all the models are heavily trained on GPT-4 outputs (or whatever the SOTA happens to be during training). All those models are, in a way, a product of inbreeding.
sumo43 · 2 years ago
Maybe true for instruct, but pretraining datasets do not usually contain GPT-4 outputs. So the base model does not rely on GPT-4 in any way.
sumo43 commented on Ask HN: Looking for a project to volunteer on? (February 2024)    · Posted by u/theycallhermax
sumo43 · 2 years ago
SEEKING VOLUNTEERS: open source self-play training for language models

we are a small team associated with EleutherAI. looking to push the frontier of open source language models through self-play. so far we have implemented SPIN. compute included.

email tyoma9k@gmail.com

edit: formatting

u/sumo43

KarmaCake day216September 9, 2023
About
Interested in LLMs that browse the web natively. you can reach me at artem@lmresearch.net
View Original