Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Do any of the open weight models from smaller labs exist if they can't distill from the SoTA models that are throwing billions of dollars of compute into pretraining?
 help



I’ve been wondering the same. And I think pretty much all the impressive small lab models were guilty of it, right? At least there is still larger players like DeepSeek and mistral to provide a bit of diversity in the market

Does it matter? The frontier models stole the whole internet, then the second-level models stole from them… It’s all theft.

Hard agree.

The question is - if the SOTA model disappear - do these follow-on models have the ability to improve themselves without distillation?

[flagged]


“Very likely yes”, I reply to an account that <1yr old with mostly comments in AI topics many of which violate the HN guidelines (including the one I’m responding to).

Strange gatekeeping response. Yep i comment on topics i'm interested in. Forgive me for not being on the platform for more than a year yet. That's a cute attitude

> The frontier models stole the whole internet

What does that even mean?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: