Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

That has me wondering now.

It's absolutely true that children learn (and even generate) language grammar from a ridiculously small number of samples compared to LLMs.

But could the availability of a world model, in the form of other sensory inputs, contribute to that capacity? Younger children who haven't fully mastered correct grammar are still able to communicate more sensibly than earlier LLMs, whereas the earlier LLMs tend toward more grammatically correct gibberish. What if the missing secret sauce to better LLM training is figuring out how to wire, say, image recognition into the training process?



It amuses me that this would be not unlike teaching an LLM with picture books.




Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: