Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I'm reminded of the story of Helen Keller, and how it took a long time for her to realize that the symbols her teacher was signing into her hand had meaning, as she was blind and deaf and only experienced the world via touch and smell. She didn't get it until her teacher spelled the word "water" as water from a pump was flowing over her hand. In other words, a multimodal experience. If the model only sees text, it can appear to be brilliant but is missing a lot. If it's also fed other channels, if it can (maybe just virtually) move around, if it can interact, the way babies do, learning about gravity by dropping things and so forth, it seems that there's lots more possibility to understand the world, not just to predict what someone will type next on the Internet.


It is important to note that Helen Keller was not born blind and deaf, though. (I am not reducing the struggle she went through. Just commentary on embodied cognition and learning.) There were around 19 months of normal speech and hearing development until then and also 3D object space traversal and object manipulation.


at least a few decades ago, this idea was called "embodied intelligence" or "embodied cognition". just FYI.


Enactivist philosophy. Karl Friston is testing this approach as CTO of an AI startup in LA.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: