The Thousand Brains Project: A New Paradigm for Sensorimotor Intelligence

:rotating_light: New Paper Alert! :rotating_light:

@vclay, @nleadholm and @jhawkins introduce “A New Paradigm for Sensorimotor Intelligence.” :robot::sparkles:

Explore how the principles of the neocortex establish a new type of AI with rapid, flexible learning - powered by sensorimotor learning, reference frames and modularity.

:memo: Read the paper:

10 Likes

Very exciting! Look forward to reading this tonight :grin:

3 Likes

Still reading through this. I’m just now working my way towards the end and I have some questions.

On agent types:
If vision in humans is being delivered to us by way of photons hitting the retina, wouldn’t vision also be a kind of “surface agent” in a way? Is there an advantage to differentiating between distant and surface agents, as opposed to developing a policy that can respond to both types of action spaces equally well?

On Policies:
I really like your approach with policies. The hypothesis-testing policy is particularly interesting to me. It almost strikes me as analogous to something like the Central Exec Network (CEN) in the brain. I wonder if it couldn’t decompose complex task-states, being those could be viewed almost as a kind of abstract space in and of themselves? (you seem to hint at working towards this behavior a little later in the reading).

In relation to the CEN, have you guys entertained emulating something like the DMN? I know discussions on the default mode network can be a bit contentious among neuroscientists, but the brain is obviously doing something while at rest. My thought is that the DMN is serving as a kind of resting-state manifold learner, mapping the complex, low-dimensional structure underlying high-dimensional brain activity. Would there be benefit to enabling something like this in the TBP framework?

Lastly, the way you describe using Euclidean distances to plot graph points almost reminds me of EC-hippocampal path encoding. It sounds like you guys are working to constrain much of the brains overall functioning to these SM/LM units. Is that the case? Would you essentially be looking to emulate things like the hippocampus and cerebellum over broader, distributed sensory-motor space? If so, do you not worry about running into issues of space complexity, similar to how transformers run into issues with arbitrarily long input sequence lengths?

Thanks for the questions.

Re. agent types

Re. policies

Re. the default mode network

  • That’s an interesting question, I don’t think it’s something we’ve considered too much. An initial thought is that something like this might emerge naturally in a large-scale Monty system when attending primarily to internally generated information (as opposed to external sensory information). Like you mention, higher cognitive functions require that the brain learn a model of itself, and so attending to internal information would probably be a natural basis for this network.

Re. Euclidean distance and computational complexity

  • Let me know if the discussion in the other thread has helped with this, but otherwise happy to discuss it more.
1 Like

Hey @nleadholm, thanks for all these replies. You’ve been busy today!

Re. Euclidean distance and computational complexity

  • Nah. I think we’ve pretty much covered everything in that other thread. No need to rehash it here.

Re. the default mode network

  • That’s an interesting thought. I’m going to try and probe some of the other members’ views on this. That said, if it is an emergent function, I’d be curious to know it becomes constrained to “task-agnostic” cortical areas only (think frontal lobe, PFC, et cetera). As opposed to cortical areas which receive constant externally-derived throughput (e.g. occipital lobe). Also, given the anti-correlated relationship between the CEN and DMN, I would suspect they share some common principle in their functioning. There’s a niggling part of my brain that keeps wanting to explore concepts of memory equivalent capacities (similar to what’s explained here: https://www.youtube.com/watch?v=nWZhgWBdgQQ), but IDK yet. I’ll need to do some thinking…

Re. agent types & policies

  • Yes! I was actually wanting to explore approaches to goal decomposition. This link is great! :smiley:

  • In regards to the agent types, I’ll need give that paper a read. My intuition says the LM should be adaptable enough to adjust to wildly different sensor-types. That ultimately those sensors should be the things responsible for dictating the resulting morphology of the LM’s overall structure (e.g. distant vs. surface agent types), not an architectural design decision bespoke to the LM itself. That said, this is 100% a me problem. Let me give that paper a read and get back to you here.

Edit: Please ignore my comment regarding agent types. Apparently I’m a dummy. I must have misconstrued something and mixed up learning modules for sensor modules. Turns out you were talking about actors in reference to the latter, which I agree with. My bad. :stuck_out_tongue:

Edit 2: I’m not sure how feasible this would be from an engineering standpoint, but how beneficial would it be for you to emulate something like Merkel disk receptors? I would imagine it’d allow for easier form perceptions of grasped objects’. Would that make processing easier for the downstream LM’s, you think?

Edit 3: Okay, so the way you’re approaching agents in the SM space makes a ton of sense. Essentially, you have retinotopic mappers and non-retinotopic mappers (e.g. tactile mapping via touch sensor), whereby the former is providing allocentric reference frames and the latter egocentric. You’re then performing transforms between the two in LM space in order to achieve things like object invariance, and the like. Then your have your motor policies (essentially sensorimotor schema) downstream of all
of this. Is this close to what’s actually going on in the TBP framework?

3 Likes

In working my way through the paper (which is FAR more readable than most I’ve encountered!), I’ve noticed a few typos and such. What is the proper way to report these?

2 Likes

If you can write them here I think that would be fine. Or if you’d prefer you can email info@thousandbrains.org.

1 Like

One of the capabilities listed in the paper caught my eye:

Recognizing an object and its pose by moving one sensor over the object.

I have no problem with this as a general goal for this stage of the development effort, but in most cases many sensors (and a lot of coordination) will be needed.

Even if we’re just talking about a fingertip moving along the surface of a cup, there are zillions of nerve cells in the skin and the musculature of the arm that need to be involved.

Or, consider some more real-world examples:

  • Hmmm; that sounds like a siren; where is it coming from?

  • Ouch; I have a sharp pain in my foot. Oh yeah, a stone got into my shoe.

It will be fascinating to find out how the motors, sensors, planning, and overall coordination need to play out.

Thanks for the follow-up questions.

@mthiboust yes that’s a good description of how we’re approaching distant vs. surface agents. Re. Merkel disk receptors, yes something like this would definitely be useful. This would then become a modality-specific feature input, similar to color for Sensor Modules that detect light.

@Rich_Morin Thank you for the suggested edits. And yes re. your question: To clarify, a single sensor-module is more than a single neuron - it’s more akin to a bundle of many neurons that capture sensory information in a patch. We stress the narrow receptive field to distinguish it from the full, high-resolution image typically fed into deep neural networks.

The other thing to highlight is that perception will be very slow and laborious with only one SMs (as recognizing the world through a straw or with a single finger tip is!), however as you say this has been our focus as part of our developmental stage, in this case to ensure we have the functionality of a single Learning Module correctly implemented. Hope that makes sense.

2 Likes