Hello, I am a beginner just starting to work with Monty, and I have a question regarding the Learning Module Outputs. I noticed that there isn’t an explicit motivation, and I am curious if this implies that the “motion” sensors are mainly used for improving the predictive model. Does this mean that the prediction motion stops when the model and actual data align? In this case, is the “building of the model” the motivation behind Monty?
I also noticed a question regarding “how to handle moving objects,” and I was wondering if the “motion” sensors should be separated from the “prediction model” to better handle and train the model. This separation would allow Monty to distinguish between movements caused by time and movements caused by motion itself, making it more capable of handling moving objects.
For example, if I input the position of the end-effector of a robotic arm as input and the six joint angles as output, the “motivation” should be to move the end-effector to a specified target point (not in a reinforcement learning sense, just as a task). Through this motion, Monty would establish the relationship between the joint angles and the end-effector position (possibly similar to the Jacobian matrix, as the relationship is linear, but it could be something else). As Monty using its model to realized its goal, he will updates its model by detecting the discrepancy between the predicted action result and the actual behavior, it would refine this model.
I think this might be achievable, and it is easy to simulate the robotic arm in ROS, next work might be use Monty to learn this relationship. The next step could involve Monty learning the dynamics of moving objects, like when a ball is moving in relation to the end-effector. Monty would need to adjust the joint angles to keep the ball stationary in space. Finally, Monty could learn more abstract nonlinear relationships, like controlling a large magnet that attracts small magnets to specific positions.
I have this question because, in the Thousand Brains theory, motivation is provided by the old brain. The old brain drives behavior based on biological needs like survival and reproduction. I am wondering whether Monty, as a model, would also need some form of “motivation” to guide its learning process and drive its model updates. This could also be important for evaluating the success of the learned world model.