Robot Arm Control via Cognitive Map Learners

arXiv cs.RO / 3/31/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a cognitive map learner (CML) approach to robot arm control by independently training CML modules for each arm segment.
  • Target locations in a 2D plane are encoded as phasor hypervectors using fractional power encoding (FPE), then factorized into joint angles via either a resonator network or a modern Hopfield network.
  • The resulting segment angles are used to drive each segment’s CML, moving the arm to the target point without relying on inverse kinematics equations.
  • The method is presented as a general solution for multi-segment 2D arms and includes a specific example for 3D arms with a rotating base.
  • Key novelty is compositionality: hierarchically trained CML modules can be combined to handle more complex tasks without task-specific retraining.

Abstract

Cognitive map learners (CML) have been shown to enable hierarchical, compositional machine learning. That is, interpedently trained CML modules can be arbitrarily composed together to solve more complex problems without task-specific retraining. This work applies this approach to control the movement of a multi-jointed robot arm, whereby each arm segment's angular position is governed by an independently trained CML. Operating in a 2D Cartesian plane, target points are encoded as phasor hypervectors according to fractional power encoding (FPE). This phasor hypervector is then factorized into a set of arm segment angles either via a resonator network or a modern Hopfield network. These arm segment angles are subsequently fed to their respective arm segment CMLs, which reposition the robot arm to the target point without the use of inverse kinematic equations. This work presents both a general solution for both a 2D robot arm with an arbitrary number of arm segments and a particular solution for a 3D arm with a single rotating base.