Bio­-inspired approaches to the control and modelling of\ud an anthropomimetic robot

Doctoral thesis English OPEN
Diamond, Alan (2013)
  • Subject: TJ0210.2

Introducing robots into human environments requires them to handle settings designed specifically for human size and morphology, however, large, conventional humanoid robots with stiff, high powered joint actuators pose a significant danger to humans. By contrast, “anthropomimetic” robots mimic both human morphology and internal structure; skeleton, muscles, compliance and high redundancy. Although far safer, their resultant compliant structure presents a formidable challenge to conventional control. Here we review, and seek to address, characteristic control issues of this class of robot, whilst exploiting their biomimetic nature by drawing upon biological motor control research. We derive a novel learning controller for discovering effective reaching actions created through sustained activation of one or more muscle synergies, an approach which draws upon strong, recent evidence from animal and humans studies, but is almost unexplored to date in musculoskeletal robot literature. Since the best synergies for a given robot will be unknown, we derive a deliberately simple reinforcement learning approach intended to allow their emergence, in particular those patterns which aid linearization of control. We also draw upon optimal control theories to encourage the emergence of smoother movement by incorporating signal dependent noise and trial repetition.\ud \ud In addition, we argue the utility of developing a detailed dynamic model of a complete robot and present a stable, physics-­‐‑based model, of the anthropomimetic ECCERobot,\ud running in real time with 55 muscles and 88 degrees of freedom. \ud \ud Using the model, we find that effective reaching actions can be learned which employ only two sequential motor co-­‐‑activation patterns, each controlled by just a single common driving signal. Factor analysis shows the emergent muscle co-­‐‑activations can be reconstructed to significant accuracy using weighted combinations of only 13 common fragments, labelled “candidate synergies”. Using these synergies as drivable units the same controller learns the same task both faster and better, however, other reaching tasks perform less well, proportional to dissimilarity; we therefore propose that modifications enabling emergence of a more generic set of synergies are required.\ud \ud Finally, we propose a continuous controller for the robot, based on model predictive control, incorporating our model as a predictive component for state estimation, delay-­‐‑\ud compensation and planning, including merging of the robot and sensed environment into a single model. We test the delay compensation mechanism by controlling a second copy of the model acting as a proxy for the real robot, finding that performance is significantly improved if a precise degree of compensation is applied and show how rapidly an un-­‐‑compensated controller fails as the model accuracy degrades.
  • References (34)
    34 references, page 1 of 4

    6.6.4   Predicting  intended  motor  signals  with  an  updateable    “buffer”  ............... 1 45  

    6.6.5   Delay  compensating  design  .............................................................................. 1 46  

    6.7   Experiment  exploring  delay  compensation  ........................................................... 1 48  

    6.7.3   Characterising  Effects  of  Model  Divergence  .................................................. 1 54  

    6.8   Discussion  and  Conclusion  ...................................................................................... 1 56  

    7.2   Original  Contributions  of  Thesis  ............................................................................. 1 60  

    7.2.3   A   low   dimensional   reaching   controller   for   biomimetic   musculoskeletal  

    modelled  robot    based  on  extracted  emergent  synergies . ......................................... 1 62  

    7.2.4   Optimal   control   principles   can   be   exploited   through   RL   trial   repetition   to  

    refine  movements . ........................................................................................................... 1 62  

  • Metrics
    0
    views in OpenAIRE
    0
    views in local repository
    13
    downloads in local repository

    The information is available from the following content providers:

    From Number Of Views Number Of Downloads
    Sussex Research Online - IRUS-UK 0 13
Share - Bookmark