Jul 27, 2015 - Ganguly, Carmena 2011: Reversible large-scale modification of cortical networks during neuroprosthetic control

Reversible large-scale modification of cortical networks during neuroprosthetic control

Monitored ensembles of neurons that were either casually linked to BMI control or indirectly invovled, found that proficient neuroprosthetic control is associated with large-scale modifications to the cortical network. [...] Specifically, there were changes in the preferred direction of both direct and indirect neurons. Notably, with learning, there was a relative decrease in the net modulation of indirect neural activity in comparison with direct activity. [...] Thus, the process of learning BMI control is associated with differential modification of neural populations based on their specific relation to movement control.

Results pretty intuitive, but analysis methods are useful to know.

Experiment
Recorded ensembles of M1 neurons while only a subset were assigned to have a causal role during control as direct neurons. The monkyes performed center-out reaching movements using a exoskeleton that constrained movements to the horizontal plane.

Used a linear decoder for motor commands. Decoder was held constant after initial training. Stability of recordings across days assessed by measuring the stationarity of spike waveforms and the interspike interval distribution, as well as directional modulation during manual control sessions.

Results

  1. Modification of preferred directions
    There is a relative remapping of the preferred directions for all neurons without any substantial systematic rotational shifts for each neural opulation.

  2. Differential modification of modulation depths
    Direct and indirect neurons had different tuning depth. This differential modulation was specifically present during proficient neuroprosthetic control and not during the initial learning period. At population level, no significant systematic differences in the mean firing rate between manual control and brain control for either populations.

  3. Reversibility of modifications
    Modulation depths for the manul-BMI-manul experimental scheme changes, while being similar during the manual sessions. This suggests reversibility in the large-scale modifications dependent on cortical state (manual vs. BMI_

  4. Stability
    Indirect neurons maintained a relatively fixed neuron-behavior relationship during brain-control through sessions across days. May suggest an active role to brain-control? Probably related to the irrotational shift of tuning properties.

Used bootstrap resampling methods for testing significance in directional tuning and mean modulation depth changes.

Jul 27, 2015 - Sanchez 2013 - Towards autonomous neuroprosthetic control using Hebbian reinforcement learing

Towards autonomous neuroprosthetic control using Hebbian reinforcement learing

Theory paper toward which the Sanchez guys have been working toward. The RLBMI decoder architecture is now presented as an actor-critic method of RL, where the neuroprosthetic controller (actor) learned the neural state-to-action mapping based on the user's evaluative feedback. The role of teh critic is to translate the user's feedback into an explicity training signal taht can be used by the factor for adaptation. The aim is to learn and automatically produce a stable neural to motor mapping and respond to pertubations by readjusting its parameters in order to maintian the performance, using a binary evaluative feedback. The metrics for the controller's performances used were:

  • speed of convergence.
  • generalization
  • accuracy and recovery from perturbation

Reinforcement Learning, control architecture, learning algorithm - Instead of using Q-learning with a NN to act as the action-reward mapping, a Hebbian reinforcement learning framework is used. I'm confused as to why this group decided to switch the RL-framework. I am not very familiar with all the ML techniques yet, but according to scholarpedia, we need to distinguish between the machine learning/TD-learning [...] and the neuronal perspective. The machine learning perspective deals with states, values, and actions, etc., whereas the neuronal perspectives tries to obtain neuronal signals related to reward-expectation or prediction-error. Regardless, the basic strategy is the same: initialize a random decoder, based on some feedback on whether the decoded action is correct or not, update the NN weights. With more trials, both the decoder and the user adapt toward each other.

Tasks

  • The sequential mode is where the controller has to perfrom a sequence of actions over mulitple steps in order to accomplish the goal of a trial. This means that there can be multiple sequences of actions that could eventually result in the end goal, and thus mulitple solutions that could be learned by the HRL. Reward/feedback is received by the controller after every action. The metric for reward in movement given in this paper is whether a specific time-step the action moves the manipulandum closer towards the targe.

    But I can imagine this approach suffering from the problem common to some greedy algorithms - sometimes the greedy approach might not give the optimal solution. Also possible that this intention estimation is too constraining - the animal might not actually want to move in the most direct trajectory possible, but this may not be important in terms of BMI for patients, efficiency is probably the most important factor there.

  • The episodic mode where the controller can select only one action in each trial, and thus that action must achieve the goal for a successful trial. The episoic learning mode is suitable for classification tasks. This is what Sanchez 2011 used for their center-out task. Since achieving the trial goals requires only a signel action here, the HRL can be encouraged to learn more quickly in the episodic task compared to the sequential task. Experience replay can be used to seed learning in this paradigm.

    Since the sequential mode is a sequence of actions, using episodic mode to intitialize the decoders for sequential mode can probably speed up the learning process.

Network was first tested in simulation experiments, developed for a reaching task in 2D grid space, useful for validating the actual computational model. Three components:

  1. Synthetic neural data generator, with each neuron tuned to one action (moving left, right, up, and down) or none. Goal of the user is to reach targets in 2D space. Input feature is firing rates in 100ms bins.
  2. Neuroprosthetic controller (actor).
  3. Behavioral paradigm - gives binary feedback to the controller.

Experiments were done in common-marmosets, utilizing a Go/No-Go motor task to move a robot arm to spatial targets to either the left or right side of the monkey. Catch Trial - traning technique to ensure the monkeys understood the necessity of the robot movements where the robot moved in the opposite direction commanded by the monkey and thus the monkey received no reward.

Results

Simulation results showed the the weights converging to achieve high success rates for both tasks, as expected. Robust against neuron loss and tuning changes.

The monkey neural data were used offline to map to four different directions. Complete BS, the analysis makes no sense since the monkey was simply trained to perform Go/No-go arm movement. This is bad.

References to Read

  • Francis J T and Song W 2011 Neuroplasticity of the sensorimotor cortex during learning Neural Plast. 2011 310737
  • Heliot R et al 2010 Learning in closed-loop brain-machine interfaces: modeling and experimental validation IEEE Trans. Syst. Man Cybern. B 40 1387–97
  • Prasad A et al 2012 Comprehensive characterization and failure modes of tungsten microwire arrays in chronic neural implants J. Neural Eng. 9 056015
  • Polikov V S, Tresco P A and Reichert W M 2005 Response of brain tissue to chronically implanted neural electrodes J. Neurosci. Methods 148 1–18
  • Mahmoudi B and Sanchez J C 2011 A symbiotic brain-machine interface through value-based decision making PLoS One 6 e14760
  • Schultz W, Tremblay L and Hollerman J R 1998 Reward prediction in primate basal ganglia and frontal cortex Neuropharmacology 37 421–9
  • Izawa E I, Aoki N and Matsushima T 2005 Neural correlates of the proximity and quantity of anticipated food rewards in the ventral striatum of domestic chicks Eur. J. Neurosci. 22 1502–12
  • Wawrzyński P 2009 Real-time reinforcement learning by sequential actor–critics and experience replay Neural Netw. 22 1484–97
  • Schultz W 2000 Multiple reward signals in the brain Nature Rev. Neurosci. 1 199–207
  • Prins N et al 2013 Feature extraction and unsupervised classification of neural population reward signals for reinforcement based BMI Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. (EMBC) (Osaka, Japan, 2013) at press
  • Ribas-Fernandes J J F et al 2011 A neural signature of hierarchical reinforcement learning Neuron 71 370–9

Jul 27, 2015 - Sanchez 2014 - Using Reinforcement Learning to Provide Stable Brain-Machine Interface Control Despite Neural Input Reorganization

Using Reinforcement Learning to Provide Stable Brain-Machine Interface Control Despite Neural Input Reorganization

Simple binary BMI-control demonstration with the RLBMI framework. Finally showed some neural data, and showed that the RLBMI approach is robust against neuron loss. They are calling their RL neural decoder in Sanchez 2013 "Associative reinforcement learning that combined elements of supervised learning with reinforcment based optimization".

Since the experiment control is so simplistic, cannot derive much value about this decoder's performance compared to other closed-loop based methods. But

The RLBMI maintained high performance when applied in a contiguous fashion across experiment sessions spanning up to 17 days. The decoder weights started from random initial conditiosn during the first session, and during subsequent sessions the system was intialized from wegiths learning in the previous session, and was then allowed to adapt as usual without any new initializations or interventions. [...] Half of the neural input signals were lost between day 9 and 16. However, the system was able to quickly adapt and this loss resulted in only a slight dip in performance.

References to Read

  • Ludwig KA, Miriani RM, Langhals NB, Joseph MD, Anderson DJ, et al. (2009) Using a common average reference to improve cortical neuron recordings from microelectrode arrays. J Neurophysiol 101: 1679–1689. doi: 10.1152/jn.90989.2008
  • Schultz W (2000) Multiple reward signals in the brain. Nat Rev Neurosci 1: 199–207. doi: 10.1038/35044563

Jul 26, 2015 - Deep Learning - Review by LeCun, Bengio, and Hinton

Nature review on Deep Learning by LeCun, Bengio, and Hinton

Representational Learning is a set of methods that allows a machine to be fed with raw data and automatically discover the representations needed for detection or classification. Deep-learning methods are repsentation-learning methods with multiple levels of representation, obtained by composing simple but non-linear modules that each transform teh representation at one level into a representation at a higher, slightly more abstract level. With the composition of enough such transformations, very complex functions can be learned.

Key advantage of Deep Learning is that it requires very little engineering by hand, so it can easily take advantage of increases in the amount of availabel computation and data. Feature extraction becomes easier. The number of nodes determine what kind of input-space transformation is possible, and there can classify data that otherwise cannot using lower-dimension techniques.

Interesting historical fact: in the late 1990s, neural nets and backpropagation were largely forsaken by the community. It was widely thought that learning useful, multistage, feature extractors with little prior knowledge was infeasible. In particular, it was commonly thought that simple gradient descent would get trapped in poor local minima.

In practice, however, poor local minima are rarely a problem with large networks. Regardless of the initial conditions, the system nearly always reaches solutions of very similar quality. Theoretical and empirical results suggest that the landscape is packed with a combinatorially large number of saddle points where the gradient is zero, and the surface curves up in most dimensions and down in the remainder [...] saddle points with only a few downward curving directions are present in very large numbers, but almost all of them have very similar values of the objective function. Hence, it does not much matter which of these saddle points the algorithm gets stuck at.

Convolutional neural networks (convNet) - four key ideas:

  • local connections: in array data, local groups of values are often highly correlated, forming distinctive local motifs that are easily detected; the local statistics of images and other signals are invariant to location.
  • shared weights: If a motif can appear in one part of the image, it could appear anywhere, hence the idea of units at different locations sharing the same weights and detecting the same pattern in different parts of the array.
  • pooling: merrge semantically similar features into one. Many natural signals are compositional hierarchies, in which higher-level features are otained by compoising lower-level ones.
  • use of many layers.

Distributed representations. two different exponential advantages over classic learning algorithms that do not use distributed representations - both arise from the power of composition and depend on the underying data-generating distribution having an appropriate compnential structure.

  • Learning distributed representations enable generalization to new combinations of the values of learned features beyond those seen during training (can be very useful BMI).
  • Composing layers of representation in a deep net brings the potential for another exponential advantage (not sure what it means).

Recurrent neural networks for tasks that involve sequential inputs. Most likely useufl for BMI. Can be augmented with an explicity memory, e.g. long short-term memory (LSTM) that use special hidden units, the natural behavior of which is to remember inputs for a long time.

Much progress shold come with systems that train end-to-end and combine ConvNets with RNNs that use reinforcment learning to decide where to look.

References to Read

  • Bottou, L. & Bousquet, O. The tradeoffs of large scale learning. In Proc. Advances in Neural Information Processing Systems 20 161–168 (2007).
  • Hinton, G. E. What kind of graphical model is the brain? In Proc. 19th International Joint Conference on Artificial intelligence 1765–1775 (2005).
  • Hinton, G. E., Osindero, S. & Teh, Y.-W. A fast learning algorithm for deep belief nets. Neural Comp. 18, 1527–1554 (2006).
    This paper introduced a novel and effective way of training very deep neural networks by pre-training one hidden layer at a time using the unsupervised learning procedure for restricted Boltzmann machines.
  • Cadieu, C. F. et al. Deep neural networks rival the representation of primate it cortex for core visual object recognition. PLoS Comp. Biol. 10, e1003963 (2014).
  • Farabet, C. et al. Large-scale FPGA-based convolutional networks. In Scaling up Machine Learning: Parallel and Distributed Approaches (eds Bekkerman, R., Bilenko, M. & Langford, J.) 399–419 (Cambridge Univ. Press, 2011).
  • Weston, J. Chopra, S. & Bordes, A. Memory networks. http://arxiv.org/abs/1410.3916 (2014).

Jul 26, 2015 - DiGiovanna 2009: Coadaptive Brain-Machine Interface via Reinforcement Learning

Coadaptive Brain-Machine Interface via Reinforcement Learning

This is one of the very first papers that uses a reinforcement learning (RL) framework for BMI decoding. The technique used is semi-supervised because only a scalar (in this case, binary) reward signal is provided after tasks. This is markedly different from the more traditional supervised learning (SL) approach to decoding that uses kinematic variables as desired signal to train a regression model, etc.

The authors claim that the RLBMI architecture involves two coupled systems - while the user learns to use the BMI controller through neural adaptation, the BMI controller learns to adapt to the user via RL. While in theory this sounds promising, not much neural data is presented to show the neural-adaptation aspect of the architecture.

Computational Architecture: Q-learning. To the controller, environment=User's Brain, State=neural activity, actions=prosthetic movement, rewards=task complete.

Experiment setup used

image 1

Experiment protocol

image 2

Goal is for the rat to move the robot arm to the other end of the room, to the lever that's lit up. During brain-control, both the rat and controller will be rewarded when the arm is naeuvered proximal to the target. So distance to goal is used as a reward metric. This is intention-estimate, which is what the closed-loop decoder adaptation (CLDA) approaches Carmena's group use.

In this RLBMI architecture, the value function estimation (VFE) is a non-trivial task. The value function Q is too big to be stored in a lookup-table, since while the total number of actions (robot movements) is 27, the number of possible states (neural vector configurations) is intractable. Thus a fully connected neural network is used, with a single layer of hidden units. Updated with *temporal difference (TD) error via backpropagation.

Weights were initialized to random. Goal for the BMI is some big radius within the goal. As training continued, the radius becomes smaller and smaller until it contains just the goal.

Neural data analysis shows rats were biased toward using a subset of the available robot actions, which moves the arm to target with not the most direct trajectories for all targets. Hidden layer feature representations should be analyzed to see how this happened, and how much of this is contributed by neural adapation vs. decoder adaptation.

Problems: RL-Deep learning is usually trained with large-batch of offline simulation data to speed up learning the value function. In the paper, the available data were reused in multiple-epoch, offline VFE training. Suggested using model-based RL that includes an environemental model to estimate future states and rewards...but this sounds just like Kalman filters with adaptive elements. Finally, rewards were prorammed by the BMI designer, but ideally they should be translated from the user's brain activity -- either the cortex or maybe basal ganglia.

Image cited: DiGiovanna, J.; Mahmoudi, B.; Fortes, J.; Principe, J.C.; Sanchez, J.C., "Coadaptive Brain–Machine Interface via Reinforcement Learning," Biomedical Engineering, IEEE Transactions on , vol.56, no.1, pp.54,64, Jan. 2009. doi: 10.1109/TBME.2008.926699

References to Read

  • R. S. Sutton and A. G. Barto. Reinforcement Learning: An Introduction, 1998, MIT Press
  • J. K. Chapin , K. A. Moxon , R. S. Markowitz and M. Nicolelis, "Real-time control of a robot arm using simultaneously recorded neurons in the motor cortex", Nat. Neurosci., vol. 2, pp. 664-670, 1999

Jul 26, 2015 - Sanchez 2011: Control of a Center-Out Reaching Task Using a Reinforcement Learning Brain-Machine Interface

Control of a Center-Out Reaching Task Using a Reinforcement Learning Brain-Machine Interface

I am a bit puzzled why this paper is published in 2011, after the DiGiovanna 2009, when the experiment less complicated, but with the same premise. I speculate it might be this paper demonstrates RLBMI experiments in rhesus monkeys, instead of rats.

From the abstract: Neural recordings obtained from the primary motor cortex were used to adapt a decoder using only sequences of neuronal activation and reinforced interaction withe the environment. From a naive state, the system was able to achieve 100% of the targets (in a center-out reaching task) without any a priori knowledge of the correct neural-to-motor mapping. Results show that the coupling of motor and reward information in an adaptive BMI decoder has the potential to create more realistic and functional models necessary for future BMI control.

Experiment: A single monkey trained to perform a center-out reaching task (to two targets in one trial, sequentially) with arm attached to exoskeleton arm (presumable so that during brain-control, the exoskeleton may preven the arm from moving. Unfortunately nothing about arm movements during brain-control were described). Monkey implanted in S1, M1, and PMd representing the right shoulder and elbow regions with Utah arrays (450um inter-electrode spacing). From 96 channels, sort between 190-240 units.

Decoder: Reinforcement learning with a multilayer perceptron neural network (MLP). Adaptation is focused on maximizing rewards through successful completion of the trials by the agent. The agent/BMI controller modeled its cursor control probelm as a Markov Decision Process (MDP), characterized by neural modulation as state s (neural data corresponding to all the units) and discrete movements performed by the RL agent as actions a - (in the experiment simply as one of the 8 directions in which the targets can be). Each action in a state will change the state of the environment with a certain probability - the transition probability. The agent expects a reward r when taking an action given a state. Q-learning is used to approximate this reward function, and the MLP is used to map state-action pairs to their expected reward values.

Pa_ss=Pr{s_t+1=ss|s_t=s,at=a}Ra_ss=E{r_t+1|s_t=s,a_t=a,s_t+1=s}Q(s_t,a_t)Q(s_t,a_t)+α(r_t+1+γmax_a(Q(s_t+1,a)Q(s_t,a_t)))

Results: Performance evolves and improves over time from the ranodmized initialization state. Accuracy around 97%. But this task is easily done with a simple logistic regression classifier. They make the distinction that the decoder here does not require an a priori training signal, as the feedback is whether the action was correct. While this demonstration is not so impressive, under more complex experiments, RL can potentially be much more useful than supervised learning method.

But again, no neural data is shown.

References to Read (not too urgent)

  • K. V. Shenoy, D. Meeker, S. Cao, S. A. Kureshi, B. Pesaran, C. A. Buneo, A. P. Batista, P. P. Mitra, J. W. Burdick, and R. A. Andersen, "Neural prosthetic control signals from plan activity", NeuroReport, vol. 14, pp. 591-597, 2003.
  • Y. Gao, M. J. Black, E. Bienenstock, W. Wu, and J. P. Donoghue, "A quantitative comparison of linear and non-linear models of motor cortical activity for the encoding and decoding of arm motions", in The 1st International IEEE EMBS Conference on Neural Engineering, Capri, Italy, 2003.

Jul 13, 2015 - KS Ch. 33: The organization and planning of movement

Movement error/variability is proportional with velocity and force.

The brain's choice of spatial coordinate system depends on the task. This can sometimes be determined by:

*Plotting the movement errors along the different components of different suspected coordinate systems.
*A likely coordinate system would result in uncorrelated errors in its principle axis/components.
*This is similar to eigenvector analysis - decompose the movements into uncorrelated components, thes components/eigenvectors is then the used coordinate system.

Examples used include [Gordon, Ghilardi, and Ghez 1994] and [Soechting and Flanders 1989].

Fitt's law describes the speed-accuracy trade-off, roughly log-inverse[Jeannerod 1988].

Stereotypical patterns are employed in many movements. Tendency to make straight-line movements characterizes a large class of movements, regardless of the motions of the joints required. Joint motions often vary while hand-trajectory reamin more invariant, also suggests planning with respect to hand [Morasso 1981].

Two-thrids Power Law - the relationship between the speed of hand motion and the degree of curvature of the hand path is roughly constant: velocity varies as a continuous function of the curvature raised to the power of two-thrids [Lacquaniti, Terzuolo, and Viviani 1983].

Feedback control cannot generate a command in anticipation of an error: It is always driven by an error. Feedforward control is based only on a desired/expected state and can therefore null the error. Most likely used to initiate an action, followed by feedback error correction.

Feedback control suffers from sensory delay. Feedforward control suffers from inaccurate estimates. Therefore, movement controls uses a combination of sensory feedback and motor prediction, using a forward model to estimate the current state of the body.

Sensory processing is different for action and perception - sensory information used to control actions is processed in neural pathways that are distinct from the afferent pathways that contribute to perception. Key points: visual information flows in two streams in the brain. Dorsal stream projects to the posterior parietal cortex, involved in use of vision for action. Ventral stream projects to the inferotemporal cortex and invovled in conscious visual perception.

Evidence for motor learning in reaching experiment: In [Brashers-Krug, Shadmehr, and Bizzi 1996], person's arm holding an apparatus reaches for targets. The apparatus then applies a CCW force against the user, disturbing his otherwise straight trajectories. The user eventually adapts and forms relatively straight lines again.
Two possible learning strategies are possible: 1) Stiffening arms to resist force; 2) Anticipate and learn a new internal model to compensate for the new forces. After turning force off, we see overcompensation in the trajectories, indicating that learning strategy 2 was used.

Dynamic motor task learning mainly through prioception and less vision. Kinematic motor task can be guided more by vision. Proprioception is critical for planning hand trajectories and controlling dynamics, is needed to update both inverse models ued to control movement and forward models used to estimate boyd positions resulting from motor commands. Derived from experiments comparing control vs. those that have lost proprioception [Ghez, Gordon, and Ghilardi 1995],[Sainburg et al., 1995].

Jul 13, 2015 - KS Ch. 37: Voluntary Movement - The Primary Motor Cortex

Focus on the control of voluntary movements of the hand and arm in primates. Cortical networks that control voluntary movement, the role of the primary motor cortex in the generation of motor commands.

Control of voluntary movement involves more than generating a particular pattern of msucle activity - involves sensory, perceptual, ad cognitive processes not rigidly compartmentalized in neural structures.

Woosley and Penfield: recognition of the motor cortex (rostral to the central sulcus) and the "cortical homunculus". Area of arm and hand concentrated in the "Fundus".

Naive understanding of voluntary movements says that voluntary motor control appears to be strictly serial processing, with only the neurons related to the last processing stage connecting to the spinal cord. This is not correct, as the brain does not even have a single, unified perceptual representation of the world.

Two main areas: Primary Motor Cortex and Premotor Cortex, which lies directly rostral of primary motor cortex. The medial part of the premotor cortex is the Supplementary motor area. However, these three areas can further be functionally organized into more areas, especially the parts of the premotor cortex.

The supplementary motor area, dorsal premotor cortex (PMd), and ventral premotor cortex (PMv) have somatotopically reciprocal connections with the primary motor cortex and with each other. Those and the primary motor cortex (M1) also have somatotopically inputs from the primary somatosensory cortex and the rostral parietal cortex (sensory areas).

Pre-supplementary and pre-dorsal premotor areas do not project to the primary motor cortex or anything more rostral. They receive higher-order cognitive information through the prefrontal cortex.

Several cortical regions project in multiple parallel tracts to subcortical areas of teh brain as well as the spinal cord. Therefore the theory of the primary motor cortex as the "final common path" from the cortex to spinal cord is incorrect, and multiple cortical regions contribute to voluntary movements.

Corticomotoneurons are corticospinal axons that extend into the ventral horn of the spinal cord and contact the spinal motor neurons. The axons of these neurons become a bigger part of the corticospinal tract moving higher in primate phylogeny. This may explain why lesions of the primary motor cortex have bigger effect on motor control in humans compared to lower mammals. Pyramidal tract neurons is the aggregate of upper motor neuron nerve fibers that travel from the cortex and terminate either in the brainstem or spinal cord. Nerve fibers usually descend down the brain in columns.

Motor commands are population encodings (Georgopolos studies) - "further studies have confirmed that similar poopulation-coding mechnisms are used in all cortical motor areas".

The motor cortex encodes both the kinematics and kinetics of movement. Experiments in which a load is applied to either oppose or assist some arm motion found the population and single neuron activity either increased or decreased accordingly, corresponding to increased or decreased muscle activity, confirming kinetics encoding. Studies in which the activity of some corticomotoneurons does not always correlate with the contraction of their target muscles, but instead correlate with carefully controlled or powerful movements hint that they may also encode kinematics. Signals about both the desired kinematics and required kinetics of movements may be generated simultaneously in different, or possibly even overlapping, populations of primary motor cortex neurons.

Hand and finger movements are directly controlled by the motor cortex. Specifically, cortical neurons controlling the hand and digits occupy the large central core of the pirmary motor cortex motor maps but also overlap extensively with populations of neurons controlling more proximial parts of the arm. We can imagine mapping the movements of the hand and digits into a component neuron space, where each neuron controls a combination of muscle activations. This is contrasted by the highly ordered representation of tactile sensory inputs from different parts of the hand and digits in the somatosensory cortex.

The motor map is dynamic and adaptable, and can experience functional reorganization. Learning a motor skill can induce reorganization, which can also decay when "out of practice" possibly due to horizontal connections and local inhibitory circuits (John Donoghue). Bizzi 2001 demonstrated 4 different motor cortex neurons during motor skill adaptation and washout - kinematic neurons (tuning does not change), dynamic neurons (tuning change in both adapation and washout), and memory neurons (change either during adaptation or washout only).

Studies found that adaptive changes in motor cortex activity lag the improvement in motor performance by several trials during adaptation. This suggests that leraning-related adjustments to motor commands are initially made elsewhere, with the cerebellum as one strong candidate. The primary motor cortex may thusb e more strongly involved in the slower process of long-term retention and recall of motor skills rather than the initial phase of learning a new skill.

The primary motor cortex is part of a distributed network of cortical motor areas, each with its own role in voluntary motor control. The primary motor cortex should be regarded as a dynamic computational map whose internal organization and spinal connections convert central signals about motor intentions and sensory feedback about the current state of the limb into motor output commands, rather than as a static map of specific muscles or movements of body parts. The tmoro cortex also provides a substrate for adatpve alterations during the acqustion of motor skills and the recovery of function after lesions.

Jul 13, 2015 - SW Ch. 7: What generates force and feedback?

What generates force and feedback?

Description of molecular muscle mechanism - actin and myosin coupling to shorten muscle fibers.

A muscle model with parallel and series springs are made to explain the active and passive forces generated by the muscle fiber. The force curve peaks around rest-length and decreases as a muscles stretches or contracts. Explains why isometric force is the greatest, i.e. force output when muscle length is not changing.

Covered how to convert forces applied by the muscle on joints to torques around those joints. Assuming constant force, this is done by relating joint angles, bone and muscle lengths. Equating angular work with linear work \[ \tau\Delta\theta = f\Delta\lambda \], we can derive the Jacobian \( \mathbf{J}=\mathbf{\frac{d\lambda}{d\theta}} \), where \( \mathbf{\theta} \) can be a vector. Using this, \( \mathbf{\tau}=-\mathbf{J^T}f \).

Muscle afferents including golgi tendon organs and muscle-spindle afferents which act as mechanical force sensors for the muscle. The muscle-spindles are innervated with \( \gamma \)-neurons at the poles. The primary muscle spindle afferents in the central nuclear bag correspond somewhat to muscle force velocity. Secondary muscle spindle afferents in the poles correspond to muscle length. \( \gamma \)-neurons innervate the poles to change length (co-activated with \( \alpha \)-neurons for the extrafusal muscles) as a type of target muscle activation. Perfect muscles length would result in 0 change in the primary afferent firing rates.

The \( \alpha \)-\( \gamma \) afferents monosynaptic connection is important in motor feedback.

This section is important for proprioception, motor movements, and motor feedback.

Jul 13, 2015 - SW Ch. 8: What provides limb stability?

Mechanisms of limb stability:

  • Antagonist muscle architecture produces an equilibrium point. Polit and Bizzi experiments with de-afferented monkeys showed passive properties can reach target, but cannot resist pertubation very well.
  • Passive, spring-like properties of limb promotes stability.
  • CNS reflexes.
  • Neuropathy - cannot sense location of limbs without vision.
  • [Mussa-Ivaldi] shows stiffness of the arm remains roughly constant when expressed in terms of joint coordinates.