Although tactile spatial acuity tests are used in both neuroscience research and clinical assessment, few automated devices exist for delivering controlled spatially structured stimuli to the skin. Consequently, investigators often apply tactile stimuli manually. Manual stimulus application is time consuming, requires great care and concentration on the part of the investigator, and leaves many stimulus parameters uncontrolled. We describe here a computer-controlled tactile stimulus system, the Tactile Automated Passive-finger Stimulator (TAPS), that applies spatially structured stimuli to the skin, controlling for onset velocity, contact force, and contact duration. TAPS is a versatile, programmable system, capable of efficiently conducting a variety of psychophysical procedures. We describe the components of TAPS, and show how TAPS is used to administer a two-interval forced-choice tactile grating orientation test.
Corresponding Author: Daniel Goldreich
22 Related JoVE Articles!
A Proboscis Extension Response Protocol for Investigating Behavioral Plasticity in Insects: Application to Basic, Biomedical, and Agricultural Research
Institutions: Arizona State University.
Insects modify their responses to stimuli through experience of associating those stimuli with events important for survival (e.g.
, food, mates, threats). There are several behavioral mechanisms through which an insect learns salient associations and relates them to these events. It is important to understand this behavioral plasticity for programs aimed toward assisting insects that are beneficial for agriculture. This understanding can also be used for discovering solutions to biomedical and agricultural problems created by insects that act as disease vectors and pests. The Proboscis Extension Response (PER) conditioning protocol was developed for honey bees (Apis mellifera
) over 50 years ago to study how they perceive and learn about floral odors, which signal the nectar and pollen resources a colony needs for survival. The PER procedure provides a robust and easy-to-employ framework for studying several different ecologically relevant mechanisms of behavioral plasticity. It is easily adaptable for use with several other insect species and other behavioral reflexes. These protocols can be readily employed in conjunction with various means for monitoring neural activity in the CNS via electrophysiology or bioimaging, or for manipulating targeted neuromodulatory pathways. It is a robust assay for rapidly detecting sub-lethal effects on behavior caused by environmental stressors, toxins or pesticides.
We show how the PER protocol is straightforward to implement using two procedures. One is suitable as a laboratory exercise for students or for quick assays of the effect of an experimental treatment. The other provides more thorough control of variables, which is important for studies of behavioral conditioning. We show how several measures for the behavioral response ranging from binary yes/no to more continuous variable like latency and duration of proboscis extension can be used to test hypotheses. And, we discuss some pitfalls that researchers commonly encounter when they use the procedure for the first time.
Neuroscience, Issue 91, PER, conditioning, honey bee, olfaction, olfactory processing, learning, memory, toxin assay
Characterization of Complex Systems Using the Design of Experiments Approach: Transient Protein Expression in Tobacco as a Case Study
Institutions: RWTH Aachen University, Fraunhofer Gesellschaft.
Plants provide multiple benefits for the production of biopharmaceuticals including low costs, scalability, and safety. Transient expression offers the additional advantage of short development and production times, but expression levels can vary significantly between batches thus giving rise to regulatory concerns in the context of good manufacturing practice. We used a design of experiments (DoE) approach to determine the impact of major factors such as regulatory elements in the expression construct, plant growth and development parameters, and the incubation conditions during expression, on the variability of expression between batches. We tested plants expressing a model anti-HIV monoclonal antibody (2G12) and a fluorescent marker protein (DsRed). We discuss the rationale for selecting certain properties of the model and identify its potential limitations. The general approach can easily be transferred to other problems because the principles of the model are broadly applicable: knowledge-based parameter selection, complexity reduction by splitting the initial problem into smaller modules, software-guided setup of optimal experiment combinations and step-wise design augmentation. Therefore, the methodology is not only useful for characterizing protein expression in plants but also for the investigation of other complex systems lacking a mechanistic description. The predictive equations describing the interconnectivity between parameters can be used to establish mechanistic models for other complex systems.
Bioengineering, Issue 83, design of experiments (DoE), transient protein expression, plant-derived biopharmaceuticals, promoter, 5'UTR, fluorescent reporter protein, model building, incubation conditions, monoclonal antibody
Methods to Explore the Influence of Top-down Visual Processes on Motor Behavior
Institutions: Rutgers University, Rutgers University, Rutgers University, Rutgers University, Rutgers University.
Kinesthetic awareness is important to successfully navigate the environment. When we interact with our daily surroundings, some aspects of movement are deliberately planned, while others spontaneously occur below conscious awareness. The deliberate component of this dichotomy has been studied extensively in several contexts, while the spontaneous component remains largely under-explored. Moreover, how perceptual processes modulate these movement classes is still unclear. In particular, a currently debated issue is whether the visuomotor system is governed by the spatial percept produced by a visual illusion or whether it is not affected by the illusion and is governed instead by the veridical percept. Bistable percepts such as 3D depth inversion illusions (DIIs) provide an excellent context to study such interactions and balance, particularly when used in combination with reach-to-grasp movements. In this study, a methodology is developed that uses a DII to clarify the role of top-down processes on motor action, particularly exploring how reaches toward a target on a DII are affected in both deliberate and spontaneous movement domains.
Behavior, Issue 86, vision for action, vision for perception, motor control, reach, grasp, visuomotor, ventral stream, dorsal stream, illusion, space perception, depth inversion
Multi-step Preparation Technique to Recover Multiple Metabolite Compound Classes for In-depth and Informative Metabolomic Analysis
Institutions: National Jewish Health, University of Colorado Denver.
Metabolomics is an emerging field which enables profiling of samples from living organisms in order to obtain insight into biological processes. A vital aspect of metabolomics is sample preparation whereby inconsistent techniques generate unreliable results. This technique encompasses protein precipitation, liquid-liquid extraction, and solid-phase extraction as a means of fractionating metabolites into four distinct classes. Improved enrichment of low abundance molecules with a resulting increase in sensitivity is obtained, and ultimately results in more confident identification of molecules. This technique has been applied to plasma, bronchoalveolar lavage fluid, and cerebrospinal fluid samples with volumes as low as 50 µl. Samples can be used for multiple downstream applications; for example, the pellet resulting from protein precipitation can be stored for later analysis. The supernatant from that step undergoes liquid-liquid extraction using water and strong organic solvent to separate the hydrophilic and hydrophobic compounds. Once fractionated, the hydrophilic layer can be processed for later analysis or discarded if not needed. The hydrophobic fraction is further treated with a series of solvents during three solid-phase extraction steps to separate it into fatty acids, neutral lipids, and phospholipids. This allows the technician the flexibility to choose which class of compounds is preferred for analysis. It also aids in more reliable metabolite identification since some knowledge of chemical class exists.
Bioengineering, Issue 89, plasma, chemistry techniques, analytical, solid phase extraction, mass spectrometry, metabolomics, fluids and secretions, profiling, small molecules, lipids, liquid chromatography, liquid-liquid extraction, cerebrospinal fluid, bronchoalveolar lavage fluid
Cortical Source Analysis of High-Density EEG Recordings in Children
Institutions: UCL Institute of Child Health, University College London.
EEG is traditionally described as a neuroimaging technique with high temporal and low spatial resolution. Recent advances in biophysical modelling and signal processing make it possible to exploit information from other imaging modalities like structural MRI that provide high spatial resolution to overcome this constraint1
. This is especially useful for investigations that require high resolution in the temporal as well as spatial domain. In addition, due to the easy application and low cost of EEG recordings, EEG is often the method of choice when working with populations, such as young children, that do not tolerate functional MRI scans well. However, in order to investigate which neural substrates are involved, anatomical information from structural MRI is still needed. Most EEG analysis packages work with standard head models that are based on adult anatomy. The accuracy of these models when used for children is limited2
, because the composition and spatial configuration of head tissues changes dramatically over development3
In the present paper, we provide an overview of our recent work in utilizing head models based on individual structural MRI scans or age specific head models to reconstruct the cortical generators of high density EEG. This article describes how EEG recordings are acquired, processed, and analyzed with pediatric populations at the London Baby Lab, including laboratory setup, task design, EEG preprocessing, MRI processing, and EEG channel level and source analysis.
Behavior, Issue 88, EEG, electroencephalogram, development, source analysis, pediatric, minimum-norm estimation, cognitive neuroscience, event-related potentials
Real-Time Impedance-based Cell Analyzer as a Tool to Delineate Molecular Pathways Involved in Neurotoxicity and Neuroprotection in a Neuronal Cell Line
Institutions: University of Zürich.
Many brain-related disorders have neuronal cell death involved in their pathophysiology. Improved in vitro
models to study neuroprotective or neurotoxic effects of drugs and downstream pathways involved would help gain insight into the molecular mechanisms of neuroprotection/neurotoxicity and could potentially facilitate drug development. However, many existing in vitro
toxicity assays have major limitations – most assess neurotoxicity and neuroprotection at a single time point, not allowing to observe the time-course and kinetics of the effect. Furthermore, the opportunity to collect information about downstream signaling pathways involved in neuroprotection in real-time would be of great importance. In the current protocol we describe the use of a real-time impedance-based cell analyzer to determine neuroprotective effects of serotonin 2A (5-HT2A
) receptor agonists in a neuronal cell line under label-free and real-time conditions using impedance measurements. Furthermore, we demonstrate that inhibitors of second messenger pathways can be used to delineate downstream molecules involved in the neuroprotective effect. We also describe the utility of this technique to determine whether an effect on cell proliferation contributes to an observed neuroprotective effect. The system utilizes special microelectronic plates referred to as E-Plates which contain alternating gold microelectrode arrays on the bottom surface of the wells, serving as cell sensors. The impedance readout is modified by the number of adherent cells, cell viability, morphology, and adhesion. A dimensionless parameter called Cell Index is derived from the electrical impedance measurements and is used to represent the cell status. Overall, the real-time impedance-based cell analyzer allows for real-time, label-free assessment of neuroprotection and neurotoxicity, and the evaluation of second messenger pathways involvement, contributing to more detailed and high-throughput assessment of potential neuroprotective compounds in vitro
, for selecting therapeutic candidates.
Neuroscience, Issue 90,
neuroscience, neuronal cell line, neurotoxicity, neuroprotection, real-time impedance-based cell analyzer, second messenger pathways, serotonin
Uncovering Beat Deafness: Detecting Rhythm Disorders with Synchronized Finger Tapping and Perceptual Timing Tasks
Institutions: University of Montpellier, Institut Universitaire de France, University of Finance and Management in Warsaw, International Laboratory for Brain, Music, and Sound Research (BRAMS).
A set of behavioral tasks for assessing perceptual and sensorimotor timing abilities in the general population (i.e.,
non-musicians) is presented here with the goal of uncovering rhythm disorders, such as beat deafness. Beat deafness is characterized by poor performance in perceiving durations in auditory rhythmic patterns or poor synchronization of movement with auditory rhythms (e.g.,
with musical beats). These tasks include the synchronization of finger tapping to the beat of simple and complex auditory stimuli and the detection of rhythmic irregularities (anisochrony detection task) embedded in the same stimuli. These tests, which are easy to administer, include an assessment of both perceptual and sensorimotor timing abilities under different conditions (e.g.,
beat rates and types of auditory material) and are based on the same auditory stimuli, ranging from a simple metronome to a complex musical excerpt. The analysis of synchronized tapping data is performed with circular statistics, which provide reliable measures of synchronization accuracy (e.g.,
the difference between the timing of the taps and the timing of the pacing stimuli) and consistency. Circular statistics on tapping data are particularly well-suited for detecting individual differences in the general population. Synchronized tapping and anisochrony detection are sensitive measures for identifying profiles of rhythm disorders and have been used with success to uncover cases of poor synchronization with spared perceptual timing. This systematic assessment of perceptual and sensorimotor timing can be extended to populations of patients with brain damage, neurodegenerative diseases (e.g.,
Parkinson’s disease), and developmental disorders (e.g.,
Attention Deficit Hyperactivity Disorder).
Behavior, Issue 97, rhythm, timing, synchronization, disorders, beat deafness, perception and action
Vision Training Methods for Sports Concussion Mitigation and Management
Institutions: University of Cincinnati, University of Cincinnati, University of Cincinnati, University of Cincinnati, University of Cincinnati, Cincinnati Children's Hospital Medical Center.
There is emerging evidence supporting the use vision training, including light board training tools, as a concussion baseline and neuro-diagnostic tool and potentially as a supportive component to concussion prevention strategies. This paper is focused on providing detailed methods for select vision training tools and reporting normative data for comparison when vision training is a part of a sports management program. The overall program includes standard vision training methods including tachistoscope, Brock’s string, and strobe glasses, as well as specialized light board training algorithms. Stereopsis is measured as a means to monitor vision training affects. In addition, quantitative results for vision training methods as well as baseline and post-testing *A and Reaction Test measures with progressive scores are reported. Collegiate athletes consistently improve after six weeks of training in their stereopsis, *A and Reaction Test scores. When vision training is initiated as a team wide exercise, the incidence of concussion decreases in players who participate in training compared to players who do not receive the vision training. Vision training produces functional and performance changes that, when monitored, can be used to assess the success of the vision training and can be initiated as part of a sports medical intervention for concussion prevention.
Behavior, Issue 99, Vision training, peripheral vision, functional peripheral vision, concussion, concussion management, diagnosis, rehabilitation, eyes, sight, seeing, sight
Making Record-efficiency SnS Solar Cells by Thermal Evaporation and Atomic Layer Deposition
Institutions: Massachusetts Institute of Technology, Massachusetts Institute of Technology, Harvard University, Massachusetts Institute of Technology, Harvard University.
Tin sulfide (SnS) is a candidate absorber material for Earth-abundant, non-toxic solar cells. SnS offers easy phase control and rapid growth by congruent thermal evaporation, and it absorbs visible light strongly. However, for a long time the record power conversion efficiency of SnS solar cells remained below 2%. Recently we demonstrated new certified record efficiencies of 4.36% using SnS deposited by atomic layer deposition, and 3.88% using thermal evaporation. Here the fabrication procedure for these record solar cells is described, and the statistical distribution of the fabrication process is reported. The standard deviation of efficiency measured on a single substrate is typically over 0.5%. All steps including substrate selection and cleaning, Mo sputtering for the rear contact (cathode), SnS deposition, annealing, surface passivation, Zn(O,S) buffer layer selection and deposition, transparent conductor (anode) deposition, and metallization are described. On each substrate we fabricate 11 individual devices, each with active area 0.25 cm2
. Further, a system for high throughput measurements of current-voltage curves under simulated solar light, and external quantum efficiency measurement with variable light bias is described. With this system we are able to measure full data sets on all 11 devices in an automated manner and in minimal time. These results illustrate the value of studying large sample sets, rather than focusing narrowly on the highest performing devices. Large data sets help us to distinguish and remedy individual loss mechanisms affecting our devices.
Engineering, Issue 99, Solar cells, thin films, thermal evaporation, atomic layer deposition, annealing, tin sulfide
Utility of Dissociated Intrinsic Hand Muscle Atrophy in the Diagnosis of Amyotrophic Lateral Sclerosis
Institutions: Westmead Hospital, University of Sydney, Australia.
The split hand
phenomenon refers to predominant wasting of thenar muscles and is an early and specific feature of amyotrophic lateral sclerosis (ALS). A novel split hand index (SI) was developed to quantify the split hand phenomenon, and its diagnostic utility was assessed in ALS patients. The split hand index was derived by dividing the product of the compound muscle action potential (CMAP) amplitude recorded over the abductor pollicis brevis and first dorsal interosseous muscles by the CMAP amplitude recorded over the abductor digiti minimi muscle. In order to assess the diagnostic utility of the split hand index, ALS patients were prospectively assessed and their results were compared to neuromuscular disorder patients. The split hand index was significantly reduced in ALS when compared to neuromuscular disorder patients (P<0.0001). Limb-onset ALS patients exhibited the greatest reduction in the split hand index, and a value of 5.2 or less reliably differentiated ALS from other neuromuscular disorders. Consequently, the split hand index appears to be a novel diagnostic biomarker for ALS, perhaps facilitating an earlier diagnosis.
Medicine, Issue 85, Amyotrophic Lateral Sclerosis (ALS), dissociated muscle atrophy, hypothenar muscles, motor neuron disease, split-hand index, thenar muscles
Automated, Quantitative Cognitive/Behavioral Screening of Mice: For Genetics, Pharmacology, Animal Cognition and Undergraduate Instruction
Institutions: Rutgers University, Koç University, New York University, Fairfield University.
We describe a high-throughput, high-volume, fully automated, live-in 24/7 behavioral testing system for assessing the effects of genetic and pharmacological manipulations on basic mechanisms of cognition and learning in mice. A standard polypropylene mouse housing tub is connected through an acrylic tube to a standard commercial mouse test box. The test box has 3 hoppers, 2 of which are connected to pellet feeders. All are internally illuminable with an LED and monitored for head entries by infrared (IR) beams. Mice live in the environment, which eliminates handling during screening. They obtain their food during two or more daily feeding periods by performing in operant (instrumental) and Pavlovian (classical) protocols, for which we have written protocol-control software and quasi-real-time data analysis and graphing software. The data analysis and graphing routines are written in a MATLAB-based language created to simplify greatly the analysis of large time-stamped behavioral and physiological event records and to preserve a full data trail from raw data through all intermediate analyses to the published graphs and statistics within a single data structure. The data-analysis code harvests the data several times a day and subjects it to statistical and graphical analyses, which are automatically stored in the "cloud" and on in-lab computers. Thus, the progress of individual mice is visualized and quantified daily. The data-analysis code talks to the protocol-control code, permitting the automated advance from protocol to protocol of individual subjects. The behavioral protocols implemented are matching, autoshaping, timed hopper-switching, risk assessment in timed hopper-switching, impulsivity measurement, and the circadian anticipation of food availability. Open-source protocol-control and data-analysis code makes the addition of new protocols simple. Eight test environments fit in a 48 in x 24 in x 78 in cabinet; two such cabinets (16 environments) may be controlled by one computer.
Behavior, Issue 84, genetics, cognitive mechanisms, behavioral screening, learning, memory, timing
Test Samples for Optimizing STORM Super-Resolution Microscopy
Institutions: National Physical Laboratory.
STORM is a recently developed super-resolution microscopy technique with up to 10 times better resolution than standard fluorescence microscopy techniques. However, as the image is acquired in a very different way than normal, by building up an image molecule-by-molecule, there are some significant challenges for users in trying to optimize their image acquisition. In order to aid this process and gain more insight into how STORM works we present the preparation of 3 test samples and the methodology of acquiring and processing STORM super-resolution images with typical resolutions of between 30-50 nm. By combining the test samples with the use of the freely available rainSTORM processing software it is possible to obtain a great deal of information about image quality and resolution. Using these metrics it is then possible to optimize the imaging procedure from the optics, to sample preparation, dye choice, buffer conditions, and image acquisition settings. We also show examples of some common problems that result in poor image quality, such as lateral drift, where the sample moves during image acquisition and density related problems resulting in the 'mislocalization' phenomenon.
Molecular Biology, Issue 79, Genetics, Bioengineering, Biomedical Engineering, Biophysics, Basic Protocols, HeLa Cells, Actin Cytoskeleton, Coated Vesicles, Receptor, Epidermal Growth Factor, Actins, Fluorescence, Endocytosis, Microscopy, STORM, super-resolution microscopy, nanoscopy, cell biology, fluorescence microscopy, test samples, resolution, actin filaments, fiducial markers, epidermal growth factor, cell, imaging
Semi-automated Optical Heartbeat Analysis of Small Hearts
Institutions: The Sanford Burnham Institute for Medical Research, The Sanford Burnham Institute for Medical Research, San Diego State University.
We have developed a method for analyzing high speed optical recordings from Drosophila
, zebrafish and embryonic mouse hearts (Fink, et. al., 2009). Our Semi-automatic Optical Heartbeat Analysis (SOHA) uses a novel movement detection algorithm that is able to detect cardiac movements associated with individual contractile and relaxation events. The program provides a host of physiologically relevant readouts including systolic and diastolic intervals, heart rate, as well as qualitative and quantitative measures of heartbeat arrhythmicity. The program also calculates heart diameter measurements during both diastole and systole from which fractional shortening and fractional area changes are calculated. Output is provided as a digital file compatible with most spreadsheet programs. Measurements are made for every heartbeat in a record increasing the statistical power of the output. We demonstrate each of the steps where user input is required and show the application of our methodology to the analysis of heart function in all three genetically tractable heart models.
Physiology, Issue 31, Drosophila, zebrafish, mouse, heart, myosin, dilated, restricted, cardiomyopathy, KCNQ, movement detection
Basics of Multivariate Analysis in Neuroimaging Data
Institutions: Columbia University.
Multivariate analysis techniques for neuroimaging data have recently received increasing attention as they have many attractive features that cannot be easily realized by the more commonly used univariate, voxel-wise, techniques1,5,6,7,8,9
. Multivariate approaches evaluate correlation/covariance of activation across brain regions, rather than proceeding on a voxel-by-voxel basis. Thus, their results can be more easily interpreted as a signature of neural networks. Univariate approaches, on the other hand, cannot directly address interregional correlation in the brain. Multivariate approaches can also result in greater statistical power when compared with univariate techniques, which are forced to employ very stringent corrections for voxel-wise multiple comparisons. Further, multivariate techniques also lend themselves much better to prospective application of results from the analysis of one dataset to entirely new datasets. Multivariate techniques are thus well placed to provide information about mean differences and correlations with behavior, similarly to univariate approaches, with potentially greater statistical power and better reproducibility checks. In contrast to these advantages is the high barrier of entry to the use of multivariate approaches, preventing more widespread application in the community. To the neuroscientist becoming familiar with multivariate analysis techniques, an initial survey of the field might present a bewildering variety of approaches that, although algorithmically similar, are presented with different emphases, typically by people with mathematics backgrounds. We believe that multivariate analysis techniques have sufficient potential to warrant better dissemination. Researchers should be able to employ them in an informed and accessible manner. The current article is an attempt at a didactic introduction of multivariate techniques for the novice. A conceptual introduction is followed with a very simple application to a diagnostic data set from the Alzheimer s Disease Neuroimaging Initiative (ADNI), clearly demonstrating the superior performance of the multivariate approach.
JoVE Neuroscience, Issue 41, fMRI, PET, multivariate analysis, cognitive neuroscience, clinical neuroscience
One Dimensional Turing-Like Handshake Test for Motor Intelligence
Institutions: Ben-Gurion University.
In the Turing test, a computer model is deemed to "think intelligently" if it can generate answers that are not distinguishable from those of a human. However, this test is limited to the linguistic aspects of machine intelligence. A salient function of the brain is the control of movement, and the movement of the human hand is a sophisticated demonstration of this function. Therefore, we propose a Turing-like handshake test, for machine motor intelligence. We administer the test through a telerobotic system in which the interrogator is engaged in a task of holding a robotic stylus and interacting with another party (human or artificial). Instead of asking the interrogator whether the other party is a person or a computer program, we employ a two-alternative forced choice method and ask which of two systems is more human-like. We extract a quantitative grade for each model according to its resemblance to the human handshake motion and name it "Model Human-Likeness Grade" (MHLG). We present three methods to estimate the MHLG. (i) By calculating the proportion of subjects' answers that the model is more human-like than the human; (ii) By comparing two weighted sums of human and model handshakes we fit a psychometric curve and extract the point of subjective equality (PSE); (iii) By comparing a given model with a weighted sum of human and random signal, we fit a psychometric curve to the answers of the interrogator and extract the PSE for the weight of the human in the weighted sum. Altogether, we provide a protocol to test computational models of the human handshake. We believe that building a model is a necessary step in understanding any phenomenon and, in this case, in understanding the neural mechanisms responsible for the generation of the human handshake.
Neuroscience, Issue 46, Turing test, Human Machine Interface, Haptics, Teleoperation, Motor Control, Motor Behavior, Diagnostics, Perception, handshake, telepresence
Utilizing Transcranial Magnetic Stimulation to Study the Human Neuromuscular System
Institutions: Ohio University.
Transcranial magnetic stimulation (TMS) has been in use for more than 20 years 1
, and has grown exponentially in popularity over the past decade. While the use of TMS has expanded to the study of many systems and processes during this time, the original application and perhaps one of the most common uses of TMS involves studying the physiology, plasticity and function of the human neuromuscular system. Single pulse TMS applied to the motor cortex excites pyramidal neurons transsynaptically 2
(Figure 1) and results in a measurable electromyographic response that can be used to study and evaluate the integrity and excitability of the corticospinal tract in humans 3
. Additionally, recent advances in magnetic stimulation now allows for partitioning of cortical versus spinal excitability 4,5
. For example, paired-pulse TMS can be used to assess intracortical facilitatory and inhibitory properties by combining a conditioning stimulus and a test stimulus at different interstimulus intervals 3,4,6-8
. In this video article we will demonstrate the methodological and technical aspects of these techniques. Specifically, we will demonstrate single-pulse and paired-pulse TMS techniques as applied to the flexor carpi radialis (FCR) muscle as well as the erector spinae (ES) musculature. Our laboratory studies the FCR muscle as it is of interest to our research on the effects of wrist-hand cast immobilization on reduced muscle performance6,9
, and we study the ES muscles due to these muscles clinical relevance as it relates to low back pain8
. With this stated, we should note that TMS has been used to study many muscles of the hand, arm and legs, and should iterate that our demonstrations in the FCR and ES muscle groups are only selected examples of TMS being used to study the human neuromuscular system.
Medicine, Issue 59, neuroscience, muscle, electromyography, physiology, TMS, strength, motor control. sarcopenia, dynapenia, lumbar
The α-test: Rapid Cell-free CD4 Enumeration Using Whole Saliva
Institutions: Weill Cornell Medical College , University of Missouri-Kansas City-School of Dentistry, University of Missouri Kansas City- School of Pharmacy, Bamenda, NWP, Cameroon, Mezam Polyclinic HIV/AIDS Treatment Center, Cameroon, Institute for Human Genetics and Biochemistry.
There is an urgent need for affordable CD4 enumeration to monitor HIV disease. CD4 enumeration is out of reach in resource-limited regions due to the time and temperature restrictions, technical sophistication, and cost of reagents, in particular monoclonal antibodies to measure CD4 on blood cells, the only currently acceptable method. A commonly used cost-saving and time-saving laboratory strategy is to calculate, rather than measure certain blood values. For example, LDL levels are calculated using the measured levels of total cholesterol, HDL, and triglycerides1
. Thus, identification of cell-free correlates that directly regulate the number of CD4+
T cells could provide an accurate method for calculating CD4 counts due to the physiological relevance of the correlates.
The number of stem cells that enter blood and are destined to become circulating CD4+
T cells is determined by the chemokine CXCL12 and its receptor CXCR4 due to their influence on locomotion2
. The process of stem cell locomotion into blood is additionally regulated by cell surface human leukocyte elastase (HLECS
) and the HLECS
-reactive active α1
proteinase inhibitor (α1
. In HIV-1 disease, α1
PI is inactivated due to disease processes 4
. In the early asymptomatic categories of HIV-1 disease, active α1
PI was found to be below normal in 100% of untreated HIV-1 patients (median=12 μM, and to achieve normal levels during the symptomatic categories4, 5
. This pattern has been attributed to immune inactivation, not to insufficient synthesis, proteolytic inactivation, or oxygenation. We observed that in HIV-1 subjects with >220 CD4 cells/μl, CD4 counts were correlated with serum levels of active α1
=0.93, p<0.0001, n=26) and inactive α1
=0.91, p<0.0001, n=26) 5
. Administration of α1
PI to HIV-1 infected and uninfected subjects resulted in dramatic increases in CD4 counts suggesting α1
PI participates in regulating the number of CD4+
T cells in blood 3
With stimulation, whole saliva contains sufficient serous exudate (plasma containing proteinaceous material that passes through blood vessel walls into saliva) to allow measurement of active α1
PI and the correlation of this measurement is evidence that it is an accurate method for calculating CD4 counts. Briefly, sialogogues such as chewing gum or citric acid stimulate the exudation of serum into whole mouth saliva. After stimulating serum exudation, the activity of serum α1
PI in saliva is measured by its capacity to inhibit elastase activity. Porcine pancreatic elastase (PPE) is a readily available inexpensive source of elastase. PPE binds to α1
PI forming a one-to-one complex that prevents PPE from cleaving its specific substrates, one of which is the colorimetric peptide, succinyl-L-Ala-L-Ala-L-Ala-p-nitroanilide (SA3
NA). Incubating saliva with a saturating concentration of PPE for 10 min at room temperature allows the binding of PPE to all the active α1
PI in saliva. The resulting inhibition of PPE by active α1
PI can be measured by adding the PPE substrate SA3
NA. (Figure 1)
. Although CD4 counts are measured in terms of blood volume (CD4 cells/μl), the concentration of α1
PI in saliva is related to the concentration of serum in saliva, not to volume of saliva since volume can vary considerably during the day and person to person6
. However, virtually all the protein in saliva is due to serum content, and the protein content of saliva is measurable7
. Thus, active α1
PI in saliva is calculated as a ratio to saliva protein content and is termed the α1
PI Index. Results presented herein demonstrate that the α1
PI Index provides an accurate and precise physiologic method for calculating CD4 counts.
Medicine, Issue 63, CD4 count, saliva, antitrypsin, hematopoiesis, T cells, HIV/AIDS, clinical
Perceptual and Category Processing of the Uncanny Valley Hypothesis' Dimension of Human Likeness: Some Methodological Issues
Institutions: University of Zurich.
Mori's Uncanny Valley Hypothesis1,2
proposes that the perception of humanlike characters such as robots and, by extension, avatars (computer-generated characters) can evoke negative or positive affect (valence) depending on the object's degree of visual and behavioral realism along a dimension of human likeness
) (Figure 1
). But studies of affective valence of subjective responses to variously realistic non-human characters have produced inconsistent findings 3, 4, 5, 6
. One of a number of reasons for this is that human likeness is not perceived as the hypothesis assumes. While the DHL can be defined following Mori's description as a smooth linear change in the degree of physical humanlike similarity, subjective perception of objects along the DHL can be understood in terms of the psychological effects of categorical perception (CP) 7
. Further behavioral and neuroimaging investigations of category processing and CP along the DHL and of the potential influence of the dimension's underlying category structure on affective experience are needed. This protocol therefore focuses on the DHL and allows examination of CP. Based on the protocol presented in the video as an example, issues surrounding the methodology in the protocol and the use in "uncanny" research of stimuli drawn from morph continua to represent the DHL are discussed in the article that accompanies the video. The use of neuroimaging and morph stimuli to represent the DHL in order to disentangle brain regions neurally responsive to physical human-like similarity from those responsive to category change and category processing is briefly illustrated.
Behavior, Issue 76, Neuroscience, Neurobiology, Molecular Biology, Psychology, Neuropsychology, uncanny valley, functional magnetic resonance imaging, fMRI, categorical perception, virtual reality, avatar, human likeness, Mori, uncanny valley hypothesis, perception, magnetic resonance imaging, MRI, imaging, clinical techniques
Detection of Architectural Distortion in Prior Mammograms via Analysis of Oriented Patterns
Institutions: University of Calgary , University of Calgary .
We demonstrate methods for the detection of architectural distortion in prior mammograms of interval-cancer cases based on analysis of the orientation of breast tissue patterns in mammograms. We hypothesize that architectural distortion modifies the normal orientation of breast tissue patterns in mammographic images before the formation of masses or tumors. In the initial steps of our methods, the oriented structures in a given mammogram are analyzed using Gabor filters and phase portraits to detect node-like sites of radiating or intersecting tissue patterns. Each detected site is then characterized using the node value, fractal dimension, and a measure of angular dispersion specifically designed to represent spiculating patterns associated with architectural distortion.
Our methods were tested with a database of 106 prior mammograms of 56 interval-cancer cases and 52 mammograms of 13 normal cases using the features developed for the characterization of architectural distortion, pattern classification via
quadratic discriminant analysis, and validation with the leave-one-patient out procedure. According to the results of free-response receiver operating characteristic analysis, our methods have demonstrated the capability to detect architectural distortion in prior mammograms, taken 15 months (on the average) before clinical diagnosis of breast cancer, with a sensitivity of 80% at about five false positives per patient.
Medicine, Issue 78, Anatomy, Physiology, Cancer Biology, angular spread, architectural distortion, breast cancer, Computer-Assisted Diagnosis, computer-aided diagnosis (CAD), entropy, fractional Brownian motion, fractal dimension, Gabor filters, Image Processing, Medical Informatics, node map, oriented texture, Pattern Recognition, phase portraits, prior mammograms, spectral analysis
Setting Limits on Supersymmetry Using Simplified Models
Institutions: University College London, CERN, Lawrence Berkeley National Laboratories.
Experimental limits on supersymmetry and similar theories are difficult to set because of the enormous available parameter space and difficult to generalize because of the complexity of single points. Therefore, more phenomenological, simplified models are becoming popular for setting experimental limits, as they have clearer physical interpretations. The use of these simplified model limits to set a real limit on a concrete theory has not, however, been demonstrated. This paper recasts simplified model limits into limits on a specific and complete supersymmetry model, minimal supergravity. Limits obtained under various physical assumptions are comparable to those produced by directed searches. A prescription is provided for calculating conservative and aggressive limits on additional theories. Using acceptance and efficiency tables along with the expected and observed numbers of events in various signal regions, LHC experimental results can be recast in this manner into almost any theoretical framework, including nonsupersymmetric theories with supersymmetry-like signatures.
Physics, Issue 81, high energy physics, particle physics, Supersymmetry, LHC, ATLAS, CMS, New Physics Limits, Simplified Models
A Chemical Screening Procedure for Glucocorticoid Signaling with a Zebrafish Larva Luciferase Reporter System
Institutions: Karlsruhe Institute of Technology - Campus North, Karlsruhe Institute of Technology - Campus North, Karlsruhe Institute of Technology - Campus South.
Glucocorticoid stress hormones and their artificial derivatives are widely used drugs to treat inflammation, but long-term treatment with glucocorticoids can lead to severe side effects. Test systems are needed to search for novel compounds influencing glucocorticoid signaling in vivo
or to determine unwanted effects of compounds on the glucocorticoid signaling pathway. We have established a transgenic zebrafish assay which allows the measurement of glucocorticoid signaling activity in vivo
and in real-time, the GRIZLY assay (Glucocorticoid Responsive In vivo
Zebrafish Luciferase activitY). The luciferase-based assay detects effects on glucocorticoid signaling with high sensitivity and specificity, including effects by compounds that require metabolization or affect endogenous glucocorticoid production. We present here a detailed protocol for conducting chemical screens with this assay. We describe data acquisition, normalization, and analysis, placing a focus on quality control and data visualization. The assay provides a simple, time-resolved, and quantitative readout. It can be operated as a stand-alone platform, but is also easily integrated into high-throughput screening workflows. It furthermore allows for many applications beyond chemical screening, such as environmental monitoring of endocrine disruptors or stress research.
Developmental Biology, Issue 79, Biochemistry, Vertebrates, Zebrafish, environmental effects (biological and animal), genetics (animal), life sciences, animal biology, animal models, biochemistry, bioengineering (general), Hormones, Hormone Substitutes, and Hormone Antagonists, zebrafish, Danio rerio, chemical screening, luciferase, glucocorticoid, stress, high-throughput screening, receiver operating characteristic curve, in vivo, animal model
An Organotypic High Throughput System for Characterization of Drug Sensitivity of Primary Multiple Myeloma Cells
Institutions: H. Lee Moffitt Cancer Center and Research Institute.
In this work we describe a novel approach that combines ex vivo
drug sensitivity assays and digital image analysis to estimate chemosensitivity and heterogeneity of patient-derived multiple myeloma (MM) cells. This approach consists in seeding primary MM cells freshly extracted from bone marrow aspirates into microfluidic chambers implemented in multi-well plates, each consisting of a reconstruction of the bone marrow microenvironment, including extracellular matrix (collagen or basement membrane matrix) and stroma (patient-derived mesenchymal stem cells) or human-derived endothelial cells (HUVECs). The chambers are drugged with different agents and concentrations, and are imaged sequentially for 96 hr through bright field microscopy, in a motorized microscope equipped with a digital camera. Digital image analysis software detects live and dead cells from presence or absence of membrane motion, and generates curves of change in viability as a function of drug concentration and exposure time. We use a computational model to determine the parameters of chemosensitivity of the tumor population to each drug, as well as the number of sub-populations present as a measure of tumor heterogeneity. These patient-tailored models can then be used to simulate therapeutic regimens and estimate clinical response.
Medicine, Issue 101, Multiple myeloma, drug sensitivity, evolution of drug resistance, computational modeling, decision support system, personalized medicine