JoVE Visualize What is visualize?
Related JoVE Video
Pubmed Article
Raman spectroscopy provides a powerful diagnostic tool for accurate determination of albumin glycation.
We present the first demonstration of glycated albumin detection and quantification using Raman spectroscopy without the addition of reagents. Glycated albumin is an important marker for monitoring the long-term glycemic history of diabetics, especially as its concentrations, in contrast to glycated hemoglobin levels, are unaffected by changes in erythrocyte life times. Clinically, glycated albumin concentrations show a strong correlation with the development of serious diabetes complications including nephropathy and retinopathy. In this article, we propose and evaluate the efficacy of Raman spectroscopy for determination of this important analyte. By utilizing the pre-concentration obtained through drop-coating deposition, we show that glycation of albumin leads to subtle, but consistent, changes in vibrational features, which with the help of multivariate classification techniques can be used to discriminate glycated albumin from the unglycated variant with 100% accuracy. Moreover, we demonstrate that the calibration model developed on the glycated albumin spectral dataset shows high predictive power, even at substantially lower concentrations than those typically encountered in clinical practice. In fact, the limit of detection for glycated albumin measurements is calculated to be approximately four times lower than its minimum physiological concentration. Importantly, in relation to the existing detection methods for glycated albumin, the proposed method is also completely reagent-free, requires barely any sample preparation and has the potential for simultaneous determination of glycated hemoglobin levels as well. Given these key advantages, we believe that the proposed approach can provide a uniquely powerful tool for quantification of glycation status of proteins in biopharmaceutical development as well as for glycemic marker determination in routine clinical diagnostics in the future.
Authors: Andrew L. Fussell, Peter Kleinebudde, Jennifer Herek, Clare J. Strachan, Herman L. Offerhaus.
Published: 07-04-2014
Traditional pharmaceutical dissolution tests determine the amount of drug dissolved over time by measuring drug content in the dissolution medium. This method provides little direct information about what is happening on the surface of the dissolving tablet. As the tablet surface composition and structure can change during dissolution, it is essential to monitor it during dissolution testing. In this work coherent anti-Stokes Raman scattering microscopy is used to image the surface of tablets during dissolution while UV absorption spectroscopy is simultaneously providing inline analysis of dissolved drug concentration for tablets containing a 50% mixture of theophylline anhydrate and ethyl cellulose. The measurements showed that in situ CARS microscopy is capable of imaging selectively theophylline in the presence of ethyl cellulose. Additionally, the theophylline anhydrate converted to theophylline monohydrate during dissolution, with needle-shaped crystals growing on the tablet surface during dissolution. The conversion of theophylline anhydrate to monohydrate, combined with reduced exposure of the drug to the flowing dissolution medium resulted in decreased dissolution rates. Our results show that in situ CARS microscopy combined with inline UV absorption spectroscopy is capable of monitoring pharmaceutical tablet dissolution and correlating surface changes with changes in dissolution rate.
27 Related JoVE Articles!
Play Button
Polymerase Chain Reaction: Basic Protocol Plus Troubleshooting and Optimization Strategies
Authors: Todd C. Lorenz.
Institutions: University of California, Los Angeles .
In the biological sciences there have been technological advances that catapult the discipline into golden ages of discovery. For example, the field of microbiology was transformed with the advent of Anton van Leeuwenhoek's microscope, which allowed scientists to visualize prokaryotes for the first time. The development of the polymerase chain reaction (PCR) is one of those innovations that changed the course of molecular science with its impact spanning countless subdisciplines in biology. The theoretical process was outlined by Keppe and coworkers in 1971; however, it was another 14 years until the complete PCR procedure was described and experimentally applied by Kary Mullis while at Cetus Corporation in 1985. Automation and refinement of this technique progressed with the introduction of a thermal stable DNA polymerase from the bacterium Thermus aquaticus, consequently the name Taq DNA polymerase. PCR is a powerful amplification technique that can generate an ample supply of a specific segment of DNA (i.e., an amplicon) from only a small amount of starting material (i.e., DNA template or target sequence). While straightforward and generally trouble-free, there are pitfalls that complicate the reaction producing spurious results. When PCR fails it can lead to many non-specific DNA products of varying sizes that appear as a ladder or smear of bands on agarose gels. Sometimes no products form at all. Another potential problem occurs when mutations are unintentionally introduced in the amplicons, resulting in a heterogeneous population of PCR products. PCR failures can become frustrating unless patience and careful troubleshooting are employed to sort out and solve the problem(s). This protocol outlines the basic principles of PCR, provides a methodology that will result in amplification of most target sequences, and presents strategies for optimizing a reaction. By following this PCR guide, students should be able to: ● Set up reactions and thermal cycling conditions for a conventional PCR experiment ● Understand the function of various reaction components and their overall effect on a PCR experiment ● Design and optimize a PCR experiment for any DNA template ● Troubleshoot failed PCR experiments
Basic Protocols, Issue 63, PCR, optimization, primer design, melting temperature, Tm, troubleshooting, additives, enhancers, template DNA quantification, thermal cycler, molecular biology, genetics
Play Button
Models and Methods to Evaluate Transport of Drug Delivery Systems Across Cellular Barriers
Authors: Rasa Ghaffarian, Silvia Muro.
Institutions: University of Maryland, University of Maryland.
Sub-micrometer carriers (nanocarriers; NCs) enhance efficacy of drugs by improving solubility, stability, circulation time, targeting, and release. Additionally, traversing cellular barriers in the body is crucial for both oral delivery of therapeutic NCs into the circulation and transport from the blood into tissues, where intervention is needed. NC transport across cellular barriers is achieved by: (i) the paracellular route, via transient disruption of the junctions that interlock adjacent cells, or (ii) the transcellular route, where materials are internalized by endocytosis, transported across the cell body, and secreted at the opposite cell surface (transyctosis). Delivery across cellular barriers can be facilitated by coupling therapeutics or their carriers with targeting agents that bind specifically to cell-surface markers involved in transport. Here, we provide methods to measure the extent and mechanism of NC transport across a model cell barrier, which consists of a monolayer of gastrointestinal (GI) epithelial cells grown on a porous membrane located in a transwell insert. Formation of a permeability barrier is confirmed by measuring transepithelial electrical resistance (TEER), transepithelial transport of a control substance, and immunostaining of tight junctions. As an example, ~200 nm polymer NCs are used, which carry a therapeutic cargo and are coated with an antibody that targets a cell-surface determinant. The antibody or therapeutic cargo is labeled with 125I for radioisotope tracing and labeled NCs are added to the upper chamber over the cell monolayer for varying periods of time. NCs associated to the cells and/or transported to the underlying chamber can be detected. Measurement of free 125I allows subtraction of the degraded fraction. The paracellular route is assessed by determining potential changes caused by NC transport to the barrier parameters described above. Transcellular transport is determined by addressing the effect of modulating endocytosis and transcytosis pathways.
Bioengineering, Issue 80, Antigens, Enzymes, Biological Therapy, bioengineering (general), Pharmaceutical Preparations, Macromolecular Substances, Therapeutics, Digestive System and Oral Physiological Phenomena, Biological Phenomena, Cell Physiological Phenomena, drug delivery systems, targeted nanocarriers, transcellular transport, epithelial cells, tight junctions, transepithelial electrical resistance, endocytosis, transcytosis, radioisotope tracing, immunostaining
Play Button
A Microplate Assay to Assess Chemical Effects on RBL-2H3 Mast Cell Degranulation: Effects of Triclosan without Use of an Organic Solvent
Authors: Lisa M. Weatherly, Rachel H. Kennedy, Juyoung Shim, Julie A. Gosse.
Institutions: University of Maine, Orono, University of Maine, Orono.
Mast cells play important roles in allergic disease and immune defense against parasites. Once activated (e.g. by an allergen), they degranulate, a process that results in the exocytosis of allergic mediators. Modulation of mast cell degranulation by drugs and toxicants may have positive or adverse effects on human health. Mast cell function has been dissected in detail with the use of rat basophilic leukemia mast cells (RBL-2H3), a widely accepted model of human mucosal mast cells3-5. Mast cell granule component and the allergic mediator β-hexosaminidase, which is released linearly in tandem with histamine from mast cells6, can easily and reliably be measured through reaction with a fluorogenic substrate, yielding measurable fluorescence intensity in a microplate assay that is amenable to high-throughput studies1. Originally published by Naal et al.1, we have adapted this degranulation assay for the screening of drugs and toxicants and demonstrate its use here. Triclosan is a broad-spectrum antibacterial agent that is present in many consumer products and has been found to be a therapeutic aid in human allergic skin disease7-11, although the mechanism for this effect is unknown. Here we demonstrate an assay for the effect of triclosan on mast cell degranulation. We recently showed that triclosan strongly affects mast cell function2. In an effort to avoid use of an organic solvent, triclosan is dissolved directly into aqueous buffer with heat and stirring, and resultant concentration is confirmed using UV-Vis spectrophotometry (using ε280 = 4,200 L/M/cm)12. This protocol has the potential to be used with a variety of chemicals to determine their effects on mast cell degranulation, and more broadly, their allergic potential.
Immunology, Issue 81, mast cell, basophil, degranulation, RBL-2H3, triclosan, irgasan, antibacterial, β-hexosaminidase, allergy, Asthma, toxicants, ionophore, antigen, fluorescence, microplate, UV-Vis
Play Button
Characterization of Complex Systems Using the Design of Experiments Approach: Transient Protein Expression in Tobacco as a Case Study
Authors: Johannes Felix Buyel, Rainer Fischer.
Institutions: RWTH Aachen University, Fraunhofer Gesellschaft.
Plants provide multiple benefits for the production of biopharmaceuticals including low costs, scalability, and safety. Transient expression offers the additional advantage of short development and production times, but expression levels can vary significantly between batches thus giving rise to regulatory concerns in the context of good manufacturing practice. We used a design of experiments (DoE) approach to determine the impact of major factors such as regulatory elements in the expression construct, plant growth and development parameters, and the incubation conditions during expression, on the variability of expression between batches. We tested plants expressing a model anti-HIV monoclonal antibody (2G12) and a fluorescent marker protein (DsRed). We discuss the rationale for selecting certain properties of the model and identify its potential limitations. The general approach can easily be transferred to other problems because the principles of the model are broadly applicable: knowledge-based parameter selection, complexity reduction by splitting the initial problem into smaller modules, software-guided setup of optimal experiment combinations and step-wise design augmentation. Therefore, the methodology is not only useful for characterizing protein expression in plants but also for the investigation of other complex systems lacking a mechanistic description. The predictive equations describing the interconnectivity between parameters can be used to establish mechanistic models for other complex systems.
Bioengineering, Issue 83, design of experiments (DoE), transient protein expression, plant-derived biopharmaceuticals, promoter, 5'UTR, fluorescent reporter protein, model building, incubation conditions, monoclonal antibody
Play Button
Modeling Neural Immune Signaling of Episodic and Chronic Migraine Using Spreading Depression In Vitro
Authors: Aya D. Pusic, Yelena Y. Grinberg, Heidi M. Mitchell, Richard P. Kraig.
Institutions: The University of Chicago Medical Center, The University of Chicago Medical Center.
Migraine and its transformation to chronic migraine are healthcare burdens in need of improved treatment options. We seek to define how neural immune signaling modulates the susceptibility to migraine, modeled in vitro using spreading depression (SD), as a means to develop novel therapeutic targets for episodic and chronic migraine. SD is the likely cause of migraine aura and migraine pain. It is a paroxysmal loss of neuronal function triggered by initially increased neuronal activity, which slowly propagates within susceptible brain regions. Normal brain function is exquisitely sensitive to, and relies on, coincident low-level immune signaling. Thus, neural immune signaling likely affects electrical activity of SD, and therefore migraine. Pain perception studies of SD in whole animals are fraught with difficulties, but whole animals are well suited to examine systems biology aspects of migraine since SD activates trigeminal nociceptive pathways. However, whole animal studies alone cannot be used to decipher the cellular and neural circuit mechanisms of SD. Instead, in vitro preparations where environmental conditions can be controlled are necessary. Here, it is important to recognize limitations of acute slices and distinct advantages of hippocampal slice cultures. Acute brain slices cannot reveal subtle changes in immune signaling since preparing the slices alone triggers: pro-inflammatory changes that last days, epileptiform behavior due to high levels of oxygen tension needed to vitalize the slices, and irreversible cell injury at anoxic slice centers. In contrast, we examine immune signaling in mature hippocampal slice cultures since the cultures closely parallel their in vivo counterpart with mature trisynaptic function; show quiescent astrocytes, microglia, and cytokine levels; and SD is easily induced in an unanesthetized preparation. Furthermore, the slices are long-lived and SD can be induced on consecutive days without injury, making this preparation the sole means to-date capable of modeling the neuroimmune consequences of chronic SD, and thus perhaps chronic migraine. We use electrophysiological techniques and non-invasive imaging to measure neuronal cell and circuit functions coincident with SD. Neural immune gene expression variables are measured with qPCR screening, qPCR arrays, and, importantly, use of cDNA preamplification for detection of ultra-low level targets such as interferon-gamma using whole, regional, or specific cell enhanced (via laser dissection microscopy) sampling. Cytokine cascade signaling is further assessed with multiplexed phosphoprotein related targets with gene expression and phosphoprotein changes confirmed via cell-specific immunostaining. Pharmacological and siRNA strategies are used to mimic and modulate SD immune signaling.
Neuroscience, Issue 52, innate immunity, hormesis, microglia, T-cells, hippocampus, slice culture, gene expression, laser dissection microscopy, real-time qPCR, interferon-gamma
Play Button
In Situ SIMS and IR Spectroscopy of Well-defined Surfaces Prepared by Soft Landing of Mass-selected Ions
Authors: Grant E. Johnson, K. Don Dasitha Gunaratne, Julia Laskin.
Institutions: Pacific Northwest National Laboratory.
Soft landing of mass-selected ions onto surfaces is a powerful approach for the highly-controlled preparation of materials that are inaccessible using conventional synthesis techniques. Coupling soft landing with in situ characterization using secondary ion mass spectrometry (SIMS) and infrared reflection absorption spectroscopy (IRRAS) enables analysis of well-defined surfaces under clean vacuum conditions. The capabilities of three soft-landing instruments constructed in our laboratory are illustrated for the representative system of surface-bound organometallics prepared by soft landing of mass-selected ruthenium tris(bipyridine) dications, [Ru(bpy)3]2+ (bpy = bipyridine), onto carboxylic acid terminated self-assembled monolayer surfaces on gold (COOH-SAMs). In situ time-of-flight (TOF)-SIMS provides insight into the reactivity of the soft-landed ions. In addition, the kinetics of charge reduction, neutralization and desorption occurring on the COOH-SAM both during and after ion soft landing are studied using in situ Fourier transform ion cyclotron resonance (FT-ICR)-SIMS measurements. In situ IRRAS experiments provide insight into how the structure of organic ligands surrounding metal centers is perturbed through immobilization of organometallic ions on COOH-SAM surfaces by soft landing. Collectively, the three instruments provide complementary information about the chemical composition, reactivity and structure of well-defined species supported on surfaces.
Chemistry, Issue 88, soft landing, mass selected ions, electrospray, secondary ion mass spectrometry, infrared spectroscopy, organometallic, catalysis
Play Button
Rejection of Fluorescence Background in Resonance and Spontaneous Raman Microspectroscopy
Authors: Zachary J. Smith, Florian Knorr, Cynthia V. Pagba, Sebastian Wachsmann-Hogiu.
Institutions: University of California, Davis, University of California, Davis.
Raman spectroscopy is often plagued by a strong fluorescent background, particularly for biological samples. If a sample is excited with a train of ultrafast pulses, a system that can temporally separate spectrally overlapping signals on a picosecond timescale can isolate promptly arriving Raman scattered light from late-arriving fluorescence light. Here we discuss the construction and operation of a complex nonlinear optical system that uses all-optical switching in the form of a low-power optical Kerr gate to isolate Raman and fluorescence signals. A single 808 nm laser with 2.4 W of average power and 80 MHz repetition rate is split, with approximately 200 mW of 808 nm light being converted to < 5 mW of 404 nm light sent to the sample to excite Raman scattering. The remaining unconverted 808 nm light is then sent to a nonlinear medium where it acts as the pump for the all-optical shutter. The shutter opens and closes in 800 fs with a peak efficiency of approximately 5%. Using this system we are able to successfully separate Raman and fluorescence signals at an 80 MHz repetition rate using pulse energies and average powers that remain biologically safe. Because the system has no spare capacity in terms of optical power, we detail several design and alignment considerations that aid in maximizing the throughput of the system. We also discuss our protocol for obtaining the spatial and temporal overlap of the signal and pump beams within the Kerr medium, as well as a detailed protocol for spectral acquisition. Finally, we report a few representative results of Raman spectra obtained in the presence of strong fluorescence using our time-gating system.
Microbiology, Issue 51, Raman scattering, all-optical gating, nonlinear optics, fluorescence, timeresolved spectroscopy.
Play Button
Strategies for Study of Neuroprotection from Cold-preconditioning
Authors: Heidi M. Mitchell, David M. White, Richard P. Kraig.
Institutions: The University of Chicago Medical Center.
Neurological injury is a frequent cause of morbidity and mortality from general anesthesia and related surgical procedures that could be alleviated by development of effective, easy to administer and safe preconditioning treatments. We seek to define the neural immune signaling responsible for cold-preconditioning as means to identify novel targets for therapeutics development to protect brain before injury onset. Low-level pro-inflammatory mediator signaling changes over time are essential for cold-preconditioning neuroprotection. This signaling is consistent with the basic tenets of physiological conditioning hormesis, which require that irritative stimuli reach a threshold magnitude with sufficient time for adaptation to the stimuli for protection to become evident. Accordingly, delineation of the immune signaling involved in cold-preconditioning neuroprotection requires that biological systems and experimental manipulations plus technical capacities are highly reproducible and sensitive. Our approach is to use hippocampal slice cultures as an in vitro model that closely reflects their in vivo counterparts with multi-synaptic neural networks influenced by mature and quiescent macroglia / microglia. This glial state is particularly important for microglia since they are the principal source of cytokines, which are operative in the femtomolar range. Also, slice cultures can be maintained in vitro for several weeks, which is sufficient time to evoke activating stimuli and assess adaptive responses. Finally, environmental conditions can be accurately controlled using slice cultures so that cytokine signaling of cold-preconditioning can be measured, mimicked, and modulated to dissect the critical node aspects. Cytokine signaling system analyses require the use of sensitive and reproducible multiplexed techniques. We use quantitative PCR for TNF-α to screen for microglial activation followed by quantitative real-time qPCR array screening to assess tissue-wide cytokine changes. The latter is a most sensitive and reproducible means to measure multiple cytokine system signaling changes simultaneously. Significant changes are confirmed with targeted qPCR and then protein detection. We probe for tissue-based cytokine protein changes using multiplexed microsphere flow cytometric assays using Luminex technology. Cell-specific cytokine production is determined with double-label immunohistochemistry. Taken together, this brain tissue preparation and style of use, coupled to the suggested investigative strategies, may be an optimal approach for identifying potential targets for the development of novel therapeutics that could mimic the advantages of cold-preconditioning.
Neuroscience, Issue 43, innate immunity, hormesis, microglia, hippocampus, slice culture, immunohistochemistry, neural-immune, gene expression, real-time PCR
Play Button
The Use of Magnetic Resonance Spectroscopy as a Tool for the Measurement of Bi-hemispheric Transcranial Electric Stimulation Effects on Primary Motor Cortex Metabolism
Authors: Sara Tremblay, Vincent Beaulé, Sébastien Proulx, Louis-Philippe Lafleur, Julien Doyon, Małgorzata Marjańska, Hugo Théoret.
Institutions: University of Montréal, McGill University, University of Minnesota.
Transcranial direct current stimulation (tDCS) is a neuromodulation technique that has been increasingly used over the past decade in the treatment of neurological and psychiatric disorders such as stroke and depression. Yet, the mechanisms underlying its ability to modulate brain excitability to improve clinical symptoms remains poorly understood 33. To help improve this understanding, proton magnetic resonance spectroscopy (1H-MRS) can be used as it allows the in vivo quantification of brain metabolites such as γ-aminobutyric acid (GABA) and glutamate in a region-specific manner 41. In fact, a recent study demonstrated that 1H-MRS is indeed a powerful means to better understand the effects of tDCS on neurotransmitter concentration 34. This article aims to describe the complete protocol for combining tDCS (NeuroConn MR compatible stimulator) with 1H-MRS at 3 T using a MEGA-PRESS sequence. We will describe the impact of a protocol that has shown great promise for the treatment of motor dysfunctions after stroke, which consists of bilateral stimulation of primary motor cortices 27,30,31. Methodological factors to consider and possible modifications to the protocol are also discussed.
Neuroscience, Issue 93, proton magnetic resonance spectroscopy, transcranial direct current stimulation, primary motor cortex, GABA, glutamate, stroke
Play Button
Differential Imaging of Biological Structures with Doubly-resonant Coherent Anti-stokes Raman Scattering (CARS)
Authors: Tyler J. Weeks, Thomas R. Huser.
Institutions: University of California, Davis, University of California, Davis.
Coherent Raman imaging techniques have seen a dramatic increase in activity over the past decade due to their promise to enable label-free optical imaging with high molecular specificity 1. The sensitivity of these techniques, however, is many orders of magnitude weaker than fluorescence, requiring milli-molar molecular concentrations 1,2. Here, we describe a technique that can enable the detection of weak or low concentrations of Raman-active molecules by amplifying their signal with that obtained from strong or abundant Raman scatterers. The interaction of short pulsed lasers in a biological sample generates a variety of coherent Raman scattering signals, each of which carry unique chemical information about the sample. Typically, only one of these signals, e.g. Coherent Anti-stokes Raman scattering (CARS), is used to generate an image while the others are discarded. However, when these other signals, including 3-color CARS and four-wave mixing (FWM), are collected and compared to the CARS signal, otherwise difficult to detect information can be extracted 3. For example, doubly-resonant CARS (DR-CARS) is the result of the constructive interference between two resonant signals 4. We demonstrate how tuning of the three lasers required to produce DR-CARS signals to the 2845 cm-1 CH stretch vibration in lipids and the 2120 cm-1 CD stretching vibration of a deuterated molecule (e.g. deuterated sugars, fatty acids, etc.) can be utilized to probe both Raman resonances simultaneously. Under these conditions, in addition to CARS signals from each resonance, a combined DR-CARS signal probing both is also generated. We demonstrate how detecting the difference between the DR-CARS signal and the amplifying signal from an abundant molecule's vibration can be used to enhance the sensitivity for the weaker signal. We further demonstrate that this approach even extends to applications where both signals are generated from different molecules, such that e.g. using the strong Raman signal of a solvent can enhance the weak Raman signal of a dilute solute.
Cellular Biology, Issue 44, Raman scattering, Four-wave mixing, Coherent anti-Stokes Raman scattering, Microscopy, Coherent Raman Scattering
Play Button
From Voxels to Knowledge: A Practical Guide to the Segmentation of Complex Electron Microscopy 3D-Data
Authors: Wen-Ting Tsai, Ahmed Hassan, Purbasha Sarkar, Joaquin Correa, Zoltan Metlagel, Danielle M. Jorgens, Manfred Auer.
Institutions: Lawrence Berkeley National Laboratory, Lawrence Berkeley National Laboratory, Lawrence Berkeley National Laboratory.
Modern 3D electron microscopy approaches have recently allowed unprecedented insight into the 3D ultrastructural organization of cells and tissues, enabling the visualization of large macromolecular machines, such as adhesion complexes, as well as higher-order structures, such as the cytoskeleton and cellular organelles in their respective cell and tissue context. Given the inherent complexity of cellular volumes, it is essential to first extract the features of interest in order to allow visualization, quantification, and therefore comprehension of their 3D organization. Each data set is defined by distinct characteristics, e.g., signal-to-noise ratio, crispness (sharpness) of the data, heterogeneity of its features, crowdedness of features, presence or absence of characteristic shapes that allow for easy identification, and the percentage of the entire volume that a specific region of interest occupies. All these characteristics need to be considered when deciding on which approach to take for segmentation. The six different 3D ultrastructural data sets presented were obtained by three different imaging approaches: resin embedded stained electron tomography, focused ion beam- and serial block face- scanning electron microscopy (FIB-SEM, SBF-SEM) of mildly stained and heavily stained samples, respectively. For these data sets, four different segmentation approaches have been applied: (1) fully manual model building followed solely by visualization of the model, (2) manual tracing segmentation of the data followed by surface rendering, (3) semi-automated approaches followed by surface rendering, or (4) automated custom-designed segmentation algorithms followed by surface rendering and quantitative analysis. Depending on the combination of data set characteristics, it was found that typically one of these four categorical approaches outperforms the others, but depending on the exact sequence of criteria, more than one approach may be successful. Based on these data, we propose a triage scheme that categorizes both objective data set characteristics and subjective personal criteria for the analysis of the different data sets.
Bioengineering, Issue 90, 3D electron microscopy, feature extraction, segmentation, image analysis, reconstruction, manual tracing, thresholding
Play Button
Determination of Protein-ligand Interactions Using Differential Scanning Fluorimetry
Authors: Mirella Vivoli, Halina R. Novak, Jennifer A. Littlechild, Nicholas J. Harmer.
Institutions: University of Exeter.
A wide range of methods are currently available for determining the dissociation constant between a protein and interacting small molecules. However, most of these require access to specialist equipment, and often require a degree of expertise to effectively establish reliable experiments and analyze data. Differential scanning fluorimetry (DSF) is being increasingly used as a robust method for initial screening of proteins for interacting small molecules, either for identifying physiological partners or for hit discovery. This technique has the advantage that it requires only a PCR machine suitable for quantitative PCR, and so suitable instrumentation is available in most institutions; an excellent range of protocols are already available; and there are strong precedents in the literature for multiple uses of the method. Past work has proposed several means of calculating dissociation constants from DSF data, but these are mathematically demanding. Here, we demonstrate a method for estimating dissociation constants from a moderate amount of DSF experimental data. These data can typically be collected and analyzed within a single day. We demonstrate how different models can be used to fit data collected from simple binding events, and where cooperative binding or independent binding sites are present. Finally, we present an example of data analysis in a case where standard models do not apply. These methods are illustrated with data collected on commercially available control proteins, and two proteins from our research program. Overall, our method provides a straightforward way for researchers to rapidly gain further insight into protein-ligand interactions using DSF.
Biophysics, Issue 91, differential scanning fluorimetry, dissociation constant, protein-ligand interactions, StepOne, cooperativity, WcbI.
Play Button
Free Radicals in Chemical Biology: from Chemical Behavior to Biomarker Development
Authors: Chryssostomos Chatgilialoglu, Carla Ferreri, Annalisa Masi, Michele Melchiorre, Anna Sansone, Michael A. Terzidis, Armida Torreggiani.
Institutions: Consiglio Nazionale delle Ricerche.
The involvement of free radicals in life sciences has constantly increased with time and has been connected to several physiological and pathological processes. This subject embraces diverse scientific areas, spanning from physical, biological and bioorganic chemistry to biology and medicine, with applications to the amelioration of quality of life, health and aging. Multidisciplinary skills are required for the full investigation of the many facets of radical processes in the biological environment and chemical knowledge plays a crucial role in unveiling basic processes and mechanisms. We developed a chemical biology approach able to connect free radical chemical reactivity with biological processes, providing information on the mechanistic pathways and products. The core of this approach is the design of biomimetic models to study biomolecule behavior (lipids, nucleic acids and proteins) in aqueous systems, obtaining insights of the reaction pathways as well as building up molecular libraries of the free radical reaction products. This context can be successfully used for biomarker discovery and examples are provided with two classes of compounds: mono-trans isomers of cholesteryl esters, which are synthesized and used as references for detection in human plasma, and purine 5',8-cyclo-2'-deoxyribonucleosides, prepared and used as reference in the protocol for detection of such lesions in DNA samples, after ionizing radiations or obtained from different health conditions.
Chemistry, Issue 74, Biochemistry, Chemical Engineering, Chemical Biology, chemical analysis techniques, chemistry (general), life sciences, radiation effects (biological, animal and plant), biomarker, biomimetic chemistry, free radicals, trans lipids, cyclopurine lesions, DNA, chromatography, spectroscopy, synthesis
Play Button
Graphene Coatings for Biomedical Implants
Authors: Ramakrishna Podila, Thomas Moore, Frank Alexis, Apparao Rao.
Institutions: Clemson University, East Carolina University, Clemson University, Clemson University.
Atomically smooth graphene as a surface coating has potential to improve implant properties. This demonstrates a method for coating nitinol alloys with nanometer thick layers of graphene for applications as a stent material. Graphene was grown on copper substrates via chemical vapor deposition and then transferred onto nitinol substrates. In order to understand how the graphene coating could change biological response, cell viability of rat aortic endothelial cells and rat aortic smooth muscle cells was investigated. Moreover, the effect of graphene-coatings on cell adhesion and morphology was examined with fluorescent confocal microscopy. Cells were stained for actin and nuclei, and there were noticeable differences between pristine nitinol samples compared to graphene-coated samples. Total actin expression from rat aortic smooth muscle cells was found using western blot. Protein adsorption characteristics, an indicator for potential thrombogenicity, were determined for serum albumin and fibrinogen with gel electrophoresis. Moreover, the transfer of charge from fibrinogen to substrate was deduced using Raman spectroscopy. It was found that graphene coating on nitinol substrates met the functional requirements for a stent material and improved the biological response compared to uncoated nitinol. Thus, graphene-coated nitinol is a viable candidate for a stent material.
Biomedical Engineering, Issue 73, Bioengineering, Medicine, Biophysics, Materials Science, Physics, Pharmacology, Toxicology, Surgery, Chemistry and Materials (General), graphene, biomedical implants, surface modification, chemical vapor deposition, protein expression, confocal microscopy, implants, stents, clinical
Play Button
Probing and Mapping Electrode Surfaces in Solid Oxide Fuel Cells
Authors: Kevin S. Blinn, Xiaxi Li, Mingfei Liu, Lawrence A. Bottomley, Meilin Liu.
Institutions: Georgia Institute of Technology , Georgia Institute of Technology .
Solid oxide fuel cells (SOFCs) are potentially the most efficient and cost-effective solution to utilization of a wide variety of fuels beyond hydrogen 1-7. The performance of SOFCs and the rates of many chemical and energy transformation processes in energy storage and conversion devices in general are limited primarily by charge and mass transfer along electrode surfaces and across interfaces. Unfortunately, the mechanistic understanding of these processes is still lacking, due largely to the difficulty of characterizing these processes under in situ conditions. This knowledge gap is a chief obstacle to SOFC commercialization. The development of tools for probing and mapping surface chemistries relevant to electrode reactions is vital to unraveling the mechanisms of surface processes and to achieving rational design of new electrode materials for more efficient energy storage and conversion2. Among the relatively few in situ surface analysis methods, Raman spectroscopy can be performed even with high temperatures and harsh atmospheres, making it ideal for characterizing chemical processes relevant to SOFC anode performance and degradation8-12. It can also be used alongside electrochemical measurements, potentially allowing direct correlation of electrochemistry to surface chemistry in an operating cell. Proper in situ Raman mapping measurements would be useful for pin-pointing important anode reaction mechanisms because of its sensitivity to the relevant species, including anode performance degradation through carbon deposition8, 10, 13, 14 ("coking") and sulfur poisoning11, 15 and the manner in which surface modifications stave off this degradation16. The current work demonstrates significant progress towards this capability. In addition, the family of scanning probe microscopy (SPM) techniques provides a special approach to interrogate the electrode surface with nanoscale resolution. Besides the surface topography that is routinely collected by AFM and STM, other properties such as local electronic states, ion diffusion coefficient and surface potential can also be investigated17-22. In this work, electrochemical measurements, Raman spectroscopy, and SPM were used in conjunction with a novel test electrode platform that consists of a Ni mesh electrode embedded in an yttria-stabilized zirconia (YSZ) electrolyte. Cell performance testing and impedance spectroscopy under fuel containing H2S was characterized, and Raman mapping was used to further elucidate the nature of sulfur poisoning. In situ Raman monitoring was used to investigate coking behavior. Finally, atomic force microscopy (AFM) and electrostatic force microscopy (EFM) were used to further visualize carbon deposition on the nanoscale. From this research, we desire to produce a more complete picture of the SOFC anode.
Materials Science, Issue 67, Chemistry, Electrical Engineering, Physics, electrochemistry, catalysts (chemical), spectroscopic chemical analysis (application), microscopes, Fuel cell, Raman, AFM, SOFC, Surface, Electrode
Play Button
An in vivo Assay to Test Blood Vessel Permeability
Authors: Maria Radu, Jonathan Chernoff.
Institutions: Fox Chase Cancer Center .
This method is based on the intravenous injection of Evans Blue in mice as the test animal model. Evans blue is a dye that binds albumin. Under physiologic conditions the endothelium is impermeable to albumin, so Evans blue bound albumin remains restricted within blood vessels. In pathologic conditions that promote increased vascular permeability endothelial cells partially lose their close contacts and the endothelium becomes permeable to small proteins such as albumin. This condition allows for extravasation of Evans Blue in tissues. A healthy endothelium prevents extravasation of the dye in the neighboring vascularized tissues. Organs with increased permeability will show significantly increased blue coloration compared to organs with intact endothelium. The level of vascular permeability can be assessed by simple visualization or by quantitative measurement of the dye incorporated per milligram of tissue of control versus experimental animal/tissue. Two powerful aspects of this assay are its simplicity and quantitative characteristics. Evans Blue dye can be extracted from tissues by incubating a specific amount of tissue in formamide. Evans Blue absorbance maximum is at 620 nm and absorbance minimum is at 740 nm. By using a standard curve for Evans Blue, optical density measurements can be converted into milligram dye captured per milligram of tissue. Statistical analysis should be used to assess significant differences in vascular permeability.
Medicine, Issue 73, Immunology, Physiology, Anatomy, Surgery, Hematology, Blood Vessels, Endothelium, Vascular, Vascular Cell Adhesion Molecule-1, permeability, in vivo, Evans Blue, Miles assay, assay, intravenous injection, mouse, animal model
Play Button
Quantifying Glomerular Permeability of Fluorescent Macromolecules Using 2-Photon Microscopy in Munich Wistar Rats
Authors: Ruben M. Sandoval, Bruce A. Molitoris.
Institutions: Indiana University School of Medicine.
Kidney diseases involving urinary loss of large essential macromolecules, such as serum albumin, have long been thought to be caused by alterations in the permeability barrier comprised of podocytes, vascular endothelial cells, and a basement membrane working in unison. Data from our laboratory using intravital 2-photon microscopy revealed a more permeable glomerular filtration barrier (GFB) than previously thought under physiologic conditions, with retrieval of filtered albumin occurring in an early subset of cells called proximal tubule cells (PTC)1,2,3. Previous techniques used to study renal filtration and establishing the characteristic of the filtration barrier involved micropuncture of the lumen of these early tubular segments with sampling of the fluid content and analysis4. These studies determined albumin concentration in the luminal fluid to be virtually non-existent; corresponding closely to what is normally detected in the urine. However, characterization of dextran polymers with defined sizes by this technique revealed those of a size similar to serum albumin had higher levels in the tubular lumen and urine; suggesting increased permeability5. Herein is a detailed outline of the technique used to directly visualize and quantify glomerular fluorescent albumin permeability in vivo. This method allows for detection of filtered albumin across the filtration barrier into Bowman's space (the initial chamber of urinary filtration); and also allows quantification of albumin reabsorption by proximal tubules and visualization of subsequent albumin transcytosis6. The absence of fluorescent albumin along later tubular segments en route to the bladder highlights the efficiency of the retrieval pathway in the earlier proximal tubule segments. Moreover, when this technique was applied to determine permeability of dextrans having a similar size to albumin virtually identical permeability values were reported2. These observations directly support the need to expand the focus of many proteinuric renal diseases to included alterations in proximal tubule cell reclamation.
Medicine, Issue 74, Biomedical Engineering, Molecular Biology, Cellular Biology, Anatomy, Physiology, Surgery, Nephrology, Kidney Diseases, Two-photon microscopy, Kidney, Glomerulus, Glomerular Sieving Coefficient (GSC), Permeability, Proximal Tubule, Proteinuria, macromolecules, 2 Photon, microscopy, intravital imaging, munich wistar rat, animal model
Play Button
Linearization of the Bradford Protein Assay
Authors: Orna Ernst, Tsaffrir Zor.
Institutions: Tel Aviv University.
Determination of microgram quantities of protein in the Bradford Coomassie brilliant blue assay is accomplished by measurement of absorbance at 590 nm. This most common assay enables rapid and simple protein quantification in cell lysates, cellular fractions, or recombinant protein samples, for the purpose of normalization of biochemical measurements. However, an intrinsic nonlinearity compromises the sensitivity and accuracy of this method. It is shown that under standard assay conditions, the ratio of the absorbance measurements at 590 nm and 450 nm is strictly linear with protein concentration. This simple procedure increases the accuracy and improves the sensitivity of the assay about 10-fold, permitting quantification down to 50 ng of bovine serum albumin. Furthermore, the interference commonly introduced by detergents that are used to create the cell lysates is greatly reduced by the new protocol. A linear equation developed on the basis of mass action and Beer's law perfectly fits the experimental data.
Cellular Biology, Issue 38, Bradford, protein assay, protein quantification, Coomassie brilliant blue
Play Button
Basics of Multivariate Analysis in Neuroimaging Data
Authors: Christian Georg Habeck.
Institutions: Columbia University.
Multivariate analysis techniques for neuroimaging data have recently received increasing attention as they have many attractive features that cannot be easily realized by the more commonly used univariate, voxel-wise, techniques1,5,6,7,8,9. Multivariate approaches evaluate correlation/covariance of activation across brain regions, rather than proceeding on a voxel-by-voxel basis. Thus, their results can be more easily interpreted as a signature of neural networks. Univariate approaches, on the other hand, cannot directly address interregional correlation in the brain. Multivariate approaches can also result in greater statistical power when compared with univariate techniques, which are forced to employ very stringent corrections for voxel-wise multiple comparisons. Further, multivariate techniques also lend themselves much better to prospective application of results from the analysis of one dataset to entirely new datasets. Multivariate techniques are thus well placed to provide information about mean differences and correlations with behavior, similarly to univariate approaches, with potentially greater statistical power and better reproducibility checks. In contrast to these advantages is the high barrier of entry to the use of multivariate approaches, preventing more widespread application in the community. To the neuroscientist becoming familiar with multivariate analysis techniques, an initial survey of the field might present a bewildering variety of approaches that, although algorithmically similar, are presented with different emphases, typically by people with mathematics backgrounds. We believe that multivariate analysis techniques have sufficient potential to warrant better dissemination. Researchers should be able to employ them in an informed and accessible manner. The current article is an attempt at a didactic introduction of multivariate techniques for the novice. A conceptual introduction is followed with a very simple application to a diagnostic data set from the Alzheimer s Disease Neuroimaging Initiative (ADNI), clearly demonstrating the superior performance of the multivariate approach.
JoVE Neuroscience, Issue 41, fMRI, PET, multivariate analysis, cognitive neuroscience, clinical neuroscience
Play Button
Intravital Microscopy of the Mouse Brain Microcirculation using a Closed Cranial Window
Authors: Pedro Cabrales, Leonardo J. M. Carvalho.
Institutions: University of California, San Diego, La Jolla Bioengineering Institute.
This experimental model was designed to assess the mouse pial microcirculation during acute and chronic, physiological and pathophysiological hemodynamic, inflammatory and metabolic conditions, using in vivo fluorescence microscopy. A closed cranial window is placed over the left parieto-occipital cortex of the mice. Local microcirculation is recorded in real time through the window using epi and fluorescence illumination, and measurements of vessels diameters and red blood cell (RBC) velocities are performed. RBC velocity is measured using real-time cross-correlation and/or fluorescent-labeled erythrocytes. Leukocyte and platelet adherence to pial vessels and assessment of perfusion and vascular leakage are made with the help of fluorescence-labeled markers such as Albumin-FITC and anti-CD45-TxR antibodies. Microcirculation can be repeatedly video-recorded over several days. We used for the first time the close window brain intravital microscopy to study the pial microcirculation to follow dynamic changes during the course of Plasmodium berghei ANKA infection in mice and show that expression of CM is associated with microcirculatory dysfunctions characterized by vasoconstriction, profound decrease in blood flow and eventually vascular collapse.
Neuroscience, Issue 45, Brain hemodynamics, blood flow, microcirculation, vascular morphology, leukocyte adherence, cerebral malaria
Play Button
Concentration Determination of Nucleic Acids and Proteins Using the Micro-volume Bio-spec Nano Spectrophotometer
Authors: Suja Sukumaran.
Institutions: Scientific Instruments.
Nucleic Acid quantitation procedures have advanced significantly in the last three decades. More and more, molecular biologists require consistent small-volume analysis of nucleic acid samples for their experiments. The BioSpec-nano provides a potential solution to the problems of inaccurate, non-reproducible results, inherent in current DNA quantitation methods, via specialized optics and a sensitive PDA detector. The BioSpec-nano also has automated functionality such that mounting, measurement, and cleaning are done by the instrument, thereby eliminating tedious, repetitive, and inconsistent placement of the fiber optic element and manual cleaning. In this study, data is presented on the quantification of DNA and protein, as well as on measurement reproducibility and accuracy. Automated sample contact and rapid scanning allows measurement in three seconds, resulting in excellent throughput. Data analysis is carried out using the built-in features of the software. The formula used for calculating DNA concentration is: Sample Concentration = DF · (OD260-OD320)· NACF (1) Where DF = sample dilution factor and NACF = nucleic acid concentration factor. The Nucleic Acid concentration factor is set in accordance with the analyte selected1. Protein concentration results can be expressed as μg/ mL or as moles/L by entering e280 and molecular weight values respectively. When residue values for Tyr, Trp and Cysteine (S-S bond) are entered in the e280Calc tab, the extinction coefficient values are calculated as e280 = 5500 x (Trp residues) + 1490 x (Tyr residues) + 125 x (cysteine S-S bond). The e280 value is used by the software for concentration calculation. In addition to concentration determination of nucleic acids and protein, the BioSpec-nano can be used as an ultra micro-volume spectrophotometer for many other analytes or as a standard spectrophotometer using 5 mm pathlength cells.
Molecular Biology, Issue 48, Nucleic acid quantitation, protein quantitation, micro-volume analysis, label quantitation
Play Button
Polymer Microarrays for High Throughput Discovery of Biomaterials
Authors: Andrew L. Hook, Chien-Yi Chang, Jing Yang, David J. Scurr, Robert Langer, Daniel G. Anderson, Steve Atkinson, Paul Williams, Martyn C. Davies, Morgan R. Alexander.
Institutions: University of Nottingham , University of Nottingham , Massachusetts Institute of Technology.
The discovery of novel biomaterials that are optimized for a specific biological application is readily achieved using polymer microarrays, which allows a combinatorial library of materials to be screened in a parallel, high throughput format1. Herein is described the formation and characterization of a polymer microarray using an on-chip photopolymerization technique 2. This involves mixing monomers at varied ratios to produce a library of monomer solutions, transferring the solution to a glass slide format using a robotic printing device and curing with UV irradiation. This format is readily amenable to many biological assays, including stem cell attachment and proliferation, cell sorting and low bacterial adhesion, allowing the ready identification of 'hit' materials that fulfill a specific biological criterion3-5. Furthermore, the use of high throughput surface characterization (HTSC) allows the biological performance to be correlated with physio-chemical properties, hence elucidating the biological-material interaction6. HTSC makes use of water contact angle (WCA) measurements, atomic force microscopy (AFM), X-ray photoelectron spectroscopy (XPS) and time-of-flight secondary ion mass spectrometry (ToF-SIMS). In particular, ToF-SIMS provides a chemically rich analysis of the sample that can be used to correlate the cell response with a molecular moiety. In some cases, the biological performance can be predicted from the ToF-SIMS spectra, demonstrating the chemical dependence of a biological-material interaction, and informing the development of hit materials5,3.
Bioengineering, Issue 59, Materials discovery, Surface characterization, Polymer library, High throughput, Cell attachment
Play Button
Improving IV Insulin Administration in a Community Hospital
Authors: Michael C. Magee.
Institutions: Wyoming Medical Center.
Diabetes mellitus is a major independent risk factor for increased morbidity and mortality in the hospitalized patient, and elevated blood glucose concentrations, even in non-diabetic patients, predicts poor outcomes.1-4 The 2008 consensus statement by the American Association of Clinical Endocrinologists (AACE) and the American Diabetes Association (ADA) states that "hyperglycemia in hospitalized patients, irrespective of its cause, is unequivocally associated with adverse outcomes."5 It is important to recognize that hyperglycemia occurs in patients with known or undiagnosed diabetes as well as during acute illness in those with previously normal glucose tolerance. The Normoglycemia in Intensive Care Evaluation-Survival Using Glucose Algorithm Regulation (NICE-SUGAR) study involved over six thousand adult intensive care unit (ICU) patients who were randomized to intensive glucose control or conventional glucose control.6 Surprisingly, this trial found that intensive glucose control increased the risk of mortality by 14% (odds ratio, 1.14; p=0.02). In addition, there was an increased prevalence of severe hypoglycemia in the intensive control group compared with the conventional control group (6.8% vs. 0.5%, respectively; p<0.001). From this pivotal trial and two others,7,8 Wyoming Medical Center (WMC) realized the importance of controlling hyperglycemia in the hospitalized patient while avoiding the negative impact of resultant hypoglycemia. Despite multiple revisions of an IV insulin paper protocol, analysis of data from usage of the paper protocol at WMC shows that in terms of achieving normoglycemia while minimizing hypoglycemia, results were suboptimal. Therefore, through a systematical implementation plan, monitoring of patient blood glucose levels was switched from using a paper IV insulin protocol to a computerized glucose management system. By comparing blood glucose levels using the paper protocol to that of the computerized system, it was determined, that overall, the computerized glucose management system resulted in more rapid and tighter glucose control than the traditional paper protocol. Specifically, a substantial increase in the time spent within the target blood glucose concentration range, as well as a decrease in the prevalence of severe hypoglycemia (BG < 40 mg/dL), clinical hypoglycemia (BG < 70 mg/dL), and hyperglycemia (BG > 180 mg/dL), was witnessed in the first five months after implementation of the computerized glucose management system. The computerized system achieved target concentrations in greater than 75% of all readings while minimizing the risk of hypoglycemia. The prevalence of hypoglycemia (BG < 70 mg/dL) with the use of the computer glucose management system was well under 1%.
Medicine, Issue 64, Physiology, Computerized glucose management, Endotool, hypoglycemia, hyperglycemia, diabetes, IV insulin, paper protocol, glucose control
Play Button
Real-time Monitoring of Ligand-receptor Interactions with Fluorescence Resonance Energy Transfer
Authors: Navneet Dogra, Julia C. Reyes, Nishi Garg, Punit Kohli.
Institutions: Southern Illinois University.
FRET is a process whereby energy is non-radiatively transferred from an excited donor molecule to a ground-state acceptor molecule through long-range dipole-dipole interactions1. In the present sensing assay, we utilize an interesting property of PDA: blue-shift in the UV-Vis electronic absorption spectrum of PDA (Figure 1) after an analyte interacts with receptors attached to PDA2,3,4,7. This shift in the PDA absorption spectrum provides changes in the spectral overlap (J) between PDA (acceptor) and rhodamine (donor) that leads to changes in the FRET efficiency. Thus, the interactions between analyte (ligand) and receptors are detected through FRET between donor fluorophores and PDA. In particular, we show the sensing of a model protein molecule streptavidin. We also demonstrate the covalent-binding of bovine serum albumin (BSA) to the liposome surface with FRET mechanism. These interactions between the bilayer liposomes and protein molecules can be sensed in real-time. The proposed method is a general method for sensing small chemical and large biochemical molecules. Since fluorescence is intrinsically more sensitive than colorimetry, the detection limit of the assay can be in sub-nanomolar range or lower8. Further, PDA can act as a universal acceptor in FRET, which means that multiple sensors can be developed with PDA (acceptor) functionalized with donors and different receptors attached on the surface of PDA liposomes.
Biochemistry, Issue 66, Molecular Biology, Chemistry, Physics, Fluorescence Resonance Energy Transfer (FRET), Polydiacetylene (PDA), Biosensor, Liposome, Sensing
Play Button
Non-contact, Label-free Monitoring of Cells and Extracellular Matrix using Raman Spectroscopy
Authors: Miriam Votteler, Daniel A. Carvajal Berrio, Marieke Pudlas, Heike Walles, Katja Schenke-Layland.
Institutions: Eberhard Karls University, Tübingen, Fraunhofer Institute of Interfacial Engineering and Biotechnology (IGB) Stuttgart, Germany, University of Stuttgart, Germany, Julius-Maximillians University, Würzburg, Germany.
Non-destructive, non-contact and label-free technologies to monitor cell and tissue cultures are needed in the field of biomedical research.1-5 However, currently available routine methods require processing steps and alter sample integrity. Raman spectroscopy is a fast method that enables the measurement of biological samples without the need for further processing steps. This laser-based technology detects the inelastic scattering of monochromatic light.6 As every chemical vibration is assigned to a specific Raman band (wavenumber in cm-1), each biological sample features a typical spectral pattern due to their inherent biochemical composition.7-9 Within Raman spectra, the peak intensities correlate with the amount of the present molecular bonds.1 Similarities and differences of the spectral data sets can be detected by employing a multivariate analysis (e.g. principal component analysis (PCA)).10 Here, we perform Raman spectroscopy of living cells and native tissues. Cells are either seeded on glass bottom dishes or kept in suspension under normal cell culture conditions (37 °C, 5% CO2) before measurement. Native tissues are dissected and stored in phosphate buffered saline (PBS) at 4 °C prior measurements. Depending on our experimental set up, we then either focused on the cell nucleus or extracellular matrix (ECM) proteins such as elastin and collagen. For all studies, a minimum of 30 cells or 30 random points of interest within the ECM are measured. Data processing steps included background subtraction and normalization.
Bioengineering, Issue 63, Raman spectroscopy, label-free analysis, living cells, extracellular matrix, tissue engineering
Play Button
Tangential Flow Ultrafiltration: A “Green” Method for the Size Selection and Concentration of Colloidal Silver Nanoparticles
Authors: Catherine B. Anders, Joshua D. Baker, Adam C. Stahler, Austin J. Williams, Jackie N. Sisco, John C. Trefry, Dawn P. Wooley, Ioana E. Pavel Sizemore.
Institutions: Wright State University, Wright State University.
Nowadays, AgNPs are extensively used in the manufacture of consumer products,1 water disinfectants,2 therapeutics,1, 3 and biomedical devices4 due to their powerful antimicrobial properties.3-6 These nanoparticle applications are strongly influenced by the AgNP size and aggregation state. Many challenges exist in the controlled fabrication7 and size-based isolation4,8 of unfunctionalized, homogenous AgNPs that are free from chemically aggressive capping/stabilizing agents or organic solvents.7-13 Limitations emerge from the toxicity of reagents, high costs or reduced efficiency of the AgNP synthesis or isolation methods (e.g., centrifugation, size-dependent solubility, size-exclusion chromatography, etc.).10,14-18 To overcome this, we recently showed that TFU permits greater control over the size, concentration and aggregation state of Creighton AgNPs (300 ml of 15.3 μg ml-1 down to 10 ml of 198.7 μg ml-1) than conventional methods of isolation such as ultracentrifugation.19 TFU is a recirculation method commonly used for the weight-based isolation of proteins, viruses and cells.20,21 Briefly, the liquid sample is passed through a series of hollow fiber membranes with pore size ranging from 1,000 kD to 10 kD. Smaller suspended or dissolved constituents in the sample will pass through the porous barrier together with the solvent (filtrate), while the larger constituents are retained (retentate). TFU may be considered a "green" method as it neither damages the sample nor requires additional solvent to eliminate toxic excess reagents and byproducts. Furthermore, TFU may be applied to a large variety of nanoparticles as both hydrophobic and hydrophilic filters are available. The two main objectives of this study were: 1) to illustrate the experimental aspects of the TFU approach through an invited video experience and 2) to demonstrate the feasibility of the TFU method for larger volumes of colloidal nanoparticles and smaller volumes of retentate. First, unfuctionalized AgNPs (4 L, 15.2 μg ml-1) were synthesized using the well-established Creighton method22,23 by the reduction of AgNO3 with NaBH4. AgNP polydispersity was then minimized via a 3-step TFU using a 50-nm filter (460 cm2) to remove AgNPs and AgNP-aggregates larger than 50 nm, followed by two 100-kD (200 cm2 and 20 cm2) filters to concentrate the AgNPs. Representative samples were characterized using transmission electron microscopy, UV-Vis absorption spectrophotometry, Raman spectroscopy, and inductively coupled plasma optical emission spectroscopy. The final retentate consisted of highly concentrated (4 ml, 8,539.9 μg ml-1) yet lowly aggregated and homogeneous AgNPs of 1-20 nm in diameter. This corresponds to a silver concentration yield of about 62%.
Chemistry, Issue 68, Biomedical Engineering, Chemical Engineering, Nanotechnology, silver nanoparticles, size selection, concentration, tangential flow ultrafiltration
Play Button
Microvolume Protein Concentration Determination using the NanoDrop 2000c Spectrophotometer
Authors: Philippe Desjardins, Joel B. Hansen, Michael Allen.
Institutions: Thermo Scientific NanoDrop Products.
Traditional spectrophotometry requires placing samples into cuvettes or capillaries. This is often impractical due to the limited sample volumes often used for protein analysis. The Thermo Scientific NanoDrop 2000c Spectrophotometer solves this issue with an innovative sample retention system that holds microvolume samples between two measurement surfaces using the surface tension properties of liquids, enabling the quantification of samples in volumes as low as 0.5-2 μL. The elimination of cuvettes or capillaries allows real time changes in path length, which reduces the measurement time while greatly increasing the dynamic range of protein concentrations that can be measured. The need for dilutions is also eliminated, and preparations for sample quantification are relatively easy as the measurement surfaces can be simply wiped with laboratory wipe. This video article presents modifications to traditional protein concentration determination methods for quantification of microvolume amounts of protein using A280 absorbance readings or the BCA colorimetric assay.
Basic Protocols, Issue 33, NanoDrop, protein measurement, protein concentration, spectrophotometer, A280, UV/Vis, BCA, microvolume, microsample, proteomics
Copyright © JoVE 2006-2015. All Rights Reserved.
Policies | License Agreement | ISSN 1940-087X
simple hit counter

What is Visualize?

JoVE Visualize is a tool created to match the last 5 years of PubMed publications to methods in JoVE's video library.

How does it work?

We use abstracts found on PubMed and match them to JoVE videos to create a list of 10 to 30 related methods videos.

Video X seems to be unrelated to Abstract Y...

In developing our video relationships, we compare around 5 million PubMed articles to our library of over 4,500 methods videos. In some cases the language used in the PubMed abstracts makes matching that content to a JoVE video difficult. In other cases, there happens not to be any content in our video library that is relevant to the topic of a given abstract. In these cases, our algorithms are trying their best to display videos with relevant content, which can sometimes result in matched videos with only a slight relation.