The full potential of magnetic resonance spectroscopic imaging (MRSI) is often limited by localization artifacts, motion-related artifacts, scanner instabilities, and long measurement times. Localized adiabatic selective refocusing (LASER) provides accurate B1-insensitive spatial excitation even at high magnetic fields. Spiral encoding accelerates MRSI acquisition, and thus, enables 3D-coverage without compromising spatial resolution. Real-time position- and shim/frequency-tracking using MR navigators correct motion- and scanner instability-related artifacts.
We propose a novel method, fMRI-Informed Regional Estimation (FIRE), which utilizes information from fMRI in E/MEG source reconstruction. FIRE takes advantage of the spatial alignment between the neural and the vascular activities, while allowing for substantial differences in their dynamics. Furthermore, with a region-based approach, FIRE estimates the model parameters for each region independently. Hence, it can be efficiently applied on a dense grid of source locations. The optimization procedure at the core of FIRE is related to the re-weighted minimum-norm algorithms.
Previous studies have suggested that speech motor system mediates suppression by silent lipreading of electromagnetic auditory cortex responses to pure tones at about 100 ms from sound onset. We used sparse sampling functional magnetic resonance imaging (fMRI) at 3 Tesla to map auditory-cortex foci of suppressant effects during silent lipreading and covert self-production. Streams of video clips were presented simultaneously with 1/3 octave noise bursts centered at 250 Hz (low frequency, LF) or 2000 Hz (mid-frequency, MF), or during no auditory stimulation.
Cognitive dysfunctions may be a significant factor in drug-seeking behavior, reducing the efficiency of rehabilitation in opioid dependence. Neurophysiological basis of these dysfunctions is poorly understood. 21 opioid-dependent patients and 15 healthy controls with no experience of illicit drugs were studied with simultaneous electroencephalography (EEG) and magnetoencephalography (MEG). Among opioid dependents 15 were benzodiazepine co-dependent.
Neuronal activation sequence information is essential for understanding brain functions. Extracting such timing information from blood oxygenation level dependent (BOLD) fMRI is confounded by interregional neurovascular differences and poorly understood relations between BOLD and electrophysiological response delays. Here, we recorded whole-head BOLD fMRI at 100 ms resolution and magnetoencephalography (MEG) during a visuomotor reaction-time task.
Neurophysiological animal models suggest that anterior auditory cortex (AC) areas process sound identity information, whereas posterior ACs specialize in sound location processing. In humans, inconsistent neuroimaging results and insufficient causal evidence have challenged the existence of such parallel AC organization. Here we transiently inhibit bilateral anterior or posterior AC areas using MRI-guided paired-pulse transcranial magnetic stimulation (TMS) while subjects listen to Reference/Probe sound pairs and perform either sound location or identity discrimination tasks.
Using simultaneous acquisition from multiple channels of a radio-frequency (RF) coil array, magnetic resonance inverse imaging (InI) achieves functional MRI acquisitions at a rate of 100ms per whole-brain volume. InI accelerates the scan by leaving out partition encoding steps and reconstructs images by solving under-determined inverse problems using RF coil sensitivity information. Hence, the correlated spatial information available in the coil array causes spatial blurring in the InI reconstruction.
Based on the infamous left-lateralized neglect syndrome, one might hypothesize that the dominating right parietal cortex has a bilateral representation of space, whereas the left parietal cortex represents only the contralateral right hemispace. Whether this principle applies to human auditory attention is not yet fully clear. Here, we explicitly tested the differences in cross-hemispheric functional coupling between the intraparietal sulcus (IPS) and auditory cortex (AC) using combined magnetoencephalography (MEG), EEG, and functional MRI (fMRI).
Natural consonant-vowel syllables are reliably classified by most listeners as voiced or voiceless. However, our previous research [Liederman, J., Frye, R., Fisher, J.M., Greenwood, K., Alexander, R., 2005. A temporally dynamic context effect that disrupts voice onset time discrimination of rapidly successive stimuli. Psychon Bull Rev. 12, 380-386] suggests that among synthetic stimuli varying systematically in voice onset time (VOT), syllables that are classified reliably as voiceless are nonetheless perceived differently within and between listeners.