Elsevier

NeuroImage

Volume 54, Issue 3, 1 February 2011, Pages 2198-2217
NeuroImage

Linked independent component analysis for multimodal data fusion

https://doi.org/10.1016/j.neuroimage.2010.09.073Get rights and content

Abstract

In recent years, neuroimaging studies have increasingly been acquiring multiple modalities of data and searching for task- or disease-related changes in each modality separately. A major challenge in analysis is to find systematic approaches for fusing these differing data types together to automatically find patterns of related changes across multiple modalities, when they exist. Independent Component Analysis (ICA) is a popular unsupervised learning method that can be used to find the modes of variation in neuroimaging data across a group of subjects. When multimodal data is acquired for the subjects, ICA is typically performed separately on each modality, leading to incompatible decompositions across modalities. Using a modular Bayesian framework, we develop a novel “Linked ICA” model for simultaneously modelling and discovering common features across multiple modalities, which can potentially have completely different units, signal- and contrast-to-noise ratios, voxel counts, spatial smoothnesses and intensity distributions. Furthermore, this general model can be configured to allow tensor ICA or spatially-concatenated ICA decompositions, or a combination of both at the same time. Linked ICA automatically determines the optimal weighting of each modality, and also can detect single-modality structured components when present. This is a fully probabilistic approach, implemented using Variational Bayes. We evaluate the method on simulated multimodal data sets, as well as on a real data set of Alzheimer's patients and age-matched controls that combines two very different types of structural MRI data: morphological data (grey matter density) and diffusion data (fractional anisotropy, mean diffusivity, and tensor mode).

Research Highlights

►Linked ICA fuses multimodal data and finds patterns of related change across modalities. ►Bayesian approach auto adapts to each modality's signal properties (CNR, smoothness). ►Flexible framework allows combination of tensor ICA and spatial-concatenation ICA. ►Improvements over spatial-concat ICA on sims and on real DTI + VBM group data.

Introduction

One of the greatest strengths of MR neuroimaging is its flexibility; by using different pulse sequences in a single scanning session, one can acquire information about the subject's tissue volume and morphology (using high-resolution structural scans), functional activity (using BOLD FMRI), white matter integrity (using diffusion-weighted imaging), perfusion (using ASL), and other distinct acquisition types. The result of this is that many recent studies have acquired these multimodal MRI data sets for each subject and analysed them separately to find changes in different aspects of the brain. For example, several recent studies have used structural and diffusion tensor imaging (DTI) to find changes in grey matter density and white matter tracts that are related to schizophrenia (Douaud et al., 2007) or learning (Scholz et al., 2009). Other possible combinations are DTI and task-related FMRI (Watkins et al., 2008) or structural, diffusion, and resting-state FMRI (Filippini et al., 2009).

A major challenge is to find systematic approaches for fusing data across multiple MRI modalities, in order to find any patterns of related change that may be present. We develop a model based on Bayesian ICA to extract linked components from multimodal data, using as inputs the subject-wise contrast images from modality-specific analyses. For example, these inputs could be GLM contrasts from FMRI, cortical thickness or VBM maps from structural MRI, and skeletonised tensor measures from diffusion-weighted imaging. ICA is a particularly effective model for finding meaningful, spatially-independent components in an unsupervised setting because it searches for non-Gaussian spatial sources that are likely to represent real structured features in the data. This is because linear mixing processes tend to turn non-Gaussian independent sources into more Gaussian observed signals, so seeking non-Gaussianity is an unsupervised way of isolating the original independent sources.

Standard ICA decompositions treat the input data as a 2D matrix, typically voxels × timepoints or voxels × subjects. Multimodal data does not naturally fit into this form and there are a number of different configurations one could consider for performing combined ICA on multimodal data:

  • Separate ICA analysis of each modality reveals the salient features for each modality. Since some of these features are caused by distributed neurological variations they could be visible (to varying degrees) in all modalities, with similar subject-courses.

    Corresponding components can then be matched up using heuristics; however there is no guarantee that components with strongly correlated subject-courses will be extracted, for example a single component in one modality might be explained as a mixture of components in another. When potential matches are found, it can be difficult to determine if they are simply noisy estimates of the same subject-course or whether the underlying subject-courses are different but correlated.

    A slightly more sophisticated approach to this is the Parallel ICA method described by Liu et al. (2009) which runs separate ICAs on each modality simultaneously; when correlated components are detected, it adds terms to the cost function to encourage these components to become more correlated in later iterations. This relies on a number of tunable constraints (learning rates and weights) to ensure convergence and balance between modalities. Furthermore, it is still not clear how to interpret paired components where the subject-courses are significantly, but not perfectly, correlated.

  • Spatial concatenation has also been used for analysing multimodal data, combining all of the data from each subject into a single dataset with more voxels. This “joint ICA” method has been used before for simultaneously analysing functional maps and grey matter maps (Calhoun et al., 2006), and has been used to extract correlations in structural grey matter/white matter density data (Xu et al., 2009). Since concatenation is a preprocessing step, the ICA model is completely unaware of which voxels belong to which modality.

    However, different modalities may have different spatial source histograms. ICA effectively assumes that each component has a single, non-Gaussian histogram as the prior distribution for all voxels in its spatial map.

    If this map consists of voxels from several different modalities, the modelled histogram (which is effectively an estimate of the source distribution) may have to compromise. For example, this can occur if one modality has a small area of strong activation (or signal change in the case of structural modalities), while the other has a large region of weak activation. This can cause sub-optimal estimates of intensities in spatial maps.

    A related problem is that the contribution each modality makes to the ICA cost function greatly depends on the scaling. One of the difficulties of concatenating multimodal data is that the modalities may have different noise levels and different numbers of voxels. If the scaling is mismatched, unsupervised methods such as PCA and ICA will be dominated by the largest-variance modalities, or those with the most voxels. Typically these concatenation methods also require the same resolution and smoothing for all modalities, rather than using optimized values for each.

    There is also an issue of noise covariance, for example due to spatial smoothing; in particular, adding more smoothing to one modality reduces the noise level but leaves the number of voxels unchanged. The proposed method deals with this explicitly using a precalculated correction for the number of effective degrees of freedom (eDOF), which is closely related to the number resolution elements (RESELs) in the image (Worsley et al., 1995).

    We also expect that some of the structured signals modelled by ICA will be observable in only one modality, and may be extremely weak or even absent in some of the other modalities. It would therefore be useful for sources to be “switched off” in the models where they are not needed, just as it is important to eliminate unneeded components in the single-modality Bayesian ICA model (Choudrey and Roberts, 2001).

  • Tensor ICA stacks the modalities to create a 3D data matrix. This has been used for multi-subject FMRI analysis, with dimensions of voxels × time × subjects (Beckmann and Smith, 2005). In the multimodal scenario this would most likely translate into voxels × subjects × modalities. This is related to the PARAFAC model (see (Nielsen, 2004) for a VB-based implementation) but with the addition of spatial-independence priors. This method assumes that each component has a single spatial map for all modalities, applied to each modality with different weightings. This can be a beneficial feature because it avoids unnecessary duplication of the spatial maps and can allow them to be inferred more accurately when the assumption holds. However this is effectively a strong prior on the nature of the spatial distribution and it may be inappropriate, for example if the number of voxels is different or if the spatial maps in different modalities are not similar.

Using a modular Bayesian framework, we have developed a novel “Linked ICA” general model that allows for either tensor ICA or spatially-concatenated ICA, or a combination of both at the same time. The same subject loading matrix is shared between all of the modalities, so each component consists of a single subject-course and one spatial map in each of the modalities. The subject-weighting matrix automatically balances information from all of the modalities. This novel Linked ICA method will be applied to a data set with four different modalities, acquired from 93 subjects (probable-Alzheimer's patients and age-matched controls). One of these modalities is a grey matter partial volume map (“GM”) derived from Voxel-Based Morphometry (VBM) methods (Ashburner and Friston, 2000), and the other three are measures of white matter integrity: Fractional Anisotropy (FA), Mean Diffusivity (MD), and an orthogonal Tensor Mode (MO) described in Ennis and Kindlmann (2006). These last three modalities have been projected onto a two-dimensional white matter surface (the “skeleton”) using a Tract-Based Spatial Statistics (TBSS) analysis (Smith et al., 2006).

Section snippets

Linked ICA model for multimodal data sets

We assume that the data set is from a group of R subjects, each scanned using several different modalities. It should be noted that the proposed method has the potential to be applied in any situation where multiple modalities have been collected across a single shared dimension (subjects, trials, timepoints, etc.). Each of the scans is prepared for analysis using whatever methods are recommended for a linear regression analysis (or a single-modality ICA) of the group data. This produces maps

Simulated multimodal data

This section presents a simulated multimodal data set, which will be analysed using linked tensor ICA and spatially-concatenated ICA to demonstrate the differences between the two approaches in terms of modelling common (multimodal) components and single-modality structured noise components.

A simulated multimodal data set was constructed with four modalities in two modality groups. The first group contains three modalities of 1000 voxels each that share the same spatial patterns with different

Discussion

The Linked ICA method presented in this paper provides a general, flexible way to perform ICA on multimodal data sets that allows components to be sparse in modalities and allows different noise levels and histograms for each modality group. This method also permits part-tensor configurations when the same spatial maps are expected across some of the modalities and allows model comparison.

Linked tensor ICA performed well in simulated data, and combined information from across modalities more

Acknowledgments

The authors would like to thank Achim Gass and Andreas Monsch for providing the structural and diffusion data, Gwenaëlle Douaud for assistance in interpreting the real data results and Salima Makni for many helpful discussions on Bayesian ICA.

References (31)

  • H. Attias

    A variational Bayesian framework for graphical models

    Adv. Neural. Inf. Process. Syst.

    (2000)
  • C.F. Beckmann

    Independent component analysis for functional magnetic resonance imaging

  • C.F. Beckmann et al.

    Probabilistic independent components analysis for functional magnetic resonance imaging

    IEEE Trans. Med. Imaging

    (2004)
  • C.M. Bishop

    Variational principal components

  • V.D. Calhoun et al.

    Method for multimodal analysis of independent source differences in schizophrenia: combining gray matter structural and auditory oddball functional data

    Hum. Brain Mapp.

    (2006)
  • Cited by (215)

    • Reviewing the neurobiology of electroconvulsive therapy on a micro- meso- and macro-level

      2023, Progress in Neuro-Psychopharmacology and Biological Psychiatry
    • Autism Is Associated With Interindividual Variations of Gray and White Matter Morphology

      2023, Biological Psychiatry: Cognitive Neuroscience and Neuroimaging
    View all citing articles on Scopus
    View full text