HASE: Framework for efficient high-dimensional association analyses =================================================================== * G.V. Roshchupkin * H.H.H. Adams * M.W. Vernooij * A. Hofman * C.M. Van Duijn * M.A. Ikram * W.J. Niessen ## ABSTRACT Large-scale data collection and processing have facilitated scientific discoveries in fields such as genomics and imaging, but cross-investigations between multiple big datasets remain impractical. Computational requirements of high-dimensional association studies are often too demanding for individual sites. Additionally, the sheer size of intermediate results is unfit for collaborative settings where summary statistics are exchanged for meta-analyses. Here we introduce the HASE framework to perform high-dimensional association studies with dramatic reduction in both computational burden and storage requirements of intermediate results. We implemented a novel meta-analytical method that yields identical power as pooled analyses without the need of sharing individual participant data. The efficiency of the framework is illustrated by associating 9 million genetic variants with 1.5 million brain imaging voxels in three cohorts (total N=4,034) followed by meta-analysis, on a standard computational infrastructure. These experiments indicate that HASE facilitates high-dimensional association studies enabling large multicenter association studies for future discoveries. ## INTRODUCTION Technological innovations have enabled the large-scale acquisition of biological information from human subjects. The emergence of these big datasets has resulted in various ’omics’ fields. Systematic and large-scale investigations of DNA sequence variations (genomics)1, gene expression (transcriptomics)2, proteins (proteomics)3, small molecule metabolites (metabolomics)4, and medical images (radiomics)5, among other data, lie at the basis of many recent biological insights. These analyses are typically unidimensional, i.e. studying only a single disease or trait of interest. Although this approach has proven its scientific merit through many discoveries, jointly investigating multiple big datasets would allow for their full exploitation, as is increasingly recognized throughout the ’omics’ world. However, the high-dimensional nature of these analyses makes them challenging and often unfeasible in current research settings. Specifically, the computational requirements for analyzing high-dimensional data are far beyond the infrastructural capabilities for single sites. Furthermore, it is incompatible with the typical collaborative approach of distributed multi-site analyses followed by meta-analysis, since the amount of generated data at every site is too large to transfer. Some studies have attempted to combine multiple big datasets5,8–10, but these methods generally rely on reducing the dimensionality or making assumptions to approximate the results, which leads to a loss of information. Here we present the High-dimensional Association Studies with Efficiency (HASE) framework, which is capable of analyzing high-dimensional data at full resolution, yielding exact association statistics (i.e. no approximations), and requiring only standard computational facilities. Additionally, the major computational burden in collaborative efforts is shifted from the individual sites to the meta-analytical level while at the same time reducing the amount of data needed to be exchanged and preserving participant privacy. HASE thus removes the current computational and logistic barriers for single-and multi-center analyses of big data. ## RESULTS ### Overview of the methods The methods are described in detail in the Online Methods. Essentially, HASE implements a high-throughput multiple linear regression algorithm that is computationally efficient when analyzing high-dimensional data of any quantitative trait. Prior to analysis, data are converted to an optimized storage format to reduce reading and writing time. Redundant calculations are removed and the high-dimensional operations are simplified into a set of matrix operations that are computationally inexpensive, thereby reducing overall computational overhead. While deriving summary statistics (e.g., beta coefficients, p-values) for every combination in the high-dimensional analysis would be computationally feasible at individual sites with our approach, it would be too large to share the intermediate results (>200GB per thousand phenotypes) in a multi-center setting. Therefore, extending from a recently proposed method, partial derivatives17 meta-analysis, we additionally developed a method that generates two relatively small datasets (e.g. 5GB for genetics data of 9 million variants and 20MB of thousand phenotypes for 4000 individuals) that are easily transferred and can subsequently be combined to calculate the full set of summary statistics, without making any approximation. This meta-analysis method additionally reduces computational overhead at individual sites by shifting the most expensive calculation to the central site. The total computational burden thus becomes even more efficient relative to conventional methods with additional sites. The HASE software is freely available from our website [www.imagene.nl/HASE/](http://www.imagene.nl/HASE/). ### Comparison of complexity and speed We compared the complexity and speed of HASE with a classical workflow, based on linear regression analyses with PLINK (version 1.9)**11 followed by meta-analysis with METAL12**; two of the most popular software packages for these tasks. **Table 1** shows that HASE dramatically reduces the complexity for the single site analysis and data transfer stages. For conventional methods, the single site analysis and data transfer have a multiplicative complexity (dependent on the number of phenotypes and determinants), whereas this is only additive for HASE. Our approach requires 2×106-fold less time on the single site stage and 3.500-fold less data to transfer for a high-dimensional association study. Additionally, the time for single site analysis does not increase significantly from analyzing a single phenotype to a million phenotypes **(Table 1).** This is due to the fact that speed is determined by the highest number of either the determinants or phenotypes. Therefore, in this case with nine million genetic variants, the complexity of ***O*(*ninp*)** is the primary factor influencing the speed, whereas ***O* (*nint***) plays a secondary role. View this table: [Table 1.](http://biorxiv.org/content/early/2016/01/21/037382/T1) Table 1. Comparison of complexity and speed between the HASE framework and a classical workflow. This drastic increase in performance is made possible through the shift of the computationally most expensive regression operation to the meta-analytical stage. For the meta-analytical stage, the HASE complexity is therefore slightly higher. However, it outperforms the classical meta-analysis using METAL (total computation time reduced 35 times), owing to the efficient implementation of our algorithm. Additionally, if HASE is only used for a high-dimensional association study of a single site, i.e. without subsequent meta-analysis with other sites, the computation time would be reduced 1400 times due to the removal of redundant calculations (for details see **Online Methods).** ### Application to real data We used HASE to perform a high-dimensional association study in 4,034 individuals from the population-based Rotterdam Study. In this proof of principle study, we relate 8,723,231 million imputed genetic variants to 1,534,602 million brain magnetic resonance imaging (MRI) voxel densities (see **Online Methods).** The analysis was performed on a small cluster of 100 CPUs and took 17 hours to complete. To demonstrate the potential of such high-dimensional analyses, we screened all genetic association results for both hippocampi (7,030 voxels) and identified the voxel with the lowest p-value. The most significant association (rs77956314; p = 3 x 10−9) corresponded to a locus on chromosome 12q24 **(Figure 1)**, which was recently discovered in a genome-wide association study of hippocampal volume encompassing 30,717 participants**13**. ![Figure 1.](http://biorxiv.org/https://www.biorxiv.org/content/biorxiv/early/2016/01/21/037382/F1.medium.gif) [Figure 1.](http://biorxiv.org/content/early/2016/01/21/037382/F1) Figure 1. Manhattan plot of the hippocampus voxel with the most significant association after screening all 7030 hippocampal voxels. The most significant association (rs77956314; p = 3 ×10−9) corresponded to a previously identified locus on chromosome 12q24. Such voxel-wise hippocampus screening would take less than 8 hours on standard laptop. Additionally, we performed the high-dimensional association studies separately in the three subcohorts of the Rotterdam Study and meta-analyzed the results using the HASE data reduction approach. It took on average 40 minutes for each subcohort to generate intermediate data for subsequent meta-analysis on a single CPU for all genetic variants and voxels. The meta-analysis was performed on the same cluster and took 17 hours to complete. Next, we compared the association results of the pooled analysis with the meta-analysis.**Figure 2** shows that the results are identical as it was predicted by theory (see **Online Methods).** ![Figure 2.](http://biorxiv.org/https://www.biorxiv.org/content/biorxiv/early/2016/01/21/037382/F2.medium.gif) [Figure 2.](http://biorxiv.org/content/early/2016/01/21/037382/F2) Figure 2. Correlation plot of voxel GWAS t-statistic estimated from pooled together data and voxel GWAS t-statistic estimated from meta-analysis of partial derivatives and encoded matrix. It took 40 min for single site to pre-compute data instead of 280 years to compute summary statistics; average data size to transfer was 17.5GBs instead of 300TBs. ## DISCUSSION We describe a framework that allows for (i) computationally-efficient high-dimensional association studies within individual sites using standard computational infrastructure and (ii) facilitates the exchange of compact summary statistics for subsequent meta-analysis for association studies in a collaborative setting. Using HASE, we performed a genome-wide and brain-wide search for genetic influences on voxel densities, and illustrate both its feasibility and potential for driving scientific discoveries. When using HASE, it is first necessary to convert the multi-dimensional data to a format that is optimized for fast reading and writing. This particular format, «hdf5», is not dependent on the architecture of the file system and can therefore be implemented on a wide range of hardware and software infrastructures. To facilitate this initial conversion step, we have built-in tools within the HASE framework for processing common file format of such big data. Furthermore, this is easily generalizable to other large data matrices in general and we foresee this initial conversion step not to form an obstacle for researchers to implement HASE. In addition to the data format, a large improvement in efficiency comes from the reduced computational complexity. High-dimensional analyses contain many redundant calculations, which were removed in the HASE software. Also, we were able to further increase efficiency by simplifying the calculations to a set of matrix operations, which are computationally inexpensive, compared to conventional linear regression algorithms. Furthermore, the implementation of partial derivatives meta-analysis allowed us to greatly reduce the size of the summary statistics that need to be shared for performing a meta-analysis. Another advantage of this approach is that it only needs to calculate the partial derivatives for each site instead of the parameter estimates (i.e., beta coefficients and standard errors). This enabled us to develop within HASE a reduction approach that encodes data prior to exchange between sites, while yielding the exact same results after meta-analysis as if the original data were used. The encoding is performed such that tracing back to original data is impossible. This guarantees protection of participant privacy and circumvents restrictions on data sharing that are unfortunately common in many research institutions. Alternative methods for solving the issues with high-dimensional data take one of two approaches. One approach is to reduce the dimensionality of the big datasets by summarizing the large amount of data into fewer variables2. Although this increases the speed, it comes at the price of losing valuable information, which these big data were primarily intended to capture. The second approach is to not perform a full analysis of all combinations of the big datasets, but instead make certain assumptions (e.g., a certain underlying pattern, or a lack of dependency on potential confounders) that allow for using statistical models that require less computing time. Again, this is a tradeoff between speed and accuracy, which is not necessary in the HASE framework, where computational efficiency is increased without introducing any approximations. Unidimensional analyses of big data, such as genome-wide association studies, have already elucidated to some extent the genetic architecture of complex diseases and other traits of interest1,14-16, but much remains unknown. Cross-investigations between multiple big datasets potentially hold the key to fulfill the promise of big data in understanding of biology7. Using the HASE framework to perform high-dimensional association studies, this hypothesis is now testable. **URLs.** Framework for efficient high-dimensional association analyses (HASE), [https://github.com/roshchupkin/HASE/;](https://github.com/roshchupkin/HASE/;) description of the framework and protocol for metaanalysis, [www.imagene.nl/HASE](http://www.imagene.nl/HASE); ## Author Contributions GVR and HHHA jointly conceived of the study, participated in its design, performed the statistical analysis, interpreted the data, and drafted the manuscript. GVR additionally developed the software. MWV, AH, and CMvD acquired data and revised the manuscript critically for important intellectual content. MAI and WJN participated in its design, interpreted the data, and revised the manuscript critically for important intellectual content. All authors read and approved the final manuscript. ## Competing Financial Interests WJN is co-founder and shareholder of Quantib BV. None of the other authors declare any competing financial interests. ## Online Methods ### Hase In high-dimensional associations analyses we test the following simple regression model: ![Formula][1] where **Y** is a **ni × np** matrix of phenotypes of interest, **ni** denotes the number of samples in the study, **np** the number of phenotypes of interest, and ɛ denotes the residual effect. **X** is a three dimensional matrix **ni × nc × nt** of independent variables, with **nc** representing the number of covariates, such as the intercept, age, sex and, for example genotype as number of alleles, and **nt** the number of independent determinants. In association analyses we are interested in estimating the p-value to test the null hypothesis that ß=0. The p-values can be directly derived from the t-statistic of our test determinants. We will rewrite the classical equation for calculating t-statistics for our multi-dimensional matrices, which will lead to a simple matrix form solution for high-dimensional association analysis: ![Formula][2] Where **T** is **np × nc × nt** matrix of t-statistics and **df** is degree of freedom of our regression model. Let’s define A = XTX,B = XTY and C = YTY, so that we can write our final equation for t-statistics: ![Formula][3] The result of this derivation is that, rather than computing all combinations of covariates and independent determinants, we only need to know three matrices: A, B and C, to calculate t-statistics and perform the full analysis. These results will be used in the section about meta-analysis. The most computationally expensive operations here are the two multi-dimensional matrix multiplications (***A***-1**B**) and (***BTA-B***), where ***A*-1** is a three dimensional matrix **nc × nc × nt** and **B** is three dimensional matrix **nc × np × nt**. Without knowledge of the data structure of these matrices, the simplest way to write the results of their multiplication would be to use Einstein’s notation for tensor multiplication: ![Formula][4] As you can see, the result is two matrices of **nc × np × nt** and **np × nt** elements respectively. Despite the seemingly complex notation, the first matrix just represents the beta coefficients for all combinations of covariates (**nc** by **np × nt** combinations) and the second is fitting values of the dependent variable for every test (**np × nt** independent determinants). However, insight into the data structure of **A** and **B** can dramatically reduce the computational burden and simplify operations. First of all, matrix **A** depends only on the covariates and number of determinants, making it unnecessary to compute it for every phenotype of interest, so we just need to calculate it once. Additionally, only the last covariate (i.e., the variable of interest) is different between tests, meaning that the (**np -1)×(np -1)xnt** part of matrix **A** remains constant during high-dimensional analyses. Matrix **B** consists of the dot product of every combination of the covariate and phenotype of interest. However, as we mentioned before, there are only (**nt + nc 1)** different covariates, and thus we can split matrix **B** in two low dimensional matrices: the first includes dot products of non-tested covariates - (**nc-1) × np** elements; the second includes the dot products only of the tested covariates - **np × nt** elements. All this allows us to achieve large gain in computation efficiency and memory usage. In **Figure 1** we show a 2D schematic representation of these two matrices for standard genome association study with the covariates being an intercept, age, sex, and genotype. This example could be easily extrapolated to any linear regression model. Applying the same splitting operation to ***BT*** it is possible to simplify tensor multiplication equation (**8**, **9**) to a low-dimensional matrix operation and rewrite the equation for t-statistics: ![Formula][5]![Formula][6]![Formula][7]![Formula][8] Then, to compute t-statistics for high-dimensional association analyses we just need to perform several matrix multiplications. #### Meta-analysis In classical meta-analysis, summary statistics such as beta coefficients and p-values are exchanged between sites. For 1.5 million phenotypes, this would yield around 400TB of data at each site, making data transfer to a centralized site impractical. In the previous section we showed that, to compute all statistics for an association study, we just need to know the **A, B** and **C** matrices. As we demonstrated before6, by exchanging these matrices between sites, it is possible to gain the same statistical power as with a pooled analysis, without sharing individual participant data, because these matrices consist of aggregate data (**Figure 1).** However, in high-dimensional association analyses, matrix **B** grows very fast, particularly the part that depends on the number of determinants and phenotypes (**b4** in **Figure 1**). If **Y** is a **ni × np** matrix of phenotypes of interest and **G** is a **ni × nt** matrix of determinants which we want to test (e.g., a genotype matrix in GWAS), then **b4 = YT** × **G.** These two matrices, **Y** and **G,** separately are not so large, but their product matrix has **np × nt** elements, which in a real application could be 106 × 107 =1013 elements and thus too large to share between sites. We propose to create a random **ni × ni** nonsingular square matrix **F** and calculate its inverse matrix **F-**1. Then by definition **F × F-1=I,** where **I** is a **ni × ni** elements identity matrix with ones on main diagonal and zeros elsewhere. Using this property, we can rewrite the equation for **b4**: ![Formula][9]![Formula][10] Therefore, instead of transferring TBs of intermediate statistics (**b4),** each side just needs to compute **A, C, YF and GF.** Sharing just the encoded matrices does not provide information on individual participants and without knowing matrix **F** it is impossible to reconstruct the real data. However, it will be possible to calculate **b4,** perform a high-dimensional meta-analysis, and avoid problems with data transfer. Additionally, this method dramatically reduces computation time by shifting all complex computations to central site, where the HASE regression algorithm should be used to handle the association analysis in time efficient way. ![Figure 1.](http://biorxiv.org/https://www.biorxiv.org/content/biorxiv/early/2016/01/21/037382/F3.medium.gif) [Figure 1.](http://biorxiv.org/content/early/2016/01/21/037382/F3) Figure 1. Explanation of the achieved speed reduction in HASE framework by removing redundant computations. In HASE multi-dimensional **A** and **B** matrices need to be calculated to perform GWAS studies. In the figure grey color means elements are parts of the matrix that are not necessary to calculate, as the **A** matrix is symmetric. The green color indicates elements that need to be calculated only once. Blue elements only have to be calculated for every SNP and yellow only for every phenotype. The red color indicates the most computationally expensive element, which needs to be calculated for every combination of phenotype and genotype. N denotes the number subjects in study. ### Supplementary Note #### Study Population The Rotterdam Study is an ongoing population-based cohort study in the Netherlands investigating diseases in the elderly and currently consists of 14,926 residents of Rotterdam who were aged 45 years or more at baseline [1,2]. The initial cohort was started in 1990 and expanded in 2000 and 2005. The whole population is subject to a set of multidisciplinary examinations every four years. MRI was implemented in 2005 and 5430 persons scanned until 2011 were eligible for this study. We excluded individuals with incomplete acquisitions, scans with artifacts hampering automated processing, participants with MRI-defined cortical infarcts, and subjects with dementia or stroke at the time of scanning. This resulted in a final study population of 4071 non-demented persons with information available on both genome-wide genotyping and MRI data. The Medical Ethics Committee of the Erasmus MC, University Medical Center Rotterdam and the review board of the Netherlands Ministry of Health, Welfare and Sports both approved the study. Informed consent was obtained from all subjects. #### Imputation of genotypes The Illumina 550K and 550K duo arrays were used for genotyping. Samples with low call rate (<97.5%), with excess autosomal heterozygosity (>0.336) or with sex-mismatch were excluded, as were outliers identified by the identity-by-state clustering analysis (outliers were defined as being >3 standard deviation (SD) from population mean or having identity-by-state probabilities >97%). A set of genotyped input SNPs with call rate >98%, MAF >0.001 and Hardy-Weinberg equilibrium (HWE) P-value > 10−6 was used for imputation. The Markov Chain Haplotyping (MACH) package version 1.0 software (Imputed to plus strand of NCBI build 37, 1000 Genomes phase I version 3) and minimac version 2012.8.6 were used for imputation. #### MRI data From August 2005 onwards, a dedicated 1.5 Tesla MRI scanner (GE Healthcare, Milwaukee, Wisconsin, USA) is operational in the Rotterdam Study research center in Ommoord. This scanner is operated by trained research technicians and all imaging data are collected according to standardized imaging protocols[2]. Brain MRI scans included a high-resolution 3D T1-weighted fast RF spoiled gradient recalled acquisition in steady state with an inversion recovery pre-pulse (FASTSPGR-IR) sequence with thin slices (voxel size<1mm3)[2]*Image processing* Voxel based morphometry (VBM) was performed according to an optimized VBM protocol [3]. First, all T1-weighted images were segmented into supratentorial gray matter (GM), white matter (WM) and cerebrospinal fluid (CSF) using a previously described k-nearest neighbor (kNN) algorithm, which was trained on six manually labeled atlases [4]. FSL software [5] was used for VBM data processing. First, all GM density maps were non-linearly registered to the standard GM probability template. For this study we chose the ICBM MNI152 GM template (Montreal Neurological Institute) with a 1x1x1 mm3 voxel resolution. The MNI152 standard-space T1-weighted average structural template is derived from 152 structural images, which have been warped and averaged into the common MNI152 co-ordinate system after high-dimensional nonlinear registration. A spatial modulation procedure was used to avoid differences in absolute GM volume due to the registration. This involved multiplying voxel density values by the Jacobian determinants estimated during spatial normalization. To gain more statistic power and decrease signal to noise ratio, all images were smoothed using a 3mm (FWHM 8mm) isotropic Gaussian kernel. #### Statistical analysis Linear regression models were fitted with voxel values of GM modulation density as the dependent variable and age, sex, and the number of minor alleles as independent variables. In total 1,534,602 voxels were processed. ## Acknowledgments The generation and management of GWAS genotype data for the Rotterdam Study are supported by the Netherlands Organization of Scientific Research NWO Investments (nr. 175.010.2005.011, 911-03-012). This study is funded by the Research Institute for Diseases in the Elderly (014-93-015; RIDE2), the Netherlands Genomics Initiative (NGI)/Netherlands Organization for Scientific Research (NWO) project nr. 050-060-810. The Rotterdam Study is funded by Erasmus Medical Center and Erasmus University, Rotterdam, Netherlands Organization for the Health Research and Development (ZonMw), the Research Institute for Diseases in the Elderly (RIDE), the Ministry of Education, Culture and Science, the Ministry for Health, Welfare and Sports, the European Commission (DG XII), and the Municipality of Rotterdam. MAI is supported by ZonMW grant number 916.13.054. HHHA is supported by the Van Leersum Grant of the Royal Netherlands Academy of Arts and Sciences. The Research is supported by the ImaGene programme of STW, The Society for Technical Scientific Research in The Netherlands. ## Footnotes * * starred authors jointly supervised this work * Received January 21, 2016. * Accepted January 21, 2016. * © 2016, Posted by Cold Spring Harbor Laboratory This pre-print is available under a Creative Commons License (Attribution-NonCommercial 4.0 International), CC BY-NC 4.0, as described at [http://creativecommons.org/licenses/by-nc/4.0/](http://creativecommons.org/licenses/by-nc/4.0/) ## References 1. 1.Wood, A. R. et al. Defining the role of common variation in the genomic and biological architecture of adult human height. Nat. Genet. 46, 1173–1186 (2014). [CrossRef](http://biorxiv.org/lookup/external-ref?access_num=10.1038/ng.3097&link_type=DOI) [PubMed](http://biorxiv.org/lookup/external-ref?access_num=25282103&link_type=MED&atom=%2Fbiorxiv%2Fearly%2F2016%2F01%2F21%2F037382.atom) 2. 2.Sambasivarao, S. V. Spatial patterns of genome wide expression profiles reflect anatomic and fiber connectivity architecture of healthy human brain. 18, 1199–1216 (2013). 3. 3.Stunnenberg, H. G. & Hubner, N. C. Genomics meets proteomics: Identifying the culprits in disease. Hum. Genet. 133, 689–700 (2014). 4. 4.Krumsiek, J. et al. Mining the Unknown: A Systems Approach to Metabolite Identification Combining Genetic and Metabolic Information. PLoS Genet. 8, e1003005 (2012). [CrossRef](http://biorxiv.org/lookup/external-ref?access_num=10.1371/journal.pgen.1003005&link_type=DOI) [PubMed](http://biorxiv.org/lookup/external-ref?access_num=23093944&link_type=MED&atom=%2Fbiorxiv%2Fearly%2F2016%2F01%2F21%2F037382.atom) 5. 5.Parmar, C. et al. Radiomic feature clusters and Prognostic Signatures specific for Lung and Head & Neck cancer. Sci. Rep. 5, 11044 (2015). [CrossRef](http://biorxiv.org/lookup/external-ref?access_num=10.1038/srep11044&link_type=DOI) [PubMed](http://biorxiv.org/lookup/external-ref?access_num=26251068&link_type=MED&atom=%2Fbiorxiv%2Fearly%2F2016%2F01%2F21%2F037382.atom) 6. 6.Medland, S. E., Jahanshad, N., Neale, B. M. & Thompson, P. M. Whole-genome analyses of whole-brain data: working within an expanded search space. Nat. Publ. Gr. 17, 791–800 (2014). 7. 7.Robinson, M. R., Wray, N. R. & Visscher, P. M. Explaining additional genetic variation in complex traits. Trends Genet. 30, 124–32 (2014). [CrossRef](http://biorxiv.org/lookup/external-ref?access_num=10.1016/j.tig.2014.02.003&link_type=DOI) [PubMed](http://biorxiv.org/lookup/external-ref?access_num=24629526&link_type=MED&atom=%2Fbiorxiv%2Fearly%2F2016%2F01%2F21%2F037382.atom) 8. 8.Zou, F. et al. Brain Expression Genome-Wide Association Study (eGWAS) Identifies Human Disease-Associated Variants. PLoS Genet. 8, e1002707 (2012). [CrossRef](http://biorxiv.org/lookup/external-ref?access_num=10.1371/journal.pgen.1002707&link_type=DOI) [PubMed](http://biorxiv.org/lookup/external-ref?access_num=22685416&link_type=MED&atom=%2Fbiorxiv%2Fearly%2F2016%2F01%2F21%2F037382.atom) 9. 9.Stein, J. L. et al. Voxelwise genome-wide association study (vGWAS). Neuroimage 53, 1160–74 (2010). [CrossRef](http://biorxiv.org/lookup/external-ref?access_num=10.1016/j.neuroimage.2010.02.032&link_type=DOI) [PubMed](http://biorxiv.org/lookup/external-ref?access_num=20171287&link_type=MED&atom=%2Fbiorxiv%2Fearly%2F2016%2F01%2F21%2F037382.atom) [Web of Science](http://biorxiv.org/lookup/external-ref?access_num=000282039300042&link_type=ISI) 10. 10.Huang, M., Nichols, T., Huang, C., Yang, Y. & Lu, Z. FVGWAS: Fast Voxelwise Genome Wide Association Analysis of Large-scale Imaging Genetic Data 1. 11. 11.Purcell, S. et al. PLINK: a tool set for whole-genome association and population-based linkage analyses. Am. J. Hum. Genet. 81, (2007). 12. 12.Willer, C. J., Li, Y. & Abecasis, G. R. METAL: Fast and efficient meta-analysis of genomewide association scans. Bioinformatics 26, 2190–2191 (2010). [CrossRef](http://biorxiv.org/lookup/external-ref?access_num=10.1093/bioinformatics/btq340&link_type=DOI) [PubMed](http://biorxiv.org/lookup/external-ref?access_num=20616382&link_type=MED&atom=%2Fbiorxiv%2Fearly%2F2016%2F01%2F21%2F037382.atom) [Web of Science](http://biorxiv.org/lookup/external-ref?access_num=000281738900017&link_type=ISI) 13. 13.Hibar, D. P. et al. Common genetic variants influence human subcortical brain structures. Nature 8, (2015). 14. 14.Purcell, S. M. et al. Common polygenic variation contributes to risk of schizophrenia and bipolar disorder. Nature 460, 748–52 (2009). [CrossRef](http://biorxiv.org/lookup/external-ref?access_num=10.1038/nature08185&link_type=DOI) [PubMed](http://biorxiv.org/lookup/external-ref?access_num=19571811&link_type=MED&atom=%2Fbiorxiv%2Fearly%2F2016%2F01%2F21%2F037382.atom) [Web of Science](http://biorxiv.org/lookup/external-ref?access_num=000268670300041&link_type=ISI) 15. 15.Polychronakos, C. & Alriyami, M. Diabetes in the post-GWAS era. Nat. Genet. 47, 1373–1374 (2015). 16. 16.Locke, A. E. et al. Genetic studies of body mass index yield new insights for obesity biology. Nature 518, 197–206 (2015). [CrossRef](http://biorxiv.org/lookup/external-ref?access_num=10.1038/nature14177&link_type=DOI) [PubMed](http://biorxiv.org/lookup/external-ref?access_num=25673413&link_type=MED&atom=%2Fbiorxiv%2Fearly%2F2016%2F01%2F21%2F037382.atom) 17. 17.Adams, H.H.H. et al. Partial derivatives meta-analysis: pooled analyses without sharing individual participant data. (2016) submitted ## References 1. 1. Albert Hofman et al. "The Rotterdam Study: 2012 objectives and design update." In: European journal of epidemiology 26.8 (Aug. 2011), pp. 657 [CrossRef](http://biorxiv.org/lookup/external-ref?access_num=10.1007/s10654-011-9610-5&link_type=DOI) [PubMed](http://biorxiv.org/lookup/external-ref?access_num=21877163&link_type=MED&atom=%2Fbiorxiv%2Fearly%2F2016%2F01%2F21%2F037382.atom) 2. 2. M. Arfan Ikram et al. "The Rotterdam Scan Study: design and update up to 2012." In: European journal of epidemiology 26.10 (Oct. 2011), pp. 811 [CrossRef](http://biorxiv.org/lookup/external-ref?access_num=10.1007/s10654-011-9624-z&link_type=DOI) [PubMed](http://biorxiv.org/lookup/external-ref?access_num=22002080&link_type=MED&atom=%2Fbiorxiv%2Fearly%2F2016%2F01%2F21%2F037382.atom) 3. 3. C. D. Good et al. "A voxel-based morphometric study of ageing in 465 normal adult human brains." In: NeuroImage 14.1 Pt 1 (July 2001), pp. 21 [CrossRef](http://biorxiv.org/lookup/external-ref?access_num=10.1006/nimg.2001.0786&link_type=DOI) [PubMed](http://biorxiv.org/lookup/external-ref?access_num=11525331&link_type=MED&atom=%2Fbiorxiv%2Fearly%2F2016%2F01%2F21%2F037382.atom) [Web of Science](http://biorxiv.org/lookup/external-ref?access_num=000169498000003&link_type=ISI) 4. 4.Henri a Vrooman et al. "Multi-spectral brain tissue segmentation using automatically trained k-Nearest-Neighbor classification." In: NeuroImage 37.1 (Aug. 2007) 5. 5.Stephen M. Smith et al. "Advances in functional and structural MR image analysis and implementation as FSL." In: NeuroImage 23 Suppl 1 (Jan. 2004) 6. 6.Adams, H.H.H. et al. Partial derivatives meta-analysis: pooled analyses without sharing individual participant data. (2016) submitted [1]: /embed/graphic-4.gif [2]: /embed/graphic-5.gif [3]: /embed/graphic-6.gif [4]: /embed/graphic-7.gif [5]: /embed/graphic-8.gif [6]: /embed/graphic-9.gif [7]: /embed/graphic-10.gif [8]: /embed/graphic-11.gif [9]: /embed/graphic-12.gif [10]: /embed/graphic-13.gif