PT - JOURNAL ARTICLE AU - Ramchandran, Maya AU - Patil, Prasad AU - Parmigiani, Giovanni TI - Tree-weighting for multi-study ensemble learners AID - 10.1101/698779 DP - 2019 Jan 01 TA - bioRxiv PG - 698779 4099 - http://biorxiv.org/content/early/2019/07/11/698779.short 4100 - http://biorxiv.org/content/early/2019/07/11/698779.full AB - Multi-study learning uses multiple training studies, separately trains classifiers on individual studies, and then forms ensembles with weights rewarding members with better cross-study prediction ability. This article considers novel weighting approaches for constructing tree-based ensemble learners in this setting. Using Random Forests as a single-study learner, we perform a comparison of either weighting each forest to form the ensemble, or extracting the individual trees trained by each Random Forest and weighting them directly. We consider weighting approaches that reward cross-study replicability within the training set. We find that incorporating multiple layers of ensembling in the training process increases the robustness of the resulting predictor. Furthermore, we explore the mechanisms by which the ensembling weights correspond to the internal structure of trees to shed light on the important features in determining the relationship between the Random Forests algorithm and the true outcome model. Finally, we apply our approach to genomic datasets and show that our method improves upon the basic multi-study learning paradigm.