papers AI Learner
The Github is limit! Click to go to the new site.

EnsembleNet: End-to-End Optimization of Multi-headed Models

2019-05-24
Hanhan Li, Joe Yue-Hei Ng, Paul Natsev

Abstract

Ensembling is a universally useful approach to boost the performance of machine learning models. However, individual models in an ensemble are typically trained independently in separate stages, without information access about the overall ensemble. In this paper, model ensembles are treated as first-class citizens, and their performance is optimized end-to-end with parameter sharing and a novel loss structure that improves generalization. On large-scale datasets including ImageNet, Youtube-8M, and Kinetics, we demonstrate a procedure that starts from a strongly performing single deep neural network, and constructs an EnsembleNet that has both a smaller size and better performance. Moreover, an EnsembleNet can be trained in one stage just like a single model without manual intervention.

Abstract (translated by Google)
URL

http://arxiv.org/abs/1905.09979

PDF

http://arxiv.org/pdf/1905.09979


Similar Posts

Comments