Mediated experts for deep convolutional networks
Journal
IEEE International Conference on Acoustics, Speech and Signal Processing
Journal Volume
2016-May
Pages
2687-2691
ISBN
9781479999880
Date Issued
2016
Author(s)
Agethen S.
Abstract
We present a new supervised architecture termed Mediated Mixture-of-Experts (MMoE) that allows us to improve classification accuracy of Deep Convolutional Networks (DCN). Our architecture achieves this with the help of expert networks: A network is trained on a disjoint subset of a given dataset and then run in parallel to other experts during deployment. A mediator is employed if experts contradict each other. This allows our framework to naturally support incremental learning, as adding new classes requires (re-)training of the new expert only. We also propose two measures to control computational complexity: An early-stopping mechanism halts experts that have low confidence in their prediction. The system allows to trade-off accuracy and complexity without further retraining. We also suggest to share low-level convo-lutional layers between experts in an effort to avoid computation of a near-duplicate feature set. We evaluate our system on a popular dataset and report improved accuracy compared to a single model of same configuration. ? 2016 IEEE.
Subjects
Deep Learning; Incremental Learning; Mixture of Experts
Type
conference paper
File(s)![Thumbnail Image]()
Loading...
Name
1511.06072.pdf
Size
400.04 KB
Format
Adobe PDF
Checksum
(MD5):b326a843c9637f66ea60c9eb881ec281