Professional Documents
Culture Documents
Nima Hatami Dept. of Elec. & Elec. Eng. Univ. of Cagliari, Italy
Agenda
Classifier ensembles Mixture of Experts (ME) model Hierarchical ME Mixture of Random Prototype-based Experts (MRPE) Hierarchical MRPE Experimental results Conclusions & Future work
PhD Day 2011
N. Hatami, Mixture of Random Prototype-based Local Experts
Classifier ensembles
Also Known as classifier fusion, combining classifiers, MCS Most real-world PR problems are too complicated for a single classifier to solve Divide-and-conquer has proved to be efficient in many of these complex situations Combination of classifiers which have complementary properties
Mixture of Experts
Jacobs et al. have proposed the ME based on the divide-and-conquer strategy One of the most popular ensemble methods used in PR and ML A set of expert networks is trained together with a gate network
Mixture of Experts
Stochastically partitions the input space of the problem into a number of subspaces Which experts becoming specialized on each subspace Uses the gating network to manage this process
Mixture of Experts
ith expert Gating network Final output
Hierarchical ME model
The HME is a well-known tree-structured architecture which can be considered a natural extension of the ME. The standard HME model hierarchically splits the input space into a nested set of subspaces
G. Armano, N. Hatami, "Random Prototype-based Oracle for selection fusion ensembles" ICPR 2010. G. Armano and N. Hatami, "Mixture of Random Prototype-based local Experts". HAIS 2010.
10
G. Armano and N. Hatami, "Mixture of Random Prototype-based local Experts". HAIS 2010.
11
G. Armano and N. Hatami, "Mixture of Random Prototype-based local Experts". HAIS 2010.
12
13
Hierarchical MRPE
Data splitting based on random prototypes has been used for each ME module
G. Armano, N. Hatami, "Hierarchical Mixture of Random Prototype-based Experts", ECML PKDD 2010.
15
16
G. Armano and N. Hatami, "Run-time Performance Analysis of the Mixture of Experts Model". CORES 2011.
18
G. Armano and N. Hatami, "Run-time Performance Analysis of the Mixture of Experts Model". CORES 2011.
19
TgME/TgMRPE=4.65!!!
Decreasing complexity too!!!
G. Armano and N. Hatami, "Run-time Performance Analysis of the Mixture of Experts Model". CORES 2011.
20
Experimental results
We used some of the UCI ML datasets 10-fold cross-validation Multi-layer perceptron (MLP) For N, number of partitions (experts), we varied it from 2 to 10
G. Armano and N. Hatami, "Mixture of Random Prototype-based local Experts". HAIS 2010.
21
Experimental results
G. Armano and N. Hatami, "Mixture of Random Prototype-based local Experts". HAIS 2010.
22
Run-time comparison
Training time Error rate Complexity (run-time)
G. Armano and N. Hatami, "Run-time Performance Analysis of the Mixture of Experts Model". CORES 2011.
23
Conclusions
A modified version of the popular ME algorithm is presented Specializes expert networks on centralized regions of input space instead of nested and stochastic regions Using simple distance-based gating thus reduces the network complexity and the training time Improves overall classification accuracy and Complexity
PhD Day 2011
24
Future work
Defining a procedure for automatically determining the number of optimal experts for each problem without resorting to complex preprocessing Adaptation of this method to simple distance-based classifiers instead of NNs Heuristics able to help in the process of partitioning the input space instead of using RP Using the error rate and the complexity for automatically estimating the optimal Ne for a given problem
25