select_feature_set_gmm — Selects an optimal combination from a set of features to classify the provided data.
select_feature_set_gmm selects an optimal subset from a set of features to solve a given classification problem. The classification problem has to be specified with annotated training data in ClassTrainDataHandle and will be classified by a Gaussian Mixture Model. Details of the properties of this classifier can be found in create_class_gmm.
The result of the operator is a trained classifier that is returned in GMMHandle. Additionally, the list of indices or names of the selected features is returned in SelectedFeatureIndices. To use this classifier, calculate for new input data all features mentioned in SelectedFeatureIndices and pass them to the classifier.
A possible application of this operator can be a comparison of different parameter sets for certain feature extraction techniques. Another application is to search for a feature that is discriminating between different classes.
To define the features that should be selected from ClassTrainDataHandle, the dimensions of the feature vectors in ClassTrainDataHandle can be grouped into subfeatures by calling set_feature_lengths_class_train_data. A subfeature can contain several subsequent elements of a feature vector. select_feature_set_gmm decides for each of these subfeatures, if it is better to use it for the classification or leave it out.
The indices of the selected subfeatures are returned in SelectedFeatureIndices. If names were set in set_feature_lengths_class_train_data, these names are returned instead of the indices. If set_feature_lengths_class_train_data was not called for ClassTrainDataHandle before, each element of the feature vector is considered as a subfeature.
The selection method SelectionMethod is either a greedy search 'greedy' (iteratively add the feature with highest gain) or the dynamically oscillating search 'greedy_oscillating' (add the feature with highest gain and test then if any of the already added features can be left out without great loss). The method 'greedy' is generally preferable, since it is faster. Only in cases when the subfeatures are low-dimensional or redundant, the method 'greedy_oscillating' should be chosen.
The optimization criterion is the classification rate of a two-fold cross-validation of the training data. The best achieved value is returned in Score.
The following generic parameters can be set in GenParamNames and GenParamValues:
Minimal number of clusters to represent a class in the training data.
Possible values: '1', '2'
Default value: '1'
Maximal number of clusters to represent a class in the training data.
Possible values: '1', '5', '10'
Default value: '1'
Type of the covariance to represent the size of a cluster.
Possible values: 'spherical', 'diag', 'full'
Default value: 'spherical'
Default value: '42'
Default value: ' 0.001'
Default value: ' 0.0001'
Randomize the input vector.
Default value: '0'
Mode to determine the a-priori probabilities of the classes.
Possible values: 'training', 'uniform'
Default value: 'training'
A more exact description of those parameters can be found in create_class_gmm and train_class_gmm.
This operator may take considerable time, depending on the size of the data set in the training file, and the number of features.
Please note, that this operator should not be called, if only a small set of training data is available. Due to the risk of overfitting the operator select_feature_set_gmm may deliver a classifier with a very high score. However, the classifier may perfom poorly when tested.
Handle of the training data.
Method to perform the selection.
Default value: 'greedy'
List of values: 'greedy', 'greedy_oscillating'
Names of generic parameters to configure the classifier.
Default value: 
List of values: 'class_priors', 'covar_type', 'max_center', 'min_centers', 'random_seed', 'randomize', 'regularize', 'threshold'
Values of generic parameters to configure the classifier.
Default value: 
Suggested values: 1, 2, 3, 'spherical', 'diag', 'full', 42, 0.001, 0.0001, 0
A trained GMM classifier using only the selected features.
The selected feature set, contains indices or names.
The achieved score using two-fold cross-validation.
* Find out which of the two features distinguishes two Classes NameFeature1 := 'Good Feature' NameFeature2 := 'Bad Feature' LengthFeature1 := 3 LengthFeature2 := 2 * Create training data create_class_train_data (LengthFeature1+LengthFeature2,\ ClassTrainDataHandle) * Define the features which are in the training data set_feature_lengths_class_train_data (ClassTrainDataHandle, [LengthFeature1,\ LengthFeature2], [NameFeature1, NameFeature2]) * Add training data * |Feat1| |Feat2| add_sample_class_train_data (ClassTrainDataHandle, 'row', [1,1,1, 2,1 ], 0) add_sample_class_train_data (ClassTrainDataHandle, 'row', [2,2,2, 2,1 ], 1) add_sample_class_train_data (ClassTrainDataHandle, 'row', [1,1,1, 3,4 ], 0) add_sample_class_train_data (ClassTrainDataHandle, 'row', [2,2,2, 3,4 ], 1) add_sample_class_train_data (ClassTrainDataHandle, 'row', [0,0,1, 5,6 ], 0) add_sample_class_train_data (ClassTrainDataHandle, 'row', [2,3,2, 5,6 ], 1) add_sample_class_train_data (ClassTrainDataHandle, 'row', [0,0,1, 5,6 ], 0) add_sample_class_train_data (ClassTrainDataHandle, 'row', [2,3,2, 5,6 ], 1) add_sample_class_train_data (ClassTrainDataHandle, 'row', [0,0,1, 5,6 ], 0) add_sample_class_train_data (ClassTrainDataHandle, 'row', [2,3,2, 5,6 ], 1) * Add more data * ... * Select the better feature with a GMM select_feature_set_gmm (ClassTrainDataHandle, 'greedy', , , GMMHandle,\ SelectedFeatureGMM, Score) clear_class_train_data (ClassTrainDataHandle) * Use the classifier * ... clear_class_gmm (GMMHandle)
If the parameters are valid, the operator select_feature_set_gmm returns the value 2 (H_MSG_TRUE). If necessary, an exception is raised.
create_class_train_data, add_sample_class_train_data, set_feature_lengths_class_train_data
select_feature_set_mlp, select_feature_set_knn, select_feature_set_svm
create_class_gmm, gray_features, region_features