Online Detection and Classification of Moving Objects Using Progressively Improving Detectors
Paper: | Online Detection and Classification of Moving Objects Using Progressively Improving Detectors |
Contact: | Omar Javed, Saad Ali, Mubarak Shah |
Introduction
Major limitations of classification algorithms like Adaboost, SVMs, or Naïve Bayes include,
- Requirement of a large amount of labeled training data.
- The parameters of the classifier are fixed after the end of training phase, i.e., these classifiers can not attune themselves to particular detection scenarios after their deployment.
To overcome above mentioned shortcomings, the algorithm proposed in this project has following properties: i) Requirement of minimal training data, ii) automated online selection of training examples after deployment for continuous improvement, and iii) near real time performance (4-5 frames/sec). The algorithm is based on a boosted classification framework, in which, aeparate views (features) of the data are used for online collection of training examples through co-training. While combined view (all features) are used to make the classification decisions. The background modeling is used to prune away stationary regions to speed up the classification process. We used global feature representations for robustness.
Feature Extraction
Features for classification are derived from Principal Component Analysis (PCA) of the appearance templates of the training examples. For each object class ci (excluding background) an appearance subspace, represented by d x mi projection matrix Si, is constructed. m is chosen such that eigenvectors represent 99% of respective subspace. Appearance features for base learners are obtained by projecting a training example ‘r’ into appearance subspace of each object class. For two object classes the feature vector v of an example will be,
The Online Co-training Framework
The boosting mechanism selects the least correlated base classifiers, which is an ideal property for co-training. The examples confidently labeled by one classifier are used to train the other classifier. In order to carry out this step, only the examples lying close to the decision boundary of the boosted classifier are useful, as classifiying such examples correctly will improve the classification performance. In other words, we employ examples with small margins for online training.
The implementation steps of the online co-training framework are:
- Determine confidence thresholds for each base classifier, using a validation data set.
- For class ci and jth base classifier hj set the confidence threshold, T_{j,ci}^{base}, (highest probability achieved by a negative example).
- Compute Margin thresholds T_{j,ci}^{base} from the validation data set.
During the test phase, select example for training, if i) more than 10% of the classifiers confidently predict the label of an example, or ii) example’s margin is less than the computed thresholds. Once an example has been labeled by the co-training mechanism, an online boosting algorithm by [Oza and Russel, 2002] is used to update the base classifiers and the boosting coefficients.
Information flow for the real-time object classification system