Boosting Classification Based Similarity Learning by using Standard Distances
نویسندگان
چکیده
Metric learning has been shown to outperform standard classification based similarity learning in a number of different contexts. In this paper, we show that the performance of classification similarity learning strongly depends on the sample format used to learn the model. We also propose an enriched classification based set-up that uses a set of standard distances to supplement the information provided by the feature vectors of the training samples. The method is compared to state-of-the-art metric learning methods, using a linear SVM for classification. Results obtained show comparable performances, slightly in favour of the method
منابع مشابه
Multiclass Semi-supervised Boosting Using Different Distance Metrics
The goal of this thesis project is to build an effective multiclass classifier which can be trained with a small amount of labeled data and a large pool of unlabeled data by applying semi-supervised learning in a boosting framework. Boosting refers to a general method of producing a very accurate classifier by combining rough and moderately inaccurate classifiers. It has attracted a significant...
متن کاملBoosting the distance estimation: Application to the K-Nearest Neighbor Classifier
In this work we introduce a new distance estimation technique by boosting and we apply it to the K-Nearest Neighbor Classifier (KNN). Instead of applying AdaBoost to a typical classification problem, we use it for learning a distance function and the resulting distance is used into K-NN. The proposed method (Boosted Distance with Nearest Neighbor) outperforms the AdaBoost classifier when the tr...
متن کاملImproving Imbalanced data classification accuracy by using Fuzzy Similarity Measure and subtractive clustering
Classification is an one of the important parts of data mining and knowledge discovery. In most cases, the data that is utilized to used to training the clusters is not well distributed. This inappropriate distribution occurs when one class has a large number of samples but while the number of other class samples is naturally inherently low. In general, the methods of solving this kind of prob...
متن کاملRelational Learning Using Constrained Confidence-Rated Boosting
In propositional learning, boosting has been a very popular technique for increasing the accuracy of classification learners. In firstorder learning, on the other hand, surprisingly little attention has been paid to boosting, perhaps due to the fact that simple forms of boosting lead to loss of comprehensibility and are too slow when used with standard ILP learners. In this paper, we show how b...
متن کاملOn boosting kernel density methods for multivariate data: density estimation and classification
Statistical learning is emerging as a promising field where a number of algorithms from machine learning are interpreted as statistical methods and vice–versa. Due to good practical performance, boosting is one of the most studied machine learning techniques. We propose algorithms for multivariate density estimation and classification. They are generated by using the traditional kernel techniqu...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2015