Supervised Machine Learning with a Novel Kernel Density Estimator

Oct 16, 2007

Yen-Jen Oyang, Darby Tien-Hao Chang, Yu-Yen Ou, Hao-Geng Hung, Chih-Peng Wu, Chien-Yu Chen

In recent years, kernel density estimation has been exploited by computer scientists to model machine learning problems. The kernel density estimation based approaches are of interest due to the low time complexity of either O(n) or O(n*log(n)) for constructing a classifier, where n is the number of sampling instances. Concerning design of kernel density estimators, one essential issue is how fast the pointwise mean square error (MSE) and/or the integrated mean square error (IMSE) diminish as the number of sampling instances increases. In this article, it is shown that with the proposed kernel function it is feasible to make the pointwise MSE of the density estimator converge at O(n^-2/3) regardless of the dimension of the vector space, provided that the probability density function at the point of interest meets certain conditions.
Oct 16, 2007

Yen-Jen Oyang, Darby Tien-Hao Chang, Yu-Yen Ou, Hao-Geng Hung, Chih-Peng Wu, Chien-Yu Chen

* The new version includes an additional theorem, Theorem 3

**Click to Read Paper and Get Code**

Knowledge Distillation with Feature Maps for Image Classification

Dec 03, 2018

Wei-Chun Chen, Chia-Che Chang, Chien-Yu Lu, Che-Rung Lee

Dec 03, 2018

Wei-Chun Chen, Chia-Che Chang, Chien-Yu Lu, Che-Rung Lee

* Knowledge Distillation, Model Compression, and Generative Adversarial Network, ACCV 2018

**Click to Read Paper and Get Code**

Supervised Machine Learning with a Novel Pointwise Density Estimator

Nov 06, 2007

Yen-Jen Oyang, Chien-Yu Chen, Darby Tien-Hao Chang, Chih-Peng Wu

This article proposes a novel density estimation based algorithm for carrying out supervised machine learning. The proposed algorithm features O(n) time complexity for generating a classifier, where n is the number of sampling instances in the training dataset. This feature is highly desirable in contemporary applications that involve large and still growing databases. In comparison with the kernel density estimation based approaches, the mathe-matical fundamental behind the proposed algorithm is not based on the assump-tion that the number of training instances approaches infinite. As a result, a classifier generated with the proposed algorithm may deliver higher prediction accuracy than the kernel density estimation based classifier in some cases.
Nov 06, 2007

Yen-Jen Oyang, Chien-Yu Chen, Darby Tien-Hao Chang, Chih-Peng Wu

* Inclusion of a new "Remarks" section

**Click to Read Paper and Get Code**