Can K-nearest neighbor classify more than two classes?

8 次查看(过去 30 天)
I want to know that K-nearest neighbor classifier can classify more than two classes? I don't understand much about this classifier.
  3 个评论
gugu
gugu 2017-12-18
Ok. I'll have a go at it. Actually, I want to classify three classes and now I'm currently using multi-svm. It doesn't meet my requirements. So, I want to change the classifier. Thank you for your answer.
Ehsan Altayef
Ehsan Altayef 2020-4-20
i need knn algorithm code, please anyone has could you send it to me on Tripoli201919@gmail.com.Thank you in advance.

请先登录,再进行评论。

回答(1 个)

the cyclist
the cyclist 2017-12-17
Yes, it can. There is an example of a 3-class classification in the documentation for fitcknn, in the Statistics and Machine Learning Toolbox.
load fisheriris
X = meas;
Y = species;
% X is a numeric matrix that contains four petal measurements for 150 irises.
% Y is a cell array of character vectors that contains the corresponding iris species.
% Train a 5-nearest neighbors classifier. It is good practice to standardize noncategorical predictor data.
Mdl = fitcknn(X,Y,'NumNeighbors',5,'Standardize',1)
  4 个评论
Image Analyst
Image Analyst 2017-12-18
Not necessarily. Denser data may make it more accurate, but just expanding the feature space so that it's bigger will not affect the results. For example if you had 500 training points in the range x=4-5, and y = 4-5, and you increased it to 2000 training points in that same area would make it more accurate. However if you still had 500 points there and just added another few thousand points in the range 0-10 in both x and y, would not necessarily make your classifications in the range 4-5 any better than what they were, since it would be using the same points as before.

请先登录,再进行评论。

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by