Our task is to classify 120 breeds of dog appropriately using machine learning methods. For this project, We had tested and observed several classical methods(kNN, SVM). However, these traditional approaches were actually not fitted to classify 120 classes. Thus, we implemented our new approach by combining classifiers and compared to our new approach (Decision Tree with SVM).
First of all, we need to mention that we did not use entire 100 examples from each class. The reason is because we had system limitation, we could not even run ML methods with that amount of data (Even Dartmouth research computing system could not run this program). Thus, we have run our apprach by choosing part out of entire example set (n = 10, 20, 30, 50 , we could not run if n is larger than 50).
As we can see above the results, we got the better accuracy result by inceasing number of samples from each breed. This is because each tree could learn more accurate 'W' and 'b' on training step. Moreover, we could get better performance result as we increased number of trees that consist of Random forest. Our trees were indepently built (reduced correlation), and they returned better results. Thus, we could get some conclusion "the more trees the better result"
Due to the time limitation, we only could try our approach with only up to 10 trees and 50 examples. However, we expect we can get close result compare to Stanford vison lab's experiment in respect that our graph.