KNN and Decision Trees
We have created two simple models using our training data in order to have a preliminary idea of what type of accuracy we are trying to improve upon. The two models we chose are KNN and Decision Tree.
KNN stands for k-nearest-neighbors and in this method, each point is classified based on its neighboring points. Meanwhile, in the Decision Tree method, the data is broken down into smaller and smaller subsets to find distinct features, while also grouping these pieces together to make a final dataset that can be classified.
MATLAB
At first, we used MATLAB to train these models using a small subset of our full-color data: 800 cat and 800 dog images. We then tested them on only 200 cat and 200 dog images. We ran the default MATLAB KNN model which only has k = 1 neighbor.
Below are two confusion matrixes for this subset of the data. As shown below, KNN had an accuracy of 53.75% and decision tree had an accuracy of 56.5% on the subset of the data.