mirror of
https://github.com/sjwhitworth/golearn.git
synced 2025-04-26 13:49:14 +08:00
150 lines
4.0 KiB
Go
150 lines
4.0 KiB
Go
// Package KNN implements a K Nearest Neighbors object, capable of both classification
|
|
// and regression. It accepts data in the form of a slice of float64s, which are then reshaped
|
|
// into a X by Y matrix.
|
|
package knn
|
|
|
|
import (
|
|
"github.com/gonum/matrix/mat64"
|
|
base "github.com/sjwhitworth/golearn/base"
|
|
pairwiseMetrics "github.com/sjwhitworth/golearn/metrics/pairwise"
|
|
util "github.com/sjwhitworth/golearn/utilities"
|
|
)
|
|
|
|
// A KNN Classifier. Consists of a data matrix, associated labels in the same order as the matrix, and a distance function.
|
|
// The accepted distance functions at this time are 'euclidean' and 'manhattan'.
|
|
type KNNClassifier struct {
|
|
base.BaseEstimator
|
|
Labels []string
|
|
DistanceFunc string
|
|
}
|
|
|
|
// Returns a new classifier
|
|
func NewKnnClassifier(labels []string, numbers []float64, rows int, cols int, distfunc string) *KNNClassifier {
|
|
if rows != len(labels) {
|
|
panic("Number of rows must equal number of labels")
|
|
}
|
|
|
|
KNN := KNNClassifier{}
|
|
KNN.Data = mat64.NewDense(rows, cols, numbers)
|
|
KNN.Labels = labels
|
|
KNN.DistanceFunc = distfunc
|
|
return &KNN
|
|
}
|
|
|
|
// Returns a classification for the vector, based on a vector input, using the KNN algorithm.
|
|
// See http://en.wikipedia.org/wiki/K-nearest_neighbors_algorithm.
|
|
func (KNN *KNNClassifier) Predict(vector []float64, K int) string {
|
|
|
|
convertedVector := util.FloatsToMatrix(vector)
|
|
// Get the number of rows
|
|
rows, _ := KNN.Data.Dims()
|
|
rownumbers := make(map[int]float64)
|
|
labels := make([]string, 0)
|
|
maxmap := make(map[string]int)
|
|
|
|
// Check what distance function we are using
|
|
switch KNN.DistanceFunc {
|
|
case "euclidean":
|
|
{
|
|
euclidean := pairwiseMetrics.NewEuclidean()
|
|
for i := 0; i < rows; i++ {
|
|
row := KNN.Data.RowView(i)
|
|
rowMat := util.FloatsToMatrix(row)
|
|
distance := euclidean.Distance(rowMat, convertedVector)
|
|
rownumbers[i] = distance
|
|
}
|
|
}
|
|
case "manhattan":
|
|
{
|
|
manhattan := pairwiseMetrics.NewEuclidean()
|
|
for i := 0; i < rows; i++ {
|
|
row := KNN.Data.RowView(i)
|
|
rowMat := util.FloatsToMatrix(row)
|
|
distance := manhattan.Distance(rowMat, convertedVector)
|
|
rownumbers[i] = distance
|
|
}
|
|
}
|
|
}
|
|
|
|
sorted := util.SortIntMap(rownumbers)
|
|
values := sorted[:K]
|
|
|
|
for _, elem := range values {
|
|
// It's when we access this map
|
|
labels = append(labels, KNN.Labels[elem])
|
|
|
|
if _, ok := maxmap[KNN.Labels[elem]]; ok {
|
|
maxmap[KNN.Labels[elem]] += 1
|
|
} else {
|
|
maxmap[KNN.Labels[elem]] = 1
|
|
}
|
|
}
|
|
|
|
sortedlabels := util.SortStringMap(maxmap)
|
|
label := sortedlabels[0]
|
|
|
|
return label
|
|
}
|
|
|
|
//A KNN Regressor. Consists of a data matrix, associated result variables in the same order as the matrix, and a name.
|
|
type KNNRegressor struct {
|
|
base.BaseEstimator
|
|
Values []float64
|
|
DistanceFunc string
|
|
}
|
|
|
|
// Mints a new classifier.
|
|
func NewKnnRegressor(values []float64, numbers []float64, x int, y int, distfunc string) *KNNRegressor {
|
|
KNN := KNNRegressor{}
|
|
KNN.Data = mat64.NewDense(x, y, numbers)
|
|
KNN.Values = values
|
|
KNN.DistanceFunc = distfunc
|
|
return &KNN
|
|
}
|
|
|
|
//Returns an average of the K nearest labels/variables, based on a vector input.
|
|
func (KNN *KNNRegressor) Predict(vector *mat64.Dense, K int) float64 {
|
|
|
|
// Get the number of rows
|
|
rows, _ := KNN.Data.Dims()
|
|
rownumbers := make(map[int]float64)
|
|
labels := make([]float64, 0)
|
|
|
|
// Check what distance function we are using
|
|
switch KNN.DistanceFunc {
|
|
case "euclidean":
|
|
{
|
|
euclidean := pairwiseMetrics.NewEuclidean()
|
|
for i := 0; i < rows; i++ {
|
|
row := KNN.Data.RowView(i)
|
|
rowMat := util.FloatsToMatrix(row)
|
|
distance := euclidean.Distance(rowMat, vector)
|
|
rownumbers[i] = distance
|
|
}
|
|
}
|
|
case "manhattan":
|
|
{
|
|
manhattan := pairwiseMetrics.NewEuclidean()
|
|
for i := 0; i < rows; i++ {
|
|
row := KNN.Data.RowView(i)
|
|
rowMat := util.FloatsToMatrix(row)
|
|
distance := manhattan.Distance(rowMat, vector)
|
|
rownumbers[i] = distance
|
|
}
|
|
}
|
|
}
|
|
|
|
sorted := util.SortIntMap(rownumbers)
|
|
values := sorted[:K]
|
|
|
|
var sum float64
|
|
for _, elem := range values {
|
|
value := KNN.Values[elem]
|
|
labels = append(labels, value)
|
|
sum += value
|
|
}
|
|
|
|
average := sum / float64(K)
|
|
return average
|
|
}
|