feat: adds accuracy, recall and precision metrics
This commit is contained in:
+2
-1
@@ -8,4 +8,5 @@ pub mod linalg;
|
|||||||
pub mod math;
|
pub mod math;
|
||||||
pub mod algorithm;
|
pub mod algorithm;
|
||||||
pub mod common;
|
pub mod common;
|
||||||
pub mod optimization;
|
pub mod optimization;
|
||||||
|
pub mod metrics;
|
||||||
+10
-1
@@ -16,9 +16,18 @@ use evd::EVDDecomposableMatrix;
|
|||||||
use qr::QRDecomposableMatrix;
|
use qr::QRDecomposableMatrix;
|
||||||
use lu::LUDecomposableMatrix;
|
use lu::LUDecomposableMatrix;
|
||||||
|
|
||||||
|
pub trait BaseVector<T: FloatExt>: Clone + Debug {
|
||||||
|
|
||||||
|
fn get(&self, i: usize) -> T;
|
||||||
|
|
||||||
|
fn set(&mut self, i: usize, x: T);
|
||||||
|
|
||||||
|
fn len(&self) -> usize;
|
||||||
|
}
|
||||||
|
|
||||||
pub trait BaseMatrix<T: FloatExt>: Clone + Debug {
|
pub trait BaseMatrix<T: FloatExt>: Clone + Debug {
|
||||||
|
|
||||||
type RowVector: Clone + Debug;
|
type RowVector: BaseVector<T> + Clone + Debug;
|
||||||
|
|
||||||
fn from_row_vector(vec: Self::RowVector) -> Self;
|
fn from_row_vector(vec: Self::RowVector) -> Self;
|
||||||
|
|
||||||
|
|||||||
@@ -9,13 +9,26 @@ use serde::ser::{Serializer, SerializeStruct};
|
|||||||
use serde::de::{Deserializer, Visitor, SeqAccess, MapAccess};
|
use serde::de::{Deserializer, Visitor, SeqAccess, MapAccess};
|
||||||
|
|
||||||
use crate::linalg::Matrix;
|
use crate::linalg::Matrix;
|
||||||
pub use crate::linalg::BaseMatrix;
|
pub use crate::linalg::{BaseMatrix, BaseVector};
|
||||||
use crate::linalg::svd::SVDDecomposableMatrix;
|
use crate::linalg::svd::SVDDecomposableMatrix;
|
||||||
use crate::linalg::evd::EVDDecomposableMatrix;
|
use crate::linalg::evd::EVDDecomposableMatrix;
|
||||||
use crate::linalg::qr::QRDecomposableMatrix;
|
use crate::linalg::qr::QRDecomposableMatrix;
|
||||||
use crate::linalg::lu::LUDecomposableMatrix;
|
use crate::linalg::lu::LUDecomposableMatrix;
|
||||||
use crate::math::num::FloatExt;
|
use crate::math::num::FloatExt;
|
||||||
|
|
||||||
|
impl<T: FloatExt> BaseVector<T> for Vec<T> {
|
||||||
|
fn get(&self, i: usize) -> T {
|
||||||
|
self[i]
|
||||||
|
}
|
||||||
|
fn set(&mut self, i: usize, x: T){
|
||||||
|
self[i] = x
|
||||||
|
}
|
||||||
|
|
||||||
|
fn len(&self) -> usize {
|
||||||
|
self.len()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone)]
|
#[derive(Debug, Clone)]
|
||||||
pub struct DenseMatrix<T: FloatExt> {
|
pub struct DenseMatrix<T: FloatExt> {
|
||||||
|
|
||||||
|
|||||||
@@ -4,13 +4,26 @@ use std::iter::Sum;
|
|||||||
use nalgebra::{MatrixMN, DMatrix, Matrix, Scalar, Dynamic, U1, VecStorage};
|
use nalgebra::{MatrixMN, DMatrix, Matrix, Scalar, Dynamic, U1, VecStorage};
|
||||||
|
|
||||||
use crate::math::num::FloatExt;
|
use crate::math::num::FloatExt;
|
||||||
use crate::linalg::BaseMatrix;
|
use crate::linalg::{BaseMatrix, BaseVector};
|
||||||
use crate::linalg::Matrix as SmartCoreMatrix;
|
use crate::linalg::Matrix as SmartCoreMatrix;
|
||||||
use crate::linalg::svd::SVDDecomposableMatrix;
|
use crate::linalg::svd::SVDDecomposableMatrix;
|
||||||
use crate::linalg::evd::EVDDecomposableMatrix;
|
use crate::linalg::evd::EVDDecomposableMatrix;
|
||||||
use crate::linalg::qr::QRDecomposableMatrix;
|
use crate::linalg::qr::QRDecomposableMatrix;
|
||||||
use crate::linalg::lu::LUDecomposableMatrix;
|
use crate::linalg::lu::LUDecomposableMatrix;
|
||||||
|
|
||||||
|
impl<T: FloatExt + 'static> BaseVector<T> for MatrixMN<T, U1, Dynamic> {
|
||||||
|
fn get(&self, i: usize) -> T {
|
||||||
|
*self.get((0, i)).unwrap()
|
||||||
|
}
|
||||||
|
fn set(&mut self, i: usize, x: T){
|
||||||
|
*self.get_mut((0, i)).unwrap() = x;
|
||||||
|
}
|
||||||
|
|
||||||
|
fn len(&self) -> usize{
|
||||||
|
self.len()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl<T: FloatExt + Scalar + AddAssign + SubAssign + MulAssign + DivAssign + Sum + 'static> BaseMatrix<T> for Matrix<T, Dynamic, Dynamic, VecStorage<T, Dynamic, Dynamic>>
|
impl<T: FloatExt + Scalar + AddAssign + SubAssign + MulAssign + DivAssign + Sum + 'static> BaseMatrix<T> for Matrix<T, Dynamic, Dynamic, VecStorage<T, Dynamic, Dynamic>>
|
||||||
{
|
{
|
||||||
type RowVector = MatrixMN<T, U1, Dynamic>;
|
type RowVector = MatrixMN<T, U1, Dynamic>;
|
||||||
@@ -340,6 +353,24 @@ mod tests {
|
|||||||
use super::*;
|
use super::*;
|
||||||
use nalgebra::{Matrix2x3, DMatrix, RowDVector};
|
use nalgebra::{Matrix2x3, DMatrix, RowDVector};
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn vec_len() {
|
||||||
|
let v = RowDVector::from_vec(vec!(1., 2., 3.));
|
||||||
|
assert_eq!(3, v.len());
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn get_set_vector() {
|
||||||
|
let mut v = RowDVector::from_vec(vec!(1., 2., 3., 4.));
|
||||||
|
|
||||||
|
let expected = RowDVector::from_vec(vec!(1., 5., 3., 4.));
|
||||||
|
|
||||||
|
v.set(1, 5.);
|
||||||
|
|
||||||
|
assert_eq!(v, expected);
|
||||||
|
assert_eq!(5., BaseVector::get(&v, 1));
|
||||||
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn get_set_dynamic() {
|
fn get_set_dynamic() {
|
||||||
let mut m = DMatrix::from_row_slice(
|
let mut m = DMatrix::from_row_slice(
|
||||||
@@ -355,7 +386,7 @@ mod tests {
|
|||||||
|
|
||||||
assert_eq!(m, expected);
|
assert_eq!(m, expected);
|
||||||
assert_eq!(10., BaseMatrix::get(&m, 1, 1));
|
assert_eq!(10., BaseMatrix::get(&m, 1, 1));
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn zeros() {
|
fn zeros() {
|
||||||
|
|||||||
@@ -9,13 +9,25 @@ use ndarray::{Array, ArrayBase, OwnedRepr, Ix2, Ix1, Axis, stack, s};
|
|||||||
use ndarray::ScalarOperand;
|
use ndarray::ScalarOperand;
|
||||||
|
|
||||||
use crate::math::num::FloatExt;
|
use crate::math::num::FloatExt;
|
||||||
use crate::linalg::BaseMatrix;
|
use crate::linalg::{BaseMatrix, BaseVector};
|
||||||
use crate::linalg::Matrix;
|
use crate::linalg::Matrix;
|
||||||
use crate::linalg::svd::SVDDecomposableMatrix;
|
use crate::linalg::svd::SVDDecomposableMatrix;
|
||||||
use crate::linalg::evd::EVDDecomposableMatrix;
|
use crate::linalg::evd::EVDDecomposableMatrix;
|
||||||
use crate::linalg::qr::QRDecomposableMatrix;
|
use crate::linalg::qr::QRDecomposableMatrix;
|
||||||
use crate::linalg::lu::LUDecomposableMatrix;
|
use crate::linalg::lu::LUDecomposableMatrix;
|
||||||
|
|
||||||
|
impl<T: FloatExt> BaseVector<T> for ArrayBase<OwnedRepr<T>, Ix1> {
|
||||||
|
fn get(&self, i: usize) -> T {
|
||||||
|
self[i]
|
||||||
|
}
|
||||||
|
fn set(&mut self, i: usize, x: T){
|
||||||
|
self[i] = x;
|
||||||
|
}
|
||||||
|
|
||||||
|
fn len(&self) -> usize{
|
||||||
|
self.len()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl<T: FloatExt + ScalarOperand + AddAssign + SubAssign + MulAssign + DivAssign + Sum> BaseMatrix<T> for ArrayBase<OwnedRepr<T>, Ix2>
|
impl<T: FloatExt + ScalarOperand + AddAssign + SubAssign + MulAssign + DivAssign + Sum> BaseMatrix<T> for ArrayBase<OwnedRepr<T>, Ix2>
|
||||||
{
|
{
|
||||||
@@ -308,6 +320,23 @@ mod tests {
|
|||||||
use super::*;
|
use super::*;
|
||||||
use ndarray::{arr1, arr2, Array2};
|
use ndarray::{arr1, arr2, Array2};
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn vec_get_set() {
|
||||||
|
let mut result = arr1(&[1., 2., 3.]);
|
||||||
|
let expected = arr1(&[1., 5., 3.]);
|
||||||
|
|
||||||
|
result.set(1, 5.);
|
||||||
|
|
||||||
|
assert_eq!(result, expected);
|
||||||
|
assert_eq!(5., BaseVector::get(&result, 1));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn vec_len() {
|
||||||
|
let v = arr1(&[1., 2., 3.]);
|
||||||
|
assert_eq!(3, v.len());
|
||||||
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn from_to_row_vec() {
|
fn from_to_row_vec() {
|
||||||
|
|
||||||
@@ -449,7 +478,7 @@ mod tests {
|
|||||||
|
|
||||||
assert_eq!(result, expected);
|
assert_eq!(result, expected);
|
||||||
assert_eq!(10., BaseMatrix::get(&result, 1, 1));
|
assert_eq!(10., BaseMatrix::get(&result, 1, 1));
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn dot() {
|
fn dot() {
|
||||||
|
|||||||
@@ -273,6 +273,7 @@ mod tests {
|
|||||||
use super::*;
|
use super::*;
|
||||||
use crate::linalg::naive::dense_matrix::*;
|
use crate::linalg::naive::dense_matrix::*;
|
||||||
use ndarray::{arr1, arr2, Array1};
|
use ndarray::{arr1, arr2, Array1};
|
||||||
|
use crate::metrics::*;
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn multiclass_objective_f() {
|
fn multiclass_objective_f() {
|
||||||
@@ -447,7 +448,7 @@ mod tests {
|
|||||||
|
|
||||||
let lr = LogisticRegression::fit(&x, &y);
|
let lr = LogisticRegression::fit(&x, &y);
|
||||||
|
|
||||||
let y_hat = lr.predict(&x);
|
let y_hat = lr.predict(&x);
|
||||||
|
|
||||||
let error: f64 = y.into_iter().zip(y_hat.into_iter()).map(|(&a, &b)| (a - b).abs()).sum();
|
let error: f64 = y.into_iter().zip(y_hat.into_iter()).map(|(&a, &b)| (a - b).abs()).sum();
|
||||||
|
|
||||||
|
|||||||
@@ -0,0 +1,45 @@
|
|||||||
|
use serde::{Serialize, Deserialize};
|
||||||
|
|
||||||
|
use crate::math::num::FloatExt;
|
||||||
|
use crate::linalg::BaseVector;
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
|
pub struct Accuracy{}
|
||||||
|
|
||||||
|
impl Accuracy {
|
||||||
|
pub fn get_score<T: FloatExt, V: BaseVector<T>>(&self, y_true: &V, y_prod: &V) -> T {
|
||||||
|
if y_true.len() != y_prod.len() {
|
||||||
|
panic!("The vector sizes don't match: {} != {}", y_true.len(), y_prod.len());
|
||||||
|
}
|
||||||
|
|
||||||
|
let n = y_true.len();
|
||||||
|
|
||||||
|
let mut positive = 0;
|
||||||
|
for i in 0..n {
|
||||||
|
if y_true.get(i) == y_prod.get(i) {
|
||||||
|
positive += 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
T::from_i64(positive).unwrap() / T::from_usize(n).unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn accuracy() {
|
||||||
|
let y_pred: Vec<f64> = vec![0., 2., 1., 3.];
|
||||||
|
let y_true: Vec<f64> = vec![0., 1., 2., 3.];
|
||||||
|
|
||||||
|
let score1: f64 = Accuracy{}.get_score(&y_pred, &y_true);
|
||||||
|
let score2: f64 = Accuracy{}.get_score(&y_true, &y_true);
|
||||||
|
|
||||||
|
assert!((score1 - 0.5).abs() < 1e-8);
|
||||||
|
assert!((score2 - 1.0).abs() < 1e-8);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
@@ -0,0 +1,34 @@
|
|||||||
|
pub mod accuracy;
|
||||||
|
pub mod recall;
|
||||||
|
pub mod precision;
|
||||||
|
|
||||||
|
use crate::math::num::FloatExt;
|
||||||
|
use crate::linalg::BaseVector;
|
||||||
|
|
||||||
|
pub struct ClassificationMetrics{}
|
||||||
|
|
||||||
|
impl ClassificationMetrics {
|
||||||
|
pub fn accuracy() -> accuracy::Accuracy{
|
||||||
|
accuracy::Accuracy {}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn recall() -> recall::Recall{
|
||||||
|
recall::Recall {}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn precision() -> precision::Precision{
|
||||||
|
precision::Precision {}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn accuracy<T: FloatExt, V: BaseVector<T>>(y_true: &V, y_prod: &V) -> T{
|
||||||
|
ClassificationMetrics::accuracy().get_score(y_true, y_prod)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn recall<T: FloatExt, V: BaseVector<T>>(y_true: &V, y_prod: &V) -> T{
|
||||||
|
ClassificationMetrics::recall().get_score(y_true, y_prod)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn precision<T: FloatExt, V: BaseVector<T>>(y_true: &V, y_prod: &V) -> T{
|
||||||
|
ClassificationMetrics::precision().get_score(y_true, y_prod)
|
||||||
|
}
|
||||||
@@ -0,0 +1,57 @@
|
|||||||
|
use serde::{Serialize, Deserialize};
|
||||||
|
|
||||||
|
use crate::math::num::FloatExt;
|
||||||
|
use crate::linalg::BaseVector;
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
|
pub struct Precision{}
|
||||||
|
|
||||||
|
impl Precision {
|
||||||
|
pub fn get_score<T: FloatExt, V: BaseVector<T>>(&self, y_true: &V, y_prod: &V) -> T {
|
||||||
|
if y_true.len() != y_prod.len() {
|
||||||
|
panic!("The vector sizes don't match: {} != {}", y_true.len(), y_prod.len());
|
||||||
|
}
|
||||||
|
|
||||||
|
let mut tp = 0;
|
||||||
|
let mut p = 0;
|
||||||
|
let n = y_true.len();
|
||||||
|
for i in 0..n {
|
||||||
|
if y_true.get(i) != T::zero() && y_true.get(i) != T::one() {
|
||||||
|
panic!("Precision can only be applied to binary classification: {}", y_true.get(i));
|
||||||
|
}
|
||||||
|
|
||||||
|
if y_prod.get(i) != T::zero() && y_prod.get(i) != T::one() {
|
||||||
|
panic!("Precision can only be applied to binary classification: {}", y_prod.get(i));
|
||||||
|
}
|
||||||
|
|
||||||
|
if y_prod.get(i) == T::one() {
|
||||||
|
p += 1;
|
||||||
|
|
||||||
|
if y_true.get(i) == T::one() {
|
||||||
|
tp += 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
T::from_i64(tp).unwrap() / T::from_i64(p).unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn precision() {
|
||||||
|
let y_true: Vec<f64> = vec![0., 1., 1., 0.];
|
||||||
|
let y_pred: Vec<f64> = vec![0., 0., 1., 1.];
|
||||||
|
|
||||||
|
let score1: f64 = Precision{}.get_score(&y_pred, &y_true);
|
||||||
|
let score2: f64 = Precision{}.get_score(&y_pred, &y_pred);
|
||||||
|
|
||||||
|
assert!((score1 - 0.5).abs() < 1e-8);
|
||||||
|
assert!((score2 - 1.0).abs() < 1e-8);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
@@ -0,0 +1,57 @@
|
|||||||
|
use serde::{Serialize, Deserialize};
|
||||||
|
|
||||||
|
use crate::math::num::FloatExt;
|
||||||
|
use crate::linalg::BaseVector;
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug)]
|
||||||
|
pub struct Recall{}
|
||||||
|
|
||||||
|
impl Recall {
|
||||||
|
pub fn get_score<T: FloatExt, V: BaseVector<T>>(&self, y_true: &V, y_prod: &V) -> T {
|
||||||
|
if y_true.len() != y_prod.len() {
|
||||||
|
panic!("The vector sizes don't match: {} != {}", y_true.len(), y_prod.len());
|
||||||
|
}
|
||||||
|
|
||||||
|
let mut tp = 0;
|
||||||
|
let mut p = 0;
|
||||||
|
let n = y_true.len();
|
||||||
|
for i in 0..n {
|
||||||
|
if y_true.get(i) != T::zero() && y_true.get(i) != T::one() {
|
||||||
|
panic!("Recall can only be applied to binary classification: {}", y_true.get(i));
|
||||||
|
}
|
||||||
|
|
||||||
|
if y_prod.get(i) != T::zero() && y_prod.get(i) != T::one() {
|
||||||
|
panic!("Recall can only be applied to binary classification: {}", y_prod.get(i));
|
||||||
|
}
|
||||||
|
|
||||||
|
if y_true.get(i) == T::one() {
|
||||||
|
p += 1;
|
||||||
|
|
||||||
|
if y_prod.get(i) == T::one() {
|
||||||
|
tp += 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
T::from_i64(tp).unwrap() / T::from_i64(p).unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn recall() {
|
||||||
|
let y_true: Vec<f64> = vec![0., 1., 1., 0.];
|
||||||
|
let y_pred: Vec<f64> = vec![0., 0., 1., 1.];
|
||||||
|
|
||||||
|
let score1: f64 = Recall{}.get_score(&y_pred, &y_true);
|
||||||
|
let score2: f64 = Recall{}.get_score(&y_pred, &y_pred);
|
||||||
|
|
||||||
|
assert!((score1 - 0.5).abs() < 1e-8);
|
||||||
|
assert!((score2 - 1.0).abs() < 1e-8);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
Reference in New Issue
Block a user