feat: + cross_validate, trait Predictor, refactoring
This commit is contained in:
+56
-27
@@ -49,13 +49,7 @@
|
||||
//! let y: Vec<f64> = vec![83.0, 88.5, 88.2, 89.5, 96.2, 98.1, 99.0,
|
||||
//! 100.0, 101.2, 104.6, 108.4, 110.8, 112.6, 114.2, 115.7, 116.9];
|
||||
//!
|
||||
//! let svr = SVR::fit(&x, &y,
|
||||
//! LinearKernel {},
|
||||
//! SVRParameters {
|
||||
//! eps: 2.0,
|
||||
//! c: 10.0,
|
||||
//! tol: 1e-3,
|
||||
//! }).unwrap();
|
||||
//! let svr = SVR::fit(&x, &y, SVRParameters::default().with_eps(2.0).with_c(10.0)).unwrap();
|
||||
//!
|
||||
//! let y_hat = svr.predict(&x).unwrap();
|
||||
//! ```
|
||||
@@ -72,25 +66,30 @@
|
||||
|
||||
use std::cell::{Ref, RefCell};
|
||||
use std::fmt::Debug;
|
||||
use std::marker::PhantomData;
|
||||
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
use crate::base::Predictor;
|
||||
use crate::error::Failed;
|
||||
use crate::linalg::BaseVector;
|
||||
use crate::linalg::Matrix;
|
||||
use crate::math::num::RealNumber;
|
||||
use crate::svm::Kernel;
|
||||
|
||||
#[derive(Serialize, Deserialize, Debug)]
|
||||
use crate::svm::{Kernel, Kernels, LinearKernel};
|
||||
|
||||
#[derive(Serialize, Deserialize, Debug, Clone)]
|
||||
/// SVR Parameters
|
||||
pub struct SVRParameters<T: RealNumber> {
|
||||
/// Epsilon in the epsilon-SVR model
|
||||
pub struct SVRParameters<T: RealNumber, M: Matrix<T>, K: Kernel<T, M::RowVector>> {
|
||||
/// Epsilon in the epsilon-SVR model.
|
||||
pub eps: T,
|
||||
/// Regularization parameter.
|
||||
pub c: T,
|
||||
/// Tolerance for stopping criterion
|
||||
/// Tolerance for stopping criterion.
|
||||
pub tol: T,
|
||||
/// The kernel function.
|
||||
pub kernel: K,
|
||||
/// Unused parameter.
|
||||
m: PhantomData<M>,
|
||||
}
|
||||
|
||||
#[derive(Serialize, Deserialize, Debug)]
|
||||
@@ -135,16 +134,52 @@ struct Cache<T: Clone> {
|
||||
data: Vec<RefCell<Option<Vec<T>>>>,
|
||||
}
|
||||
|
||||
impl<T: RealNumber> Default for SVRParameters<T> {
|
||||
impl<T: RealNumber, M: Matrix<T>, K: Kernel<T, M::RowVector>> SVRParameters<T, M, K> {
|
||||
/// Epsilon in the epsilon-SVR model.
|
||||
pub fn with_eps(mut self, eps: T) -> Self {
|
||||
self.eps = eps;
|
||||
self
|
||||
}
|
||||
/// Regularization parameter.
|
||||
pub fn with_c(mut self, c: T) -> Self {
|
||||
self.c = c;
|
||||
self
|
||||
}
|
||||
/// Tolerance for stopping criterion.
|
||||
pub fn with_tol(mut self, tol: T) -> Self {
|
||||
self.tol = tol;
|
||||
self
|
||||
}
|
||||
/// The kernel function.
|
||||
pub fn with_kernel<KK: Kernel<T, M::RowVector>>(&self, kernel: KK) -> SVRParameters<T, M, KK> {
|
||||
SVRParameters {
|
||||
eps: self.eps,
|
||||
c: self.c,
|
||||
tol: self.tol,
|
||||
kernel: kernel,
|
||||
m: PhantomData
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: RealNumber, M: Matrix<T>> Default for SVRParameters<T, M, LinearKernel> {
|
||||
fn default() -> Self {
|
||||
SVRParameters {
|
||||
eps: T::from_f64(0.1).unwrap(),
|
||||
c: T::one(),
|
||||
tol: T::from_f64(1e-3).unwrap(),
|
||||
kernel: Kernels::linear(),
|
||||
m: PhantomData
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: RealNumber, M: Matrix<T>, K: Kernel<T, M::RowVector>> Predictor<M, M::RowVector> for SVR<T, M, K> {
|
||||
fn predict(&self, x: &M) -> Result<M::RowVector, Failed> {
|
||||
self.predict(x)
|
||||
}
|
||||
}
|
||||
|
||||
impl<T: RealNumber, M: Matrix<T>, K: Kernel<T, M::RowVector>> SVR<T, M, K> {
|
||||
/// Fits SVR to your data.
|
||||
/// * `x` - _NxM_ matrix with _N_ observations and _M_ features in each observation.
|
||||
@@ -153,9 +188,8 @@ impl<T: RealNumber, M: Matrix<T>, K: Kernel<T, M::RowVector>> SVR<T, M, K> {
|
||||
/// * `parameters` - optional parameters, use `Default::default()` to set parameters to default values.
|
||||
pub fn fit(
|
||||
x: &M,
|
||||
y: &M::RowVector,
|
||||
kernel: K,
|
||||
parameters: SVRParameters<T>,
|
||||
y: &M::RowVector,
|
||||
parameters: SVRParameters<T, M, K>,
|
||||
) -> Result<SVR<T, M, K>, Failed> {
|
||||
let (n, _) = x.shape();
|
||||
|
||||
@@ -165,12 +199,12 @@ impl<T: RealNumber, M: Matrix<T>, K: Kernel<T, M::RowVector>> SVR<T, M, K> {
|
||||
));
|
||||
}
|
||||
|
||||
let optimizer = Optimizer::new(x, y, &kernel, ¶meters);
|
||||
let optimizer = Optimizer::new(x, y, ¶meters.kernel, ¶meters);
|
||||
|
||||
let (support_vectors, weight, b) = optimizer.smo();
|
||||
|
||||
Ok(SVR {
|
||||
kernel,
|
||||
kernel: parameters.kernel,
|
||||
instances: support_vectors,
|
||||
w: weight,
|
||||
b,
|
||||
@@ -243,7 +277,7 @@ impl<'a, T: RealNumber, M: Matrix<T>, K: Kernel<T, M::RowVector>> Optimizer<'a,
|
||||
x: &M,
|
||||
y: &M::RowVector,
|
||||
kernel: &'a K,
|
||||
parameters: &SVRParameters<T>,
|
||||
parameters: &SVRParameters<T, M, K>,
|
||||
) -> Optimizer<'a, T, M, K> {
|
||||
let (n, _) = x.shape();
|
||||
|
||||
@@ -513,12 +547,7 @@ mod tests {
|
||||
let y_hat = SVR::fit(
|
||||
&x,
|
||||
&y,
|
||||
LinearKernel {},
|
||||
SVRParameters {
|
||||
eps: 2.0,
|
||||
c: 10.0,
|
||||
tol: 1e-3,
|
||||
},
|
||||
SVRParameters::default().with_eps(2.0).with_c(10.0),
|
||||
)
|
||||
.and_then(|lr| lr.predict(&x))
|
||||
.unwrap();
|
||||
@@ -552,7 +581,7 @@ mod tests {
|
||||
114.2, 115.7, 116.9,
|
||||
];
|
||||
|
||||
let svr = SVR::fit(&x, &y, LinearKernel {}, Default::default()).unwrap();
|
||||
let svr = SVR::fit(&x, &y, Default::default()).unwrap();
|
||||
|
||||
let deserialized_svr: SVR<f64, DenseMatrix<f64>, LinearKernel> =
|
||||
serde_json::from_str(&serde_json::to_string(&svr).unwrap()).unwrap();
|
||||
|
||||
Reference in New Issue
Block a user