Merge pull request #117 from morenol/lmm/fix_clippy

Fix clippy warnings
This commit is contained in:
VolodymyrOrlov
2021-10-27 11:01:16 -07:00
committed by GitHub
11 changed files with 23 additions and 24 deletions
+5 -5
View File
@@ -117,7 +117,7 @@ impl<T: Debug + PartialEq, F: RealNumber, D: Distance<T, F>> CoverTree<T, F, D>
}
let e = self.get_data_value(self.root.idx);
let mut d = self.distance.distance(&e, p);
let mut d = self.distance.distance(e, p);
let mut current_cover_set: Vec<(F, &Node<F>)> = Vec::new();
let mut zero_set: Vec<(F, &Node<F>)> = Vec::new();
@@ -175,7 +175,7 @@ impl<T: Debug + PartialEq, F: RealNumber, D: Distance<T, F>> CoverTree<T, F, D>
if ds.0 <= upper_bound {
let v = self.get_data_value(ds.1.idx);
if !self.identical_excluded || v != p {
neighbors.push((ds.1.idx, ds.0, &v));
neighbors.push((ds.1.idx, ds.0, v));
}
}
}
@@ -200,7 +200,7 @@ impl<T: Debug + PartialEq, F: RealNumber, D: Distance<T, F>> CoverTree<T, F, D>
let mut zero_set: Vec<(F, &Node<F>)> = Vec::new();
let e = self.get_data_value(self.root.idx);
let mut d = self.distance.distance(&e, p);
let mut d = self.distance.distance(e, p);
current_cover_set.push((d, &self.root));
while !current_cover_set.is_empty() {
@@ -230,7 +230,7 @@ impl<T: Debug + PartialEq, F: RealNumber, D: Distance<T, F>> CoverTree<T, F, D>
for ds in zero_set {
let v = self.get_data_value(ds.1.idx);
if !self.identical_excluded || v != p {
neighbors.push((ds.1.idx, ds.0, &v));
neighbors.push((ds.1.idx, ds.0, v));
}
}
@@ -287,7 +287,7 @@ impl<T: Debug + PartialEq, F: RealNumber, D: Distance<T, F>> CoverTree<T, F, D>
if point_set.is_empty() {
self.new_leaf(p)
} else {
let max_dist = self.max(&point_set);
let max_dist = self.max(point_set);
let next_scale = (max_scale - 1).min(self.get_scale(max_dist));
if next_scale == std::i64::MIN {
let mut children: Vec<Node<F>> = Vec::new();
+2 -2
View File
@@ -74,7 +74,7 @@ impl<T, F: RealNumber, D: Distance<T, F>> LinearKNNSearch<T, F, D> {
}
for i in 0..self.data.len() {
let d = self.distance.distance(&from, &self.data[i]);
let d = self.distance.distance(from, &self.data[i]);
let datum = heap.peek_mut();
if d < datum.distance {
datum.distance = d;
@@ -104,7 +104,7 @@ impl<T, F: RealNumber, D: Distance<T, F>> LinearKNNSearch<T, F, D> {
let mut neighbors: Vec<(usize, F, &T)> = Vec::new();
for i in 0..self.data.len() {
let d = self.distance.distance(&from, &self.data[i]);
let d = self.distance.distance(from, &self.data[i]);
if d <= radius {
neighbors.push((i, d, &self.data[i]));
+1 -2
View File
@@ -53,8 +53,7 @@ impl<'a, T: PartialOrd + Debug> HeapSelection<T> {
if self.sorted {
&self.heap[0]
} else {
&self
.heap
self.heap
.iter()
.max_by(|a, b| a.partial_cmp(b).unwrap())
.unwrap()
+1 -1
View File
@@ -6,7 +6,7 @@
clippy::upper_case_acronyms
)]
#![warn(missing_docs)]
#![warn(missing_doc_code_examples)]
#![warn(rustdoc::missing_doc_code_examples)]
//! # SmartCore
//!
+1 -1
View File
@@ -330,7 +330,7 @@ impl<T: RealNumber> DenseMatrix<T> {
cur_r: 0,
max_c: self.ncols,
max_r: self.nrows,
m: &self,
m: self,
}
}
}
+2 -2
View File
@@ -178,7 +178,7 @@ impl<T: RealNumber + ScalarOperand> BaseVector<T> for ArrayBase<OwnedRepr<T>, Ix
}
fn copy_from(&mut self, other: &Self) {
self.assign(&other);
self.assign(other);
}
}
@@ -385,7 +385,7 @@ impl<T: RealNumber + ScalarOperand + AddAssign + SubAssign + MulAssign + DivAssi
}
fn copy_from(&mut self, other: &Self) {
self.assign(&other);
self.assign(other);
}
fn abs_mut(&mut self) -> &Self {
@@ -50,14 +50,14 @@ impl<T: RealNumber> FirstOrderOptimizer<T> for GradientDescent<T> {
let f_alpha = |alpha: T| -> T {
let mut dx = step.clone();
dx.mul_scalar_mut(alpha);
f(&dx.add_mut(&x)) // f(x) = f(x .+ gvec .* alpha)
f(dx.add_mut(&x)) // f(x) = f(x .+ gvec .* alpha)
};
let df_alpha = |alpha: T| -> T {
let mut dx = step.clone();
let mut dg = gvec.clone();
dx.mul_scalar_mut(alpha);
df(&mut dg, &dx.add_mut(&x)); //df(x) = df(x .+ gvec .* alpha)
df(&mut dg, dx.add_mut(&x)); //df(x) = df(x .+ gvec .* alpha)
gvec.dot(&dg)
};
@@ -66,7 +66,7 @@ impl<T: RealNumber> FirstOrderOptimizer<T> for GradientDescent<T> {
let ls_r = ls.search(&f_alpha, &df_alpha, alpha, fx, df0);
alpha = ls_r.alpha;
fx = ls_r.f_x;
x.add_mut(&step.mul_scalar_mut(alpha));
x.add_mut(step.mul_scalar_mut(alpha));
df(&mut gvec, &x);
gnorm = gvec.norm2();
}
+3 -3
View File
@@ -117,14 +117,14 @@ impl<T: RealNumber> LBFGS<T> {
let f_alpha = |alpha: T| -> T {
let mut dx = state.s.clone();
dx.mul_scalar_mut(alpha);
f(&dx.add_mut(&state.x)) // f(x) = f(x .+ gvec .* alpha)
f(dx.add_mut(&state.x)) // f(x) = f(x .+ gvec .* alpha)
};
let df_alpha = |alpha: T| -> T {
let mut dx = state.s.clone();
let mut dg = state.x_df.clone();
dx.mul_scalar_mut(alpha);
df(&mut dg, &dx.add_mut(&state.x)); //df(x) = df(x .+ gvec .* alpha)
df(&mut dg, dx.add_mut(&state.x)); //df(x) = df(x .+ gvec .* alpha)
state.x_df.dot(&dg)
};
@@ -206,7 +206,7 @@ impl<T: RealNumber> FirstOrderOptimizer<T> for LBFGS<T> {
) -> OptimizerResult<T, X> {
let mut state = self.init_state(x0);
df(&mut state.x_df, &x0);
df(&mut state.x_df, x0);
let g_converged = state.x_df.norm(T::infinity()) < self.g_atol;
let mut converged = g_converged;
+2 -2
View File
@@ -377,7 +377,7 @@ impl<'a, T: RealNumber, M: Matrix<T>, K: Kernel<T, M::RowVector>> Optimizer<'a,
Optimizer {
x,
y,
parameters: &parameters,
parameters,
svmin: 0,
svmax: 0,
gmin: T::max_value(),
@@ -589,7 +589,7 @@ impl<'a, T: RealNumber, M: Matrix<T>, K: Kernel<T, M::RowVector>> Optimizer<'a,
for i in 0..self.sv.len() {
let v = &self.sv[i];
let z = v.grad - gm;
let k = cache.get(sv1, &v);
let k = cache.get(sv1, v);
let mut curv = km + v.k - T::two() * k;
if curv <= T::zero() {
curv = self.tau;
+2 -2
View File
@@ -380,7 +380,7 @@ impl<T: RealNumber> DecisionTreeClassifier<T> {
depth: 0,
};
let mut visitor = NodeVisitor::<T, M>::new(0, samples, &order, &x, &yi, 1);
let mut visitor = NodeVisitor::<T, M>::new(0, samples, &order, x, &yi, 1);
let mut visitor_queue: LinkedList<NodeVisitor<'_, T, M>> = LinkedList::new();
@@ -541,7 +541,7 @@ impl<T: RealNumber> DecisionTreeClassifier<T> {
- T::from(tc).unwrap() / T::from(n).unwrap()
* impurity(&self.parameters.criterion, &true_count, tc)
- T::from(fc).unwrap() / T::from(n).unwrap()
* impurity(&self.parameters.criterion, &false_count, fc);
* impurity(&self.parameters.criterion, false_count, fc);
if self.nodes[visitor.node].split_score == Option::None
|| gain > self.nodes[visitor.node].split_score.unwrap()
+1 -1
View File
@@ -280,7 +280,7 @@ impl<T: RealNumber> DecisionTreeRegressor<T> {
depth: 0,
};
let mut visitor = NodeVisitor::<T, M>::new(0, samples, &order, &x, &y_m, 1);
let mut visitor = NodeVisitor::<T, M>::new(0, samples, &order, x, &y_m, 1);
let mut visitor_queue: LinkedList<NodeVisitor<'_, T, M>> = LinkedList::new();