Fix clippy::or_fun_call
This commit is contained in:
@@ -137,13 +137,13 @@ impl<T: RealNumber> RandomForestClassifier<T> {
|
|||||||
yi[i] = classes.iter().position(|c| yc == *c).unwrap();
|
yi[i] = classes.iter().position(|c| yc == *c).unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
let mtry = parameters.m.unwrap_or(
|
let mtry = parameters.m.unwrap_or_else(|| {
|
||||||
(T::from(num_attributes).unwrap())
|
(T::from(num_attributes).unwrap())
|
||||||
.sqrt()
|
.sqrt()
|
||||||
.floor()
|
.floor()
|
||||||
.to_usize()
|
.to_usize()
|
||||||
.unwrap(),
|
.unwrap()
|
||||||
);
|
});
|
||||||
|
|
||||||
let classes = y_m.unique();
|
let classes = y_m.unique();
|
||||||
let k = classes.len();
|
let k = classes.len();
|
||||||
|
|||||||
@@ -65,7 +65,6 @@
|
|||||||
//! ```
|
//! ```
|
||||||
|
|
||||||
#![allow(
|
#![allow(
|
||||||
clippy::or_fun_call,
|
|
||||||
clippy::needless_range_loop,
|
clippy::needless_range_loop,
|
||||||
clippy::ptr_arg,
|
clippy::ptr_arg,
|
||||||
clippy::len_without_is_empty,
|
clippy::len_without_is_empty,
|
||||||
|
|||||||
@@ -24,8 +24,8 @@ impl HCVScore {
|
|||||||
let contingency = contingency_matrix(&labels_true, &labels_pred);
|
let contingency = contingency_matrix(&labels_true, &labels_pred);
|
||||||
let mi: T = mutual_info_score(&contingency);
|
let mi: T = mutual_info_score(&contingency);
|
||||||
|
|
||||||
let homogeneity = entropy_c.map(|e| mi / e).unwrap_or(T::one());
|
let homogeneity = entropy_c.map(|e| mi / e).unwrap_or_else(T::one);
|
||||||
let completeness = entropy_k.map(|e| mi / e).unwrap_or(T::one());
|
let completeness = entropy_k.map(|e| mi / e).unwrap_or_else(T::one);
|
||||||
|
|
||||||
let v_measure_score = if homogeneity + completeness == T::zero() {
|
let v_measure_score = if homogeneity + completeness == T::zero() {
|
||||||
T::zero()
|
T::zero()
|
||||||
|
|||||||
+6
-2
@@ -561,7 +561,9 @@ impl<'a, T: RealNumber, M: Matrix<T>, K: Kernel<T, M::RowVector>> Optimizer<'a,
|
|||||||
(
|
(
|
||||||
idx_1,
|
idx_1,
|
||||||
idx_2,
|
idx_2,
|
||||||
k_v_12.unwrap_or(self.kernel.apply(&self.sv[idx_1].x, &self.sv[idx_2].x)),
|
k_v_12.unwrap_or_else(|| {
|
||||||
|
self.kernel.apply(&self.sv[idx_1].x, &self.sv[idx_2].x)
|
||||||
|
}),
|
||||||
)
|
)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
@@ -597,7 +599,9 @@ impl<'a, T: RealNumber, M: Matrix<T>, K: Kernel<T, M::RowVector>> Optimizer<'a,
|
|||||||
(
|
(
|
||||||
idx_1,
|
idx_1,
|
||||||
idx_2,
|
idx_2,
|
||||||
k_v_12.unwrap_or(self.kernel.apply(&self.sv[idx_1].x, &self.sv[idx_2].x)),
|
k_v_12.unwrap_or_else(|| {
|
||||||
|
self.kernel.apply(&self.sv[idx_1].x, &self.sv[idx_2].x)
|
||||||
|
}),
|
||||||
)
|
)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -376,7 +376,8 @@ impl<T: RealNumber> DecisionTreeClassifier<T> {
|
|||||||
let node = &self.nodes[node_id];
|
let node = &self.nodes[node_id];
|
||||||
if node.true_child == None && node.false_child == None {
|
if node.true_child == None && node.false_child == None {
|
||||||
result = node.output;
|
result = node.output;
|
||||||
} else if x.get(row, node.split_feature) <= node.split_value.unwrap_or(T::nan())
|
} else if x.get(row, node.split_feature)
|
||||||
|
<= node.split_value.unwrap_or_else(T::nan)
|
||||||
{
|
{
|
||||||
queue.push_back(node.true_child.unwrap());
|
queue.push_back(node.true_child.unwrap());
|
||||||
} else {
|
} else {
|
||||||
@@ -529,7 +530,7 @@ impl<T: RealNumber> DecisionTreeClassifier<T> {
|
|||||||
for i in 0..n {
|
for i in 0..n {
|
||||||
if visitor.samples[i] > 0 {
|
if visitor.samples[i] > 0 {
|
||||||
if visitor.x.get(i, self.nodes[visitor.node].split_feature)
|
if visitor.x.get(i, self.nodes[visitor.node].split_feature)
|
||||||
<= self.nodes[visitor.node].split_value.unwrap_or(T::nan())
|
<= self.nodes[visitor.node].split_value.unwrap_or_else(T::nan)
|
||||||
{
|
{
|
||||||
true_samples[i] = visitor.samples[i];
|
true_samples[i] = visitor.samples[i];
|
||||||
tc += true_samples[i];
|
tc += true_samples[i];
|
||||||
|
|||||||
@@ -282,7 +282,8 @@ impl<T: RealNumber> DecisionTreeRegressor<T> {
|
|||||||
let node = &self.nodes[node_id];
|
let node = &self.nodes[node_id];
|
||||||
if node.true_child == None && node.false_child == None {
|
if node.true_child == None && node.false_child == None {
|
||||||
result = node.output;
|
result = node.output;
|
||||||
} else if x.get(row, node.split_feature) <= node.split_value.unwrap_or(T::nan())
|
} else if x.get(row, node.split_feature)
|
||||||
|
<= node.split_value.unwrap_or_else(T::nan)
|
||||||
{
|
{
|
||||||
queue.push_back(node.true_child.unwrap());
|
queue.push_back(node.true_child.unwrap());
|
||||||
} else {
|
} else {
|
||||||
@@ -401,7 +402,7 @@ impl<T: RealNumber> DecisionTreeRegressor<T> {
|
|||||||
for i in 0..n {
|
for i in 0..n {
|
||||||
if visitor.samples[i] > 0 {
|
if visitor.samples[i] > 0 {
|
||||||
if visitor.x.get(i, self.nodes[visitor.node].split_feature)
|
if visitor.x.get(i, self.nodes[visitor.node].split_feature)
|
||||||
<= self.nodes[visitor.node].split_value.unwrap_or(T::nan())
|
<= self.nodes[visitor.node].split_value.unwrap_or_else(T::nan)
|
||||||
{
|
{
|
||||||
true_samples[i] = visitor.samples[i];
|
true_samples[i] = visitor.samples[i];
|
||||||
tc += true_samples[i];
|
tc += true_samples[i];
|
||||||
|
|||||||
Reference in New Issue
Block a user