Release 0.3 (#235)
This commit is contained in:
+1
-1
@@ -2,7 +2,7 @@
|
||||
//! Computes the area under the receiver operating characteristic (ROC) curve that is equal to the probability that a classifier will rank a
|
||||
//! randomly chosen positive instance higher than a randomly chosen negative one.
|
||||
//!
|
||||
//! SmartCore calculates ROC AUC from Wilcoxon or Mann-Whitney U test.
|
||||
//! `smartcore` calculates ROC AUC from Wilcoxon or Mann-Whitney U test.
|
||||
//!
|
||||
//! Example:
|
||||
//! ```
|
||||
|
||||
+1
-1
@@ -4,7 +4,7 @@
|
||||
//! In a feedback loop you build your model first, then you get feedback from metrics, improve it and repeat until your model achieve desirable performance.
|
||||
//! Evaluation metrics helps to explain the performance of a model and compare models based on an objective criterion.
|
||||
//!
|
||||
//! Choosing the right metric is crucial while evaluating machine learning models. In SmartCore you will find metrics for these classes of ML models:
|
||||
//! Choosing the right metric is crucial while evaluating machine learning models. In `smartcore` you will find metrics for these classes of ML models:
|
||||
//!
|
||||
//! * [Classification metrics](struct.ClassificationMetrics.html)
|
||||
//! * [Regression metrics](struct.RegressionMetrics.html)
|
||||
|
||||
Reference in New Issue
Block a user