Simple python package to sanitize in a standard way ML-related labels.
Project description
Simple python package to sanitize in a standard way ML-related labels.
How do I install this package?
As usual, just download it using pip:
pip install sanitize_ml_labels
Tests Coverage
I have strived to mantain a 100% code coverage in this project:
Module |
statements |
missing |
excluded |
coverage |
---|---|---|---|---|
Total |
84 |
0 |
0 |
100% |
sanitize _ml_labels/__init__.py |
3 |
0 |
0 |
100% |
sanitize_ml _labels/__version__.py |
1 |
0 |
0 |
100% |
sanitize_ml_labels/is _normalized_metric.py |
10 |
0 |
0 |
100% |
sanitize_ml_labels /sanitize_ml_labels.py |
70 |
0 |
0 |
100% |
You can verify the test coverage of this repository by running in its root:
pytest --cov
Why do I need this?
So you have some kind of plot and you have some ML-related labels. Since I always rename and sanitize them the same way, I have prepared this package to always sanitize them in a standard fashion.
Usage examples
Here you have a couple of common examples: you have a set of metrics to normalize or a set of model names to normalize.
from sanitize_ml_labels import sanitize_ml_labels
# Example for metrics
labels = [
"acc",
"loss",
"auroc",
"lr"
]
sanitize_ml_labels(labels)
# ["Accuracy", "Loss", "AUROC", "Learning rate"]
# Example for models
labels = [
"vanilla mlp",
"vanilla cnn",
"vanilla ffnn",
"vanilla perceptron"
]
sanitize_ml_labels(labels)
# ["MLP", "CNN", "FFNN", "Perceptron"]
Extra utilities
Since I always use metric sanitization alongside axis normalization, it is useful to know which axis should be maxed between zero and one to avoid any visualization bias to the metrics.
For this reason I have created the method is_normalized_metric
, which after having normalized the given metric
validates it against known normalized metrics (metrics between 0 and 1, is there another name? I could not figure out a better one).
Analogously, I have also created the method is_absolutely_normalized_metric
to validate a metric for the range
between -1 and 1.
from sanitize_ml_labels import is_normalized_metric, is_absolutely_normalized_metric
is_normalized_metric("MSE") # False
is_normalized_metric("acc") # True
is_normalized_metric("accuracy") # True
is_normalized_metric("AUROC") # True
is_normalized_metric("auprc") # True
is_absolutely_normalized_metric("auprc") # False
is_absolutely_normalized_metric("MCC") # True
is_absolutely_normalized_metric("Markedness") # True
New features and issues
As always, for new features and issues you can either open a new issue and pull request. A pull request will always be the quicker way, but I’ll look into the issues when I get the time.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Hashes for sanitize_ml_labels-1.0.33.tar.gz
Algorithm | Hash digest | |
---|---|---|
SHA256 | b633321560de922983fc4a4a8193239e6fe9963163dfcca2e32c26f5b86fb43a |
|
MD5 | fa1634ae5074aeea9e0a160d77e43eab |
|
BLAKE2b-256 | fd717f329b40e604f4c2a85b42c3b9781c9776a7c87cd9bdc42ba9f197385c46 |