Welcome to the Model Metrics Python Library Documentation!
Note
This documentation is for model_metrics
version 0.0.3a
.
Welcome to Model Metrics! Model Metrics is a versatile Python library designed to streamline the evaluation and interpretation of machine learning models. It provides a robust framework for generating predictions, computing model metrics, analyzing feature importance, and visualizing results. Whether you’re working with SHAP values, model coefficients, confusion matrices, ROC curves, precision-recall plots, and other key performance indicators.
Project Links
What is Model Evaluation?
Model evaluation is a fundamental aspect of the machine learning lifecycle. It involves assessing the performance of predictive models using various metrics to ensure accuracy, reliability, and fairness. Proper evaluation helps in understanding how well a model generalizes to unseen data, detects potential biases, and optimizes performance. This step is critical before deploying any model into production.
Purpose of Model Metrics Library
The model_metrics
library is a robust framework designed to simplify and
standardize the evaluation of machine learning models. It provides an extensive
set of tools to assess model performance, compare different approaches, and
validate results with statistical rigor. Key functionalities include:
Performance Metrics: A suite of functions to compute essential metrics such as accuracy, precision, recall, F1-score, ROC-AUC, log loss, and RMSE, among others.
Custom Evaluation Pipelines: Predefined and customizable pipelines for automating model evaluation workflows.
Visualization Tools: Functions to generate confusion matrices, ROC curves, precision-recall curves, calibration plots, and lift and gain charts.
Comparison and Benchmarking: Frameworks to compare multiple models based on key metrics and statistical significance tests.
Key Features
Comprehensive Evaluation: Supports a wide range of model evaluation methods, ensuring a holistic assessment of predictive performance.
User-Friendly: Designed for ease of use, with intuitive functions and well-documented workflows.
Customizable and Extensible: Allows users to configure metric calculations and integrate with different machine learning frameworks.
Seamless Integration: Works with popular libraries such as
Scikit-Learn
,XGBoost
,LightGBM
, andTensorFlow
, provided that model objects follow standard prediction interfaces likepredict()
,predict_proba()
, ordecision_function()
. Special considerations may be required for deep learning models, time-series models, or custom transformers that return non-standard outputs.Detailed Reports: Provides automated summaries and visual insights to aid in model selection and decision-making.
Prerequisites
Before you install model_metrics
, ensure your system meets the following requirements:
Python: version
3.7.4
or higher is required to runmodel_metrics
.
Additionally, model_metrics
depends on the following packages, which will be automatically installed when you install model_metrics
:
matplotlib
: version3.5.3
or higher, but capped at3.9.2
numpy
: version1.21.6
or higher, but capped at2.1.0
pandas
: version1.3.5
or higher, but capped at2.2.3
plotly
: version5.18.0
or higher, but capped at5.24.0
scikit-learn
: version1.0.2
or higher, but capped at1.5.2
shap
: version0.41.0
or higher, but capped below0.46.0
statsmodels
: version0.12.2
or higher, but capped below0.14.4
tqdm`
: version4.66.4
or higher, but capped below4.67.1
Installation
You can install model_metrics
directly from PyPI:
pip install model_metrics
Description
This guide provides detailed instructions and examples for using the functions
provided in the model_metrics
library and how to use them effectively in your projects.
For most of the ensuing examples, we will leverage the Census Income Data (1994) from
the UCI Machine Learning Repository [1]. This dataset provides a rich source of
information for demonstrating the functionalities of the model_metrics
.