MLTE (pronounced "melt") is a framework and infrastructure for evaluating machine learning models and systems. To get started with the MLTE Python package, continuing reading below. The MLTE framework can be found in the documentation, along with a more in-depth guide to using MLTE that expands on the quick start guide below. For examples of use cases, see the demo folder.
The MLTE Python package is available on PyPI, and the MLTE framework is described in our documentation. Install the latest version of the package with pip or conda:
$ pip install mlteTo use the web UI (frontend/backend functionality), the frontend optional dependencies are needed; to use relational database storage, the rdbs optional dependencies are needed; and to use the GPU measurements, the gpu optional dependencies are needed. To install all optional dependencies:
$ pip install "mlte[frontend,rdbs,gpu]"MLTE can be imported and used as a regular library to access most of its functionality by importing the mlte package. Before most operations can be done on MLTE, a context and store need to be set via set_context("model_name", "model_version") and set_store("store_uri"), which can be imported as follows:
from mlte.session import set_context, set_storeset_context() indicates the model and version being used for the script, and can be any string. set_store() indicates the location of the store being used for artifacts and other entities, with four store type options described in the documentation. The MLTE context and store can alternatively be set by environment variables before starting the script (MLTE_CONTEXT_MODEL, MLTE_CONTEXT_VERSION, and MLTE_STORE_URI), and can later be overridden using the set methods above.
For a simple example of using the MLTE library, see the simple demo Jupyter notebooks available here.
The MLTE web-based user interface (UI) allows you to view stored artifacts, create/edit some of them, and review existing models and test catalogs. To access the UI, first start the backend server with the following command:
$ mlte backendThere are a number of flags that can be used to specify parameters; see the backend section of the using MLTE page for details. The default artifact store puts artifacts into a non-persistent, in-memory store. For example, running the backend with a store located in a folder called store relative to the folder where you are running MLTE would use the following command:
$ mlte backend --store-uri fs://storeOnce the backend is running, you can run the UI with the following command:
$ mlte uiAfter this, go to the hosted address (defaults to http://localhost:8000) to view the MLTE UI homepage. You will need to log in to access the functionality in the UI, which you can do by using the default user. You can later use the UI to set up new users as well.
NOTE: you should change the default user's password as soon as you can, if you are not on a local setup.
- Default user: admin
- Default password: admin1234
If you want to build MLTE from its source code in this repository, see the development section of the MLTE docs for details on setting up a local environment to build and run MLTE from source.
To build the MLTE wheel from source in an isolated Docker environment, without setting up a local environment (the output will be in the ./dist folder), run the following command:
$ make build-in-dockerThe MLTE Python package is best used in conjunction with the MLTE process framework. For more details on using the package, see our documentation page on using MLTE.
All information relating to development of MLTE from demo information, development setup, development guidelines, and QA/CI information can be found here.
If you're interested in learning more about this work, you can read our paper. While not required, it is highly encouraged and greatly appreciated if you cite our paper when you use MLTE for academic research.
@INPROCEEDINGS{10173876,
author={Maffey, Katherine R. and Dotterrer, Kyle and Niemann, Jennifer and Cruickshank, Iain and Lewis, Grace A. and Kästner, Christian},
booktitle={2023 IEEE/ACM 45th International Conference on Software Engineering: New Ideas and Emerging Results (ICSE-NIER)},
title={MLTEing Models: Negotiating, Evaluating, and Documenting Model and System Qualities},
year={2023},
volume={},
number={},
pages={31-36},
keywords={Measurement;Machine learning;Production;Organizations;Software;Stakeholders;Software engineering;machine learning;test and evaluation;machine learning evaluation;responsible AI},
doi={10.1109/ICSE-NIER58687.2023.00012}
}