Skip to content

openscilab/pymilo



Codecov PyPI version built with Python3 GitHub repo size Discord Channel

Overview

PyMilo is an open source Python package that provides a simple, efficient, and safe way for users to export pre-trained machine learning models in a transparent way. By this, the exported model can be used in other environments, transferred across different platforms, and shared with others. PyMilo allows the users to export the models that are trained using popular Python libraries like scikit-learn, and then use them in deployment environments, or share them without exposing the underlying code or dependencies. The transparency of the exported models ensures reliability and safety for the end users, as it eliminates the risks of binary or pickle formats.

PyPI Counter PyPI Downloads
Github Stars
Branch main dev
CI
Code Quality CodeFactor codebeat badge

Installation

PyPI

Source code

Conda

Usage

Import/Export

Imagine you want to train a LinearRegression model representing this equation: $y = x_0 + 2x_1 + 3$. You will create data points (X, y) and train your model as follows.

import numpy as np
from sklearn.linear_model import LinearRegression
X = np.array([[1, 1], [1, 2], [2, 2], [2, 3]])
y = np.dot(X, np.array([1, 2])) + 3
# y = 1 * x_0 + 2 * x_1 + 3
model = LinearRegression().fit(X, y)
pred = model.predict(np.array([[3, 5]]))
# pred = [16.] (=1 * 3 + 2 * 5 + 3)

Using PyMilo Export class you can easily serialize and export your trained model into a JSON file.

from pymilo import Export
Export(model).save("model.json")

Export

The Export class facilitates exporting of machine learning models to JSON files.

Parameter Description
model The machine learning model to be exported
Property Description
data The serialized model data including all learned parameters
version The scikit-learn version used to train the model
type The type/class name of the exported model
Method Description
save Save the exported model to a JSON file
to_json Return the model as a JSON string representation
batch_export Export multiple models to individual JSON files in a directory

You can check out your model as a JSON file now.

{
    "data": {
        "fit_intercept": true,
        "copy_X": true,
        "n_jobs": null,
        "positive": false,
        "n_features_in_": 2,
        "coef_": {
            "pymiloed-ndarray-list": [
                1.0000000000000002,
                1.9999999999999991
            ],
            "pymiloed-ndarray-dtype": "float64",
            "pymiloed-ndarray-shape": [
                2
            ],
            "pymiloed-data-structure": "numpy.ndarray"
        },
        "rank_": 2,
        "singular_": {
            "pymiloed-ndarray-list": [
                1.618033988749895,
                0.6180339887498948
            ],
            "pymiloed-ndarray-dtype": "float64",
            "pymiloed-ndarray-shape": [
                2
            ],
            "pymiloed-data-structure": "numpy.ndarray"
        },
        "intercept_": {
            "value": 3.0000000000000018,
            "np-type": "numpy.float64"
        }
    },
    "sklearn_version": "1.4.2",
    "pymilo_version": "0.8",
    "model_type": "LinearRegression"
}

You can see all the learned parameters of the model in this file and change them if you want. This JSON representation is a transparent version of your model.

Now let's load it back. You can do it easily by using PyMilo Import class.

from pymilo import Import
model = Import("model.json").to_model()
pred = model.predict(np.array([[3, 5]]))
# pred = [16.] (=1 * 3 + 2 * 5 + 3)

Import

The Import class facilitates importing of serialized models from JSON files, JSON strings, or URLs.

Parameter Description
file_adr Path to the JSON file containing the serialized model
json_dump JSON string representation of the serialized model
url URL to download the serialized model from
Property Description
data The deserialized model data
version The scikit-learn version of the original model
type The type/class name of the imported model
Method Description
to_model Convert the imported data back to a scikit-learn model
batch_import Import multiple models from JSON files in a directory

This loaded model is exactly the same as the original trained model.

ML streaming

You can easily serve your ML model from a remote server using ML streaming feature of PyMilo.

⚠️ ML streaming feature exists in versions >=1.0

⚠️ In order to use ML streaming feature, make sure you've installed the streaming mode of PyMilo

⚠️ The ML streaming feature is under construction and is not yet considered stable.

You can choose either REST or WebSocket as the communication medium protocol.

Server

Let's assume you are in the remote server and you want to import the exported JSON file and start serving your model through REST protocol!

from pymilo import Import
from pymilo.streaming import PymiloServer, CommunicationProtocol
my_model = Import("model.json").to_model()
communicator = PymiloServer(
    model=my_model,
    port=8000,
    communication_protocol=CommunicationProtocol["REST"],
    ).communicator
communicator.run()

PymiloServer

The PymiloServer class facilitates streaming machine learning models over a network.

Parameter Description
port Port number for the server to listen on (default: 8000)
host Host address for the server (default: "127.0.0.1")
compressor Compression method from Compression enum
communication_protocol Communication protocol from CommunicationProtocol enum

The compressor parameter accepts values from the Compression enum including NULL (no compression), GZIP, ZLIB, LZMA, or BZ2. The communication_protocol parameter accepts values from the CommunicationProtocol enum including REST or WEBSOCKET.

Method Description
init_client Initialize a new client with the given client ID
remove_client Remove an existing client by client ID
init_ml_model Initialize a new ML model for a given client
set_ml_model Set or update the ML model for a client
remove_ml_model Remove an existing ML model for a client
get_ml_models Get all ML model IDs for a client
execute_model Execute model methods or access attributes
grant_access Allow a client to access another client's model
revoke_access Revoke access to a client's model
get_allowed_models Get models a client is allowed to access

Now PymiloServer runs on port 8000 and exposes REST API to upload, download and retrieve attributes either data attributes like model._coef or method attributes like model.predict(x_test).

ℹ️ By default, PymiloServer listens on the loopback interface (127.0.0.1). To make it accessible over a local network (LAN), specify your machine’s LAN IP address in the host parameter of the PymiloServer constructor.

Client

By using PymiloClient you can easily connect to the remote PymiloServer and execute any functionalities that the given ML model has, let's say you want to run predict function on your remote ML model and get the result:

from pymilo.streaming import PymiloClient, CommunicationProtocol
pymilo_client = PymiloClient(
    mode=PymiloClient.Mode.LOCAL,
    server_url="SERVER_URL",
    communication_protocol=CommunicationProtocol["REST"],
    )
pymilo_client.toggle_mode(PymiloClient.Mode.DELEGATE)
result = pymilo_client.predict(x_test)

PymiloClient

The PymiloClient class facilitates working with remote PyMilo servers.

Parameter Description
model The local ML model to wrap around
mode Operating mode (LOCAL or DELEGATE)
compressor Compression method from Compression enum
server_url URL of the PyMilo server
communication_protocol Communication protocol from CommunicationProtocol enum

The mode parameter accepts two values LOCAL to execute operations on the local model, or DELEGATE to delegate operations to the remote server. The compressor parameter accepts values from the Compression enum including NULL (no compression), GZIP, ZLIB, LZMA, or BZ2. The communication_protocol parameter accepts values from the CommunicationProtocol enum including REST or WEBSOCKET.

Method Description
toggle_mode Switch between LOCAL and DELEGATE modes
register Register the client with the remote server
deregister Deregister the client from the server
register_ml_model Register an ML model with the server
deregister_ml_model Deregister an ML model from the server
upload Upload the local model to the remote server
download Download the remote model to local
get_ml_models Get all registered ML models for this client
grant_access Grant access to this client's model to another client
revoke_access Revoke access previously granted to another client
get_allowance Get clients who have access to this client's models
get_allowed_models Get models this client is allowed to access from another client

ℹ️ If you've deployed PymiloServer locally (on port 8000 for instance), then SERVER_URL would be http://127.0.0.1:8000 or ws://127.0.0.1:8000 based on the selected protocol for the communication medium.

You can also download the remote ML model into your local and execute functions locally on your model.

Calling download function on PymiloClient will sync the local model that PymiloClient wraps upon with the remote ML model, and it doesn't save model directly to a file.

pymilo_client.download()

If you want to save the ML model to a file in your local, you can use Export class.

from pymilo import Export
Export(pymilo_client.model).save("model.json")

Now that you've synced the remote model with your local model, you can run functions.

pymilo_client.toggle_mode(mode=PymiloClient.Mode.LOCAL)
result = pymilo_client.predict(x_test)

PymiloClient wraps around the ML model, either to the local ML model or the remote ML model, and you can work with PymiloClient in the exact same way that you did with the ML model, you can run exact same functions with same signature.

ℹ️ Through the usage of toggle_mode function you can specify whether PymiloClient applies requests on the local ML model pymilo_client.toggle_mode(mode=Mode.LOCAL) or delegates it to the remote server pymilo_client.toggle_mode(mode=Mode.DELEGATE)

Supported ML models

scikit-learn PyTorch
Linear Models ✅ -
Neural Networks ✅ -
Trees ✅ -
Clustering ✅ -
Naïve Bayes ✅ -
Support Vector Machines (SVMs) ✅ -
Nearest Neighbors ✅ -
Ensemble Models ✅ -
Pipeline Model ✅ -
Preprocessing Models ✅ -
Cross Decomposition Models ✅ -
Feature Extractor Models ✅ -

Details are available in Supported Models.

Issues & bug reports

Just fill an issue and describe it. We'll check it ASAP! or send an email to pymilo@openscilab.com.

  • Please complete the issue template

You can also join our discord server

Discord Channel

Contributing

We welcome contributions! Please read our Contributing Guidelines before submitting any changes.

Acknowledgments

Python Software Foundation (PSF) grants PyMilo library partially for versions 1.0, 1.1. PSF is the organization behind Python. Their mission is to promote, protect, and advance the Python programming language and to support and facilitate the growth of a diverse and international community of Python programmers.

Python Software Foundation

Trelis Research grants PyMilo library partially for version 1.0. Trelis Research provides tools and tutorials for businesses and developers looking to fine-tune and deploy large language models.

Trelis Research

Cite

If you use PyMilo in your research, we would appreciate citations to the following paper :

Rostami, A., Haghighi, S., Sabouri, S., & Zolanvari, A. (2024). PyMilo: A Python Library for ML I/O. arXiv e-prints, arXiv-2501.

Zenodo DOI 10.5281/zenodo.14713078
@article{rostami2024pymilo,
  title={PyMilo: A Python Library for ML I/O},
  author={Rostami, AmirHosein and Haghighi, Sepand and Sabouri, Sadra and Zolanvari, Alireza},
  journal={arXiv e-prints},
  pages={arXiv--2501},
  year={2024}
}

Show your support

Star this repo

Give a ⭐️ if this project helped you!

Donate to our project

If you do like our project and we hope that you do, can you please support us? Our project is not and is never going to be working for profit. We need the money just so we can continue doing what we do ;-) .

PyMilo Donation