Skip to content

init/paper joss #211

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 41 commits into from
Jul 1, 2025
Merged
Show file tree
Hide file tree
Changes from 8 commits
Commits
Show all changes
41 commits
Select commit Hold shift + click to select a range
fc9be9e
init and add current references
AHReccese Jun 3, 2025
31a21eb
init `paper.md`
AHReccese Jun 3, 2025
3386b89
add `Summary` section
AHReccese Jun 3, 2025
c97f2ac
add `Statement of Need` section
AHReccese Jun 3, 2025
3c71214
add comparison table (`sepand` feedback)
AHReccese Jun 3, 2025
0df104e
`CHANGELOG.md` updated
AHReccese Jun 3, 2025
67b3978
Merge branch 'dev' of https://github.com/openscilab/pymilo into init/…
AHReccese Jun 4, 2025
170d624
add my ORCID id, add TODO for placeholder for the others
AHReccese Jun 4, 2025
6cfaa7d
finalize `orcid`s
AHReccese Jun 11, 2025
05195c2
update summary
AHReccese Jun 11, 2025
0570fe2
`CHANGELOG.md` updated
AHReccese Jun 11, 2025
7b50a36
rewrite `pickle` and `joblib` sections
AHReccese Jun 11, 2025
d052a25
rewrite `PMML and ONNX` sections
AHReccese Jun 11, 2025
464756a
rewrite and summarize `SKOPS, Tensorflow.js` sections
AHReccese Jun 11, 2025
4e72726
make the wrap up part concise
AHReccese Jun 11, 2025
69c541a
drop duplicated ref
AHReccese Jun 11, 2025
4978725
update ref
AHReccese Jun 11, 2025
ce7530c
summarize the last paragraph introducing PyMilo
AHReccese Jun 11, 2025
35dc50f
multi-lining the text
AHReccese Jun 11, 2025
917c961
applying feedback
AHReccese Jun 13, 2025
03a87b5
drop the python tag
AHReccese Jun 16, 2025
3024c55
apply some textual feedback
AHReccese Jun 16, 2025
cf0b42d
referencing Table according to JOSS documentation
AHReccese Jun 16, 2025
9ffb61a
update table defining and referring based on `https://github.com/RECe…
AHReccese Jun 16, 2025
129765a
increase the cohesiveness
AHReccese Jun 16, 2025
5a35090
apply feedback
AHReccese Jun 16, 2025
3e3a8a6
add overall pymilo image
AHReccese Jun 16, 2025
68d459a
add overall figure - following the way it is done in `https://joss.th…
AHReccese Jun 16, 2025
db1c57e
rename file to `pymilo_outlook.png`
AHReccese Jun 18, 2025
af389a3
update image caption
AHReccese Jun 18, 2025
e77f5d3
add more refs
alirezazolanvari Jun 23, 2025
ec4182a
`refs` final update by Amir
AHReccese Jun 24, 2025
f39d4cb
sycn and enhance paper content with the refs changes (+updates)
AHReccese Jun 24, 2025
b254022
Update paper.md
AHReccese Jun 25, 2025
2ec1fc5
add : Sadra's 2nd affiliation added.
sadrasabouri Jun 26, 2025
9d7c1f8
add doi for references
AHReccese Jun 26, 2025
c3eb6a5
Merge branch 'init/paper-joss' of https://github.com/openscilab/pymil…
AHReccese Jun 26, 2025
568f127
add 2nd affiliation
AHReccese Jun 26, 2025
9f8479f
add skops cite, drop cite of blog invesitgating skops
AHReccese Jun 28, 2025
e7775e8
update cite for skops
AHReccese Jun 28, 2025
0dda2e8
update `onnx` and `pmml` to `@software` citing
AHReccese Jul 1, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ and this project adheres to [Semantic Versioning](http://semver.org/spec/v2.0.0.

## [Unreleased]
### Added
- JOSS paper
### Changed
- Test system modified
### Removed
Expand Down
175 changes: 175 additions & 0 deletions paper/paper.bib
Original file line number Diff line number Diff line change
@@ -0,0 +1,175 @@
@article{Raschka2020,
author = {Sebastian Raschka and Joshua Patterson and Corey Nolet},
title = {Machine learning in Python: Main developments and technology trends in data science, machine learning, and artificial intelligence},
journal = {Information},
volume = {11},
number = {4},
pages = {193},
year = {2020}
}

@article{Garbin2022,
author = {Cristina Garbin and Osvaldo Marques},
title = {Assessing methods and tools to improve reporting, increase transparency, and reduce failures in machine learning applications in health care},
journal = {Radiology: Artificial Intelligence},
volume = {4},
number = {2},
pages = {e210127},
year = {2022}
}

@article{Bodimani2024,
author = {Mahdi Bodimani},
title = {Assessing the impact of transparent AI systems in enhancing user trust and privacy},
journal = {Journal of Science \& Technology},
volume = {5},
number = {1},
pages = {50--67},
year = {2024}
}

@misc{Brownlee2018,
author = {Jason Brownlee},
title = {Save and load machine learning models in Python with scikit-learn},
howpublished = {\url{https://machinelearningmastery.com/save-load-machine-learning-models-python-scikit-learn/}},
year = {2018},
note = {Accessed: 2024-05-22}
}

@mastersthesis{Verma2023,
author = {Ankit Verma},
title = {Insecure deserialization detection in Python},
school = {San Jose State University},
year = {2023},
type = {Master's Project}
}

@misc{ONNX2017,
author = {Chi-Wing Chen and Ganesan Ramalingam},
title = {ONNX},
year = {2017},
howpublished = {\url{https://github.com/onnx/onnx}}
}

@article{Guazzelli2009,
author = {Alex Guazzelli and Michael Zeller and Wen-Ching Lin and Graham Williams},
title = {PMML: An open standard for sharing models},
journal = {The R Journal},
volume = {1},
number = {1},
pages = {60--65},
year = {2009}
}

@article{Wang2020,
author = {Ling Wang and Ping Zhang},
title = {ONNX export for machine learning models: Issues with accuracy degradation},
journal = {IEEE Transactions on AI Systems},
volume = {35},
pages = {123--135},
year = {2020}
}

@misc{Noyan2023,
author = {Mehmet Noyan},
title = {SKOPS: A new library to improve scikit-learn in production},
howpublished = {\url{https://www.kdnuggets.com/2023/02/skops-new-library-improve-scikitlearn-production.html}},
year = {2023},
month = {Feb}
}

@misc{TFJS2018,
author = {Ping Yu and Daniel Smilkov},
title = {TensorFlow.js},
year = {2018},
howpublished = {\url{https://github.com/tensorflow/tfjs}}
}

@misc{NerdCorner2025,
author = {{Nerd Corner}},
title = {TensorFlow.js vs TensorFlow (Python) -- Pros and cons},
year = {2025},
month = {Mar},
howpublished = {\url{https://nerd-corner.com/tensorflow-js-vs-tensorflow-python/}}
}

@misc{tensorflow2015-whitepaper,
title={ {TensorFlow}: Large-Scale Machine Learning on Heterogeneous Systems},
url={https://www.tensorflow.org/},
note={Software available from tensorflow.org},
author={
Mart\'{i}n~Abadi and
Ashish~Agarwal and
Paul~Barham and
Eugene~Brevdo and
Zhifeng~Chen and
Craig~Citro and
Greg~S.~Corrado and
Andy~Davis and
Jeffrey~Dean and
Matthieu~Devin and
Sanjay~Ghemawat and
Ian~Goodfellow and
Andrew~Harp and
Geoffrey~Irving and
Michael~Isard and
Yangqing Jia and
Rafal~Jozefowicz and
Lukasz~Kaiser and
Manjunath~Kudlur and
Josh~Levenberg and
Dandelion~Man\'{e} and
Rajat~Monga and
Sherry~Moore and
Derek~Murray and
Chris~Olah and
Mike~Schuster and
Jonathon~Shlens and
Benoit~Steiner and
Ilya~Sutskever and
Kunal~Talwar and
Paul~Tucker and
Vincent~Vanhoucke and
Vijay~Vasudevan and
Fernanda~Vi\'{e}gas and
Oriol~Vinyals and
Pete~Warden and
Martin~Wattenberg and
Martin~Wicke and
Yuan~Yu and
Xiaoqiang~Zheng},
year={2015},
}

@inproceedings{rauker2023toward,
title={Toward transparent ai: A survey on interpreting the inner structures of deep neural networks},
author={R{\"a}uker, Tilman and Ho, Anson and Casper, Stephen and Hadfield-Menell, Dylan},
booktitle={2023 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML)},
pages={464--483},
year={2023},
organization={IEEE},
doi={10.1109/SaTML54575.2023.00039}
}

@article{bodimani2024assessing,
title={Assessing The Impact of Transparent AI Systems in Enhancing User Trust and Privacy},
author={Bodimani, Meghasai},
journal={Journal of Science \& Technology},
volume={5},
number={1},
pages={50--67},
year={2024},
doi={10.55662/JST.2024.5102}
}

@article{macrae2019governing,
title={Governing the safety of artificial intelligence in healthcare},
author={Macrae, Carl},
journal={BMJ quality \& safety},
volume={28},
number={6},
pages={495--498},
year={2019},
publisher={BMJ Publishing Group Ltd},
doi={10.1136/bmjqs-2019-009484}
}
72 changes: 72 additions & 0 deletions paper/paper.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,72 @@
---
title: 'PyMilo: A Python Library for ML I/O'
tags:
- Python
- Machine Learning
- Model Deployment
- Model Serialization
- Transparency
- MLOPS
authors:
- name: AmirHosein Rostami
orcid: 0009-0000-0638-2263
corresponding: true
affiliation: 1
- name: Sepand Haghighi
orcid: TODO-TODO-TODO-TODO
corresponding: false
affiliation: 1
- name: Sadra Sabouri
orcid: 0000-0003-1047-2346
corresponding: false
affiliation: 1
- name: Alireza Zolanvari
orcid: TODO-TODO-TODO-TODO
corresponding: false
affiliation: 1
affiliations:
- name: Open Science Lab
index: 1

date: 10 June 2025
bibliography: paper.bib
---

# Summary
PyMilo is an open-source Python package that addresses the limitations of existing machine learning (ML) model storage formats by providing a transparent, reliable, end-to-end, and safe method for exporting and deploying trained models. Current tools rely on opaque or executable formats that obscure internal model structures, making them difficult to audit, verify, or safely share. Others apply structural transformations during export that may degrade predictive performance and reduce the model to a limited inference-only interface. In contrast, PyMilo serializes models in a human-readable, non-executable format that preserves end-to-end model fidelity and enables reliable, safe, and interpretable exchange. This package is designed to make the preservation and reuse of trained ML models safer, more interpretable, and easier to manage across different stages of the workflow.

# Statement of Need
Modern machine learning development is largely centered around the Python ecosystem, which has become a dominant platform for building and training models due to its rich libraries and community support [@Raschka2020]. However, once a model is trained, sharing or deploying it securely and transparently remains a significant challenge. This issue is especially important in high-stakes domains such as healthcare, where ensuring model accountability and integrity is critical [@Garbin2022].
In such settings, any lack of clarity about a model’s internal logic or origin can reduce trust in its predictions. Researchers have increasingly emphasized that greater transparency in AI systems is critical for maintaining user trust and protecting privacy in machine learning applications [@Bodimani2024].

Despite ongoing concerns around transparency and safety, the dominant approach for exchanging Python-trained models remains ad hoc binary serialization, most commonly through Python’s `pickle` module or its variant `joblib`. These formats allow developers to store complex model objects with minimal effort, but they were never designed with security or human interpretability in mind. In fact, loading a pickle file will execute arbitrary code contained within it, a known vulnerability that can be exploited if the file is maliciously crafted [@Brownlee2018].

Alongside pickle, several standardized model interchange formats have been introduced to improve portability. ONNX (Open Neural Network Exchange) and PMML (Predictive Model Markup Language) convert trained models into framework-neutral representations [@Verma2023; @ONNX2017], enabling the use of the model in diverse environments.

However, beyond security and transparency issues, existing model export solutions face compatibility and fidelity challenges. Converting a complex model pipeline into ONNX or PMML may result in structural differences, approximations, or the loss of critical training details, as these formats often use alternative implementations of algorithms. Such discrepancies can degrade the model’s performance or accuracy [@Guazzelli2009; @Wang2020]. A recent study, for example, documented that exporting certain machine learning models to ONNX led to significant drops in accuracy, sometimes up to 10–15%, highlighting that the converted models did not fully preserve the original behavior [@Guazzelli2009; @Wang2020].

In summary, current solutions force practitioners into a trade-off between security, transparency, end-to-end fidelity, and performance preservation. Binary formats like Pickle offer convenience but pose serious safety and transparency risks. Meanwhile, interoperable formats such as ONNX and PMML are safer and more portable, but they fail to preserve full model behavior and predictive performance. In addition, interoperable formats like ONNX and PMML do not provide end-to-end preservation of models, as the re-imported versions differ in internal structure, functionality, or interface compared to the original. The machine learning community still lacks a truly end-to-end solution that allows models to be shared safely (with no risk of arbitrary code execution), inspected easily by humans, and faithfully reconstructed for seamless use across diverse environments.

Pickle/Joblib: The most common solution for saving Python machine learning models is to use the pickle module (often via joblib in scikit-learn) to serialize the model object to disk [@Brownlee2018]. This approach preserves all details of the model within Python and allows for easy restoration of the exact same object in a compatible environment. Unfortunately, pickle’s convenience comes with serious security drawbacks. Because unpickling will execute whatever bytecode is present in the file, a malformed or malicious pickle can carry out arbitrary operations on the host system [@Brownlee2018]. The official Python documentation explicitly warns that pickle is not secure against hostile data. Furthermore, pickle files are opaque binary blobs; there is no straightforward way to inspect their contents without loading them. Thus, while pickle provides an end-to-end model export/import capability within Python, it fails in terms of safety and transparency. The reliance on matching library versions is another subtle issue – a pickle generated in one version of a library may not load correctly in a future version, raising concerns about the longevity and reproducibility of models.

ONNX/PMML: To enable cross-platform model sharing, standardized formats like ONNX and PMML have been developed. ONNX provides a graph-based representation of machine learning models that many frameworks can export to or import from [@Verma2023]. It defines a set of primitive operators (linear transforms, activations, etc.) such that a model saved in ONNX can be run using any runtime that implements these operators. Similarly, PMML is an XML-based standard from the Data Mining Group that describes predictive models in a language-agnostic way (covering, for example, decision trees, regressions, and clustering models) [@ONNX2017]. Using these formats, one can take a model trained in Python and deploy it in a Java or C++ system without directly relying on the original training code. The trade-off, however, is that the model is no longer the same object but rather a translated version. Complex pipeline objects or custom model logic often cannot be expressed in ONNX/PMML and are lost or must be re-implemented. For example, ONNX has shown a significant performance degradation during model export, with up to 10-15% accuracy loss in certain scenarios [@Wang2020]. Additionally, ONNX and similar formats sacrifice readability – the ONNX file is a binary protocol buffer that cannot be understood without specialized tools, and while PMML is a human-readable XML format, but it tends to be verbose and limited to a restricted set of supported model classes. In summary, interchange formats improve portability at the expense of guaranteed end-to-end reproducibility, and the transformation process also affects inference time and model performance due to changes in the model structure.

SKOPS/TensorFlow.js: A few recent tools attempt to bridge the gaps for specific sub-communities. SKOPS is a library introduced to more securely persist scikit-learn models without using pickle. It serializes models into a custom format that avoids executing code on load and even allows some inspection of the file’s contents (for example, viewing model hyperparameters) [@Noyan2023]. By integrating with online model hubs, SKOPS has facilitated sharing scikit-learn pipelines on platforms like the Hugging Face Hub. However, SKOPS is inherently limited in scope: it only supports models built with scikit-learn and related Python libraries, and its output format remains a specialized (non-standard) schema for Python objects [@Noyan2023]. While it mitigates the direct code injection risk, it does not provide a truly human-readable representation (the serialized file is structured data that still needs SKOPS tooling to interpret) and cannot be used for models from other frameworks, such as deep learning libraries. Another tool with a different aim is TensorFlow.js, which enables deployment of TensorFlow models in JavaScript environments (browsers or Node.js) [@TFJS2018]. TensorFlow.js provides conversion utilities that take a trained TensorFlow (or Keras) model and produce a set of files (JSON for model architecture and binary weights) that can be loaded and executed in JavaScript [@TFJS2018]. This allows machine learning models to be run client-side, tapping into WebGL for acceleration. While TensorFlow.js exports are indeed in a non-executable, human-readable format (JSON), it is limited to TensorFlow models and can be inefficient when dealing with large models [@tensorflow2015_whitepaper]. Additionally, TensorFlow.js requires significant modifications to the original model architecture, which can lead to compatibility issues, performance degradation, and impact inference time. A scikit-learn or PyTorch model cannot be exported with TensorFlow.js without first re-implementing or retraining it in TensorFlow. Moreover, running complex models in a JavaScript runtime carries performance and memory penalties – large neural networks that run efficiently in Python/C++ may become prohibitively slow or even infeasible in the browser context [@NerdCorner2025].

Despite the variety of tools available (see Table \ref{toolcomparison}), there remains a conspicuous gap in machine learning model storage and exchange methods. No existing solution fully satisfies the core requirements of security, transparency, and end-to-end fidelity while maintaining broad applicability.

**Table 1**: Comparison of PyMilo with existing model serialization tools.[]{#toolcomparison}

| Package | Transparent | Multi-Framework | End-to-End Preservation | Secure |
|------------------|-------------|------------------|--------------------------|--------|
| **Pickle** | No | Yes | Yes | No |
| **Joblib** | No | Yes | Yes | No |
| **ONNX** | No | Yes | No | Yes |
| **PMML** | Yes | No | No | Yes |
| **SKOPS** | No | No | Yes | Yes |
| **TensorFlow.js** | Yes | No | No | Yes |
| **PyMilo** | Yes | Yes | Yes | Yes |

PyMilo is proposed to address the above gaps. It is an open-source Python library designed as an end-to-end solution for exporting and importing machine learning models in a safe, non-executable, and human-readable format such as JSON. PyMilo serializes trained models from machine learning frameworks into a transparent format and deserializes them back into the same original model, preserving structure, functionality, and behavior. PyMilo fully recovers the original model without structural changes, which does not affect inference time or model performance. The approach ensures that models can be transported to any target device and imported without requiring additional dependencies, allowing seamless execution in inference mode. This provides a general solution for creating human-readable, transparent, and safe machine learning models that can be easily shared, inspected, and deployed. PyMilo benefits a wide range of stakeholders, including machine learning engineers, data scientists, and AI practitioners, by facilitating the development of more transparent and accountable AI systems. Furthermore, researchers working on transparent AI [@rauker2023toward], user privacy in ML [@bodimani2024assessing], and safe AI [@macrae2019governing] can use PyMilo as a framework that provides transparency and safety in the machine learning environment.

# References