Bodywork deploys machine learning projects developed in Python, to Kubernetes. It helps you:
- serve models as microservices
- execute batch jobs
- run reproducible pipelines
On demand, or on a schedule. It automates repetitive DevOps tasks and frees machine learning engineers to focus on what they do best - solving data problems with machine learning.
Where does Bodywork Fit?
Bodywork is aimed at teams who want to deploy machine learning projects in containers. It will deliver your project's Python modules directly from your Git repository into Docker containers and manage their deployment to a Kubernetes cluster.
Where do I Install Bodywork?
Bodywork is distributed as a Python package that exposes a command line interface for configuring Kubernetes to run Bodywork deployments. It takes just one command to schedule a pipeline hosted on GitHub to run every evening,
What does Bodywork Do?
When triggered, Bodywork clones your project's Git repository, analyses the configuration provided in a
bodywork.yaml file, and then manages the deployment of the projects' stages - creating new Bodywork containers to run the Python modules that define each one. At no point is there any need to build Docker images, push them to a container registry or to configure Kubernetes directly.
This process is shown below for a
train-and-serve pipeline with two stages: train model (as a batch job), then serve the trained model (as a microservice with a REST API).
What will I need to Do?
Divide your project into discrete stages and create an executable Python module for each one. Bundle these files together with a
bodywork.yaml configuration file, into a Git repository and you're ready to go.
You do not need to tie yourself to new APIs - just add
bodywork.yaml to your existing codebase and watch as Bodywork pulls each stage into its own container and deploys to Kubernetes.
CI/CD for Machine Learning
Because Bodywork can run deployments on a schedule, every time cloning the latest version of your codebase in the target branch, this system naturally forms an end-to-end CI/CD platform for your machine learning project, as illustrated below.
This is the GitOps pattern for cloud native continuous delivery.
- continuous deployment - batch jobs, model-scoring services as well as complex ML pipelines, using pre-built Bodywork containers to orchestrate end-to-end machine learning workflows.
- resilience - Bodywork handles automatic retires for batch jobs and automatic roll-backs for service deployments, without any downtime.
- horizontal scaling - Bodywork can back your service endpoints with as many container replicas as you need to handle your API traffic volumes.
- no APIs to learn - Bodywork does not require you to re-write your machine learning projects to conform to our view of how your codebase should be engineered. All you need to do is provide executable Python modules for starting service applications and running batch jobs.
- multi-cloud - Bodywork deploys to Kubernetes clusters, which are available as managed services from all major cloud providers. Kubernetes is indifferent to where it is running, so changing cloud provider is as easy as pointing to a different cluster.
- written in Python - the native language of machine learning and data science, so your team can have full visibility of what Bodywork is doing and how.
- open-source - Bodywork is built and maintained by machine learning engineers, for machine learning engineers, who are committed to keeping it 100% open-source.
Bodywork brings DevOps to your machine learning projects and will form the basis of your Machine Learning Operations (MLOps) platform. It will ensure that your projects are always trained with the latest data, the most recent models are always deployed and your machine learning systems remain highly-available.