Deliver AI/ML-as-a-Service
Automate Your Complex AI/ML Applications Using Kubernetes

Self Service AI/ML & Big Data on Kubernetes

Cloud-native AI/ML for Your Enterprise

The Robin Cloud Native Platform extends Kubernetes with application automation, built-in storage, and networking, to deliver a production-ready solution for AI/ML. Robin automates the provisioning and management of AI/ML clusters so you can deliver an “as-a-service” experience with one-click simplicity to data engineers, data scientists, and developers.

FreedomChoice
TensorFlow
Pytorch
KubeFlow
Spark
Caffe2
Jupyter
Apache-Zappelin
MLFlow
XGBoost
R-Studio

Use cases

Separation of compute and storage

Disaggregate compute and storage, the cloud-native way

Enjoy cloud-like cost efficiencies by disaggregating compute and storage. Scale independently so you can store massive data on inexpensive hardware, while only utilizing compute when needed. For high performance, turn on data locality with one click.

Improved hardware utilization

Get more out of your hardware investments

Consolidate multiple AI/ML workloads and improve hardware utilization without compromising SLAs or QoS. The Robin platform provides multi-tenancy with RBAC which enables sharing resources between clusters. For example, if a cluster runs the majority of its batch jobs during the night, it can borrow resources from an adjacent application cluster with daytime peaks, and vice versa.

Self-service experience

Self-service provisioning in an app-store experience

Create a private app-store that your developers can use to provision AI/ML clusters, eliminating IT tickets and procedural delays. Robin’s application automation platform simplifies the deployment of complex, AI/ML applications on Kubernetes. Simply log in to an app-store experience, click on the desired application, and allocate the compute and storage resources. Your application will be deployed and ready to use within minutes.

Resources