MLOps company Iterative has announced a new open source compute orchestration tool using Terraform, a solution by HashiCorp. According to the company, Terraform Provider Iterative (TPI) is the first product on HashiCorp’s Terraform technology stack to simplify ML training on any cloud while helping infrastructure and ML teams to save significant time and money in maintaining and configuring their training resources.
Built on Terraform by HashiCorp, an open-source infrastructure as code software tool that provides a consistent CLI workflow to manage hundreds of cloud services, TPI allows data scientists to deploy workloads without having to figure out the infrastructure.
TPI complements Terraform with additional functionality, customized for machine learning use cases. It automatically provisions and de-provisions compute resources once an experiment is finished running, helping to reduce costs by up to 90%. Also, ML teams can use spot instances to train experiments without worrying about losing all their progress if a spot instance terminates. TPI automatically migrates training jobs to a new spot instance when the existing instance terminates so that the workload can pick up where it left off.
TPI delivers a tool that lets both data science and software development teams collaborate using the same language and tool. This simplifies compute management and allows for ML models to be delivered into production faster.
With TPI, data scientists only need to configure the resources they need once and are able to deploy anywhere and everywhere in minutes. Once it is configured as part of an ML model experiment pipeline, users can deploy on AWS, GCP, Azure, on-prem, or with Kubernetes.