Our website publishes news, press releases, opinion and advertorials on various financial organizations, products and services which are commissioned from various Companies, Organizations, PR agencies, Bloggers etc. These commissioned articles are commercial in nature. This is not to be considered as financial advice and should be considered only for information purposes. It does not reflect the views or opinion of our website and is not to be considered an endorsement or a recommendation. We cannot guarantee the accuracy or applicability of any information provided with respect to your individual or personal circumstances. Please seek Professional advice from a qualified professional before making any financial decisions. We link to various third-party websites, affiliate sales networks, and to our advertising partners websites. When you view or click on certain links available on our articles, our partners may compensate us for displaying the content to you or make a purchase or fill a form. This will not incur any additional charges to you. To make things simpler for you to identity or distinguish advertised or sponsored articles or links, you may consider all articles or links hosted on our site as a commercial article placement. We will not be responsible for any loss you may suffer as a result of any omission or inaccuracy on the website.
Home Technology Charmed Kubeflow 1.6 is now available from Canonical

Charmed Kubeflow 1.6 is now available from Canonical

by uma

 

The latest release of Canonical’s end-to-end MLOps platform brings advanced AI/ML training capabilities

8 September 2022- Canonical, the publisher of Ubuntu, announces today the release of Charmed Kubeflow 1.6, an end-to-end MLOps platform with optimised complex model training capabilities.  Charmed Kubeflow is Canonical’s enterprise-ready distribution of Kubeflow, an open-source machine learning toolkit designed for use with Kubernetes.

Charmed Kubeflow 1.6 follows the same release cadence as the Kubeflow upstream project. This latest version follows the Kubeflow roadmap and comes with performance enhancements and more advanced model training capabilities.  

Speed up digital transformation: take models from concept to production 

Charmed Kubeflow helps data science teams automate tasks and boost productivity, helping companies lower costs. The platform’s components use charms – Kubernetes operators that automate maintenance and security operations. Charms accelerate workload deployment, allowing data scientists to take models to market more efficiently.

According to a 2022 IBM Index AI Report, adoption of AI/ML within enterprises reached 35% last year. The benefits are clear. For instance, banks applying ML to build recommendation earnings have boosted sales by 10%, saved 20% in CAPEX, increased cash collections by 20%, and reduced churn by 20%.  Identifying tools that bring automation to data science workflows has become essential to getting a faster return on investment from AI/ML projects.

Improved data processing and tracking with MLOps

Besides accelerating deployments, Charmed Kubeflow 1.6 makes data processing more seamless.  Kubeflow’s 2022 survey reveals that data processing and transformation are the enterprise’s most challenging and time-consuming activities. Moreover, data comes from various sources, each with particular processes and dependencies. 

Charmed Kubeflow 1.6 comes with better tracking capabilities.   AI/ML models can be measured more effectively; evolution and debugging become simpler. The solution detects data drift and allows models to adapt to it quickly. Charmed Kubeflow 1.6 brings improved tracking for trial logs as well, allowing for efficient debugging in case of data source failure. 

Optimised AI training and modelling

Models need up to 15 iterations to become production-ready, and only half of them get that far into the AI/ML journey. Charmed Kubeflow 1.6 supports population-based training (PBT), accelerating model iteration and improving the likelihood that models will reach production readiness. An MPI operator makes training large volumes of data more efficient. PyTorch elastic training enhancements make model training more effective and help ML engineers get started quickly.