Course
digicode: RAI500
MLOps Practices with Red Hat OpenShift AI
Course facts
Download as PDF- Implementing a successful MLOps adoption journey using proven open culture and practices for customer innovation
- Combining the best Open Source tools to create a full MLOps workflow, blending continuous discovery, continuous training, and continuous delivery
- Practicing cross-functional collaboration to improve team alignment and delivery efficiency by working beyond traditional silos
- Mastering the end-to-end journey of a Predictive Intelligent Application, from ideation and inner loop experimentation to production
- Automating model training processes by transitioning from experimentation to production-ready training pipelines
- Executing advanced deployments, including autoscaling and patterns like canary and blue-green, to ensure safe and seamless model rollouts
- Maintaining optimal model performance through continuous monitoring and by introducing data versioning for enhanced traceability
- Securing machine learning models by implementing automated security guardrails and utilizing robust Feature Stores
This course is a five-day immersive class, offering attendees an opportunity to experience and implement a successful MLOps adoption journey. While many AI or data science training programs focus on a particular framework or technology, this course covers how the best Open Source tools fit together in a full MLOps workflow. It blends continuous discovery, continuous training, and continuous delivery in a highly engaging experience simulating real-world machine learning scenarios.
To achieve the learning objectives, participants should include multiple roles from across the organization. Data scientists, machine learning engineers, platform engineers, architects, and product owners will gain experience working beyond their traditional silos. The daily routine simulates a real-world delivery team, where cross-functional teams learn how collaboration breeds innovation. Armed with shared experiences and best practices, the team can apply what it has learned to help the organization's culture and mission succeed in the pursuit of new projects and improved processes.
What is MLOps?
Brainstorm and explore what principles, practices, and cultural elements make up a MLOps model for ML model developments and deployments.
Inner Loop
Familiarize ourselves with the necessary tools for experimenting and building our model; we will create a workbench, explore the dataset, start tracking our experiments, and deploy our models.
Training Pipelines
Transition to automating the previous steps for productionizing our model training.
Outer Loop
Introduction to MLOps: a set of practices that automate and simplify machine learning workflows and deployments.
Here we will create our MLOps environment where the continuous training pipeline, automated deployment, and the supporting toolings will be running.
Monitoring
Machine learning models can be influenced by various factors, including changes in data patterns, shifts in user behavior, and evolving external conditions. By implementing continuous monitoring, we will proactively identify these changes, assess their impact on model accuracy, and make necessary adjustments to maintain optimal performance.
Data Versioning
Enhance traceability by introducing versioning for our datasets as they change over time.
Advanced Deployments
Properly handle pre- and post-processing for data and predictions, explore autoscaling to handle loads, and introduce advanced deployment patterns like canary and blue-green deployments to ensure safe and seamless model rollouts.
Feature Stores
Robust ways of dealing with data features and their changes, as well as making sure features are homogeneous between training and serving.
Security
Implement automated security guardrails to stay compliant with the organizations security practices and extend them to the models.
This experience demonstrates how individuals across different roles must learn to share, collaborate, and work toward a common goal to achieve positive outcomes and drive innovation.
It is especially valuable for:
- MLOps Platform Users: Data scientists, data engineers, and application developers.
- MLOps Platform Providers: Machine learning engineers, MLOps engineers, and platform engineers.
- MLOps Platform Stakeholders: Architects and IT managers.
The scenario incorporates technical aspects of working with machine learning systems, offering practical insights into how these roles can align their efforts.
You will learn how to continuously deliver value to your customers by accelerating the deployment of new models to market. Our instructors will share experiences and best practices learned from engaging directly with customers during Red Hat services engagements.
- High level understanding of AI or Red Hat AI Foundations is beneficial