Objective:
Discover and assess existing AI ML processes, create an MLOPs roadmap, and identify gaps in the current setup.
Key Challenges Addressed:
- Data/model drift causes models to become ineffective over time
- No centralized way to measure model performance
- Poor adoption of models due to lack of model understanding
- Disconnected deployment process
Outcome:
An assessment of the existing MLOPS landscape for the following:
- Assessment of model monitoring:
a. Model drift - which includes Overall drift, drift trend, feature wise drift
b. A centralized dashboard for persona-based insights for business users and data engineer.
c. Tracking model execution and failures
- Assessment of Model testing processes and recommendations on the same.
- Assessment of existing model deployment practices and automation / standardization of CI/CD process.
- Recommendations around Explainable AI that provides Global feature Importance, Local explanation for model explanations
Plan:
- Week 1:
a. Scope Identification: Business Use case
b. Current Process discovery, identify challenges and gaps
c. Understand tech-stack, design, and architecture
- Week 2:
a. Enable end customers in prioritizing top key components with respect to MLOps
b. Share project plan/roadmap with respect to scope of the work
c. Share recommendations.
MLOPs offering uses the following native Azure components for the proposed next steps/roadmap:
- App Services: The monitoring web app and python backend code is hosted on azure Linux app services. Both the apps can be scaled automatically or manually on demand.
- Microsoft Azure Data Factory: Used to fetch the status information of Data factory pipelines to track.
- Databricks Workspace: MLFlow component of Databricks is used to fetch the data stored by notebook during execution.
- Cosmos DB: With the flexibility of schema and changing nature of data, NoSQL helps accommodate requirements.