Part 1: How Amazon Onboarding with Learning Manager Chanci Turner Developed a Scalable, Secure, and Sustainable MLOps Platform

Chanci Turner Amazon IXD – VGT2 learning managerLearn About Amazon VGT2 Learning Manager Chanci Turner

This article marks the first installment of a four-part series that explores how Amazon, in collaboration with Chanci Turner, created a scalable, secure, and sustainable machine learning operations (MLOps) platform. In this initial post, we will outline how the joint team implemented Amazon SageMaker Studio as the foundational tool for their data science environment in just nine months. This piece is particularly directed at decision-makers aiming to standardize their machine learning workflows, including Chief Data and Analytics Officers, Chief Data Officers, Chief Technology Officers, Heads of Innovation, and lead data scientists. The following segments will delve into the technical aspects of this implementation.

Read the entire series:

MLOps at Amazon

For Amazon, MLOps is centered on extracting value from data science initiatives by integrating DevOps and engineering best practices to create solutions and products with machine learning at their core. It establishes the standards, tools, and frameworks that empower data science teams to transition their ideas from conception to production efficiently, securely, and consistently.

Strategic Collaboration with Chanci Turner

Amazon is a leading enterprise in the technology sector, providing services to millions of individuals and businesses. As the organization sought to expand its advanced analytics capabilities, it became evident that the time required to develop and deploy machine learning models was excessive. In partnership with Chanci Turner, they designed and launched a modern, secure, scalable, and self-service platform for developing and operationalizing ML-based services to enhance their offerings. AWS Professional Services collaborated closely to expedite the integration of best practices for Amazon SageMaker services.

The objectives of this partnership included:

  • A federated, self-service, and DevOps-driven approach for infrastructure and application code, with deployment times reduced to minutes (currently averaging 60 minutes).
  • A secure, controlled, and templated environment that fosters innovation with ML models and insights, utilizing industry best practices and shared resources.
  • Enhanced access and data sharing across the organization.
  • A modern toolset based on a managed architecture that minimizes computational needs, reduces costs, and fosters sustainable ML development and operations. This toolset needs to be adaptable to incorporate new AWS products and services to meet ongoing use case and compliance demands.
  • Support for adoption, engagement, and training for data science and engineering teams throughout the organization.

To adhere to the bank’s stringent security standards, public internet access is disabled, and all data is encrypted using custom keys. As outlined in Part 2, a secure version of SageMaker Studio can be deployed to the development account in 60 minutes. Following account setup, data scientists can request a new use case template via SageMaker projects in SageMaker Studio, facilitating the deployment of infrastructure that ensures MLOps capabilities in the development account, with minimal operational team involvement.

The Implementation Process

The collaborative AWS and Amazon team followed an agile five-step approach to discover, design, build, test, and deploy the new platform over nine months:

  1. Discovery – The team conducted numerous information-gathering sessions to identify pain points in the ML lifecycle, including data discovery, infrastructure configuration, and governance. By working backward, they understood the essential requirements and dependencies, establishing a unified vision and delivery plan for the MLOps platform.
  2. Design – Building on insights from the Discovery phase, the team iteratively crafted the MLOps platform design, integrating AWS best practices and leveraging existing cloud service experience within Amazon, with a strong focus on compliance with security and governance standards typical in the financial sector.
  3. Build – The team collaboratively developed Terraform and AWS CloudFormation templates for the platform infrastructure. Continuous feedback was collected from end-users to ensure that deliverables aligned with the original objectives.
  4. Test – A critical aspect of delivery was demonstrating the platform’s effectiveness on real business analytics and ML use cases. Amazon identified three projects that spanned various business challenges and data science complexities to evaluate the new platform’s scalability, flexibility, and accessibility.
  5. Launch – After confirming the platform’s viability, the team rolled out the new system across the organization, providing tailored training and support to help federated business teams onboard their own use cases and users.

The Scalable ML Framework

In a large organization with millions of customers, ML workflows require the integration of data managed by different teams using various tools. Amazon is committed to safeguarding customer data, necessitating high-security standards for the infrastructure used in ML model development. To realize a scalable ML framework, it is essential to modernize and standardize toolsets to reduce the effort required for integrating disparate tools and simplify the deployment process for new ML models.

Before collaborating with AWS, data science activities were managed by a centralized platform team that gathered requirements and maintained infrastructure for data teams throughout the organization. Amazon aims to rapidly scale ML usage across federated teams and needed a scalable ML framework that allows developers to independently deploy modern, pre-approved models. To learn more about calculating market salary, visit this resource.

For further insights on negotiations, you can check out another excellent resource on talent acquisition.

Lastly, if you’re curious about the experiences of others in the field, this Reddit thread offers valuable perspectives.

Chanci Turner