<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Heba K. Ahmed</title>
    <description>The latest articles on DEV Community by Heba K. Ahmed (@hkamel12).</description>
    <link>https://dev.to/hkamel12</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F601697%2F44bf722c-b537-403a-b45d-e330d435814b.jpg</url>
      <title>DEV Community: Heba K. Ahmed</title>
      <link>https://dev.to/hkamel12</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/hkamel12"/>
    <language>en</language>
    <item>
      <title>Machine Learning Best Practices for Public Sector Organizations</title>
      <dc:creator>Heba K. Ahmed</dc:creator>
      <pubDate>Tue, 02 Nov 2021 14:11:55 +0000</pubDate>
      <link>https://dev.to/hkamel12/machine-learning-best-practices-for-public-sector-organizations-4ema</link>
      <guid>https://dev.to/hkamel12/machine-learning-best-practices-for-public-sector-organizations-4ema</guid>
      <description>&lt;h1&gt;
  
  
  Challenges for public sector
&lt;/h1&gt;

&lt;p&gt;Government, education, and nonprofit organizations face several challenges in implementing ML programs to accomplish their mission objectives. This section outlines some of the challenges in seven critical areas of an ML implementation. These are outlined as follows:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; &lt;strong&gt;Data Ingestion and Preparation&lt;/strong&gt;: Identifying, collecting, and transforming data is the foundation for ML. Once the data is extracted, it needs to be organized so that it is available for consumption with the necessary approvals in compliance with public sector guidelines.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;Model Training and Tuning&lt;/strong&gt;: One of the major challenges facing public sector organizations is the ability to create a common platform that provides ML algorithms and optimize model training performance with minimal resources without compromising on the quality of ML models.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;ML Operations (MLOps)&lt;/strong&gt;: Integrating ML into business operations requires significant planning and preparation. Also, effectively monitor the model in production to ensure that ML models do not lose their effectiveness&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;Management &amp;amp; Governance&lt;/strong&gt;: Public sector organizations face increased scrutiny to ensure that public funds are being properly utilized to serve mission needs. In addition, any underlying infrastructure, software, and licenses need to be maintained and managed.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;Security &amp;amp; Compliance&lt;/strong&gt;: The sensitive nature of the work done by public sector organizations results in increased security requirements at all levels of an ML platform.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;Cost Optimization&lt;/strong&gt;: The challenge facing public sector agencies is the need to account for the resources used, and to monitor the usage against specified cost centers and task orders.&lt;/li&gt;
&lt;li&gt; &lt;strong&gt;Bias &amp;amp; Explainability&lt;/strong&gt;: Public sector organizations need to invest significant time with appropriate tools, techniques to demonstrate explainability and lack of bias in their ML models.&lt;/li&gt;
&lt;/ol&gt;

&lt;h1&gt;
  
  
  Best Practices
&lt;/h1&gt;

&lt;p&gt;AWS Cloud provides several fully managed services that supply developers and data scientists with the ability to prepare, build, train, and deploy ML models.&lt;/p&gt;

&lt;h2&gt;
  
  
  Data Ingestion and Preparation
&lt;/h2&gt;

&lt;p&gt;The AWS Cloud provides services that enable public sector customers to overcome challenges in data ingestion, data preparation, and data quality. These are further described as follows:&lt;/p&gt;

&lt;h3&gt;
  
  
  Data Ingestion
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Streaming Data&lt;/strong&gt;: For streaming data, Amazon Kinesis and Amazon Managed Streaming for Apache Kafka (Amazon MSK) enable the collection, processing, and analysis of data in real time.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;a href="https://aws.amazon.com/kinesis/data-streams/"&gt;Amazon Kinesis Data Streams (KDS)&lt;/a&gt; is a service that enables ingestion of streaming data.&lt;/li&gt;
&lt;li&gt;  Ingestion of streaming videos can be done using &lt;a href="https://aws.amazon.com/kinesis/video-streams/"&gt;Amazon Kinesis Video Streams&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://aws.amazon.com/kinesis/data-firehose/"&gt;Amazon Kinesis Data Firehose&lt;/a&gt; is a service that can be used to deliver real-time streaming data to a chosen destination.&lt;/li&gt;
&lt;li&gt;  you can use &lt;a href="https://aws.amazon.com/streaming-data/what-is-kafka/"&gt;Amazon MSK&lt;/a&gt; to build and run applications that use Apache Kafka to process streaming data without needing Apache Kafka infrastructure management expertise.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Batch Data&lt;/strong&gt;: There are a number of mechanisms available for data ingestion in batch format. &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  With &lt;a href="https://aws.amazon.com/dms/"&gt;AWS Database Migration Services (AWS DMS)&lt;/a&gt;, you can replicate and ingest existing databases while the source databases remain fully operational.&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://aws.amazon.com/datasync/?whats-new-cards.sort-by=item.additionalFields.postDateTime&amp;amp;whats-new-cards.sort-order=desc"&gt;AWS DataSync&lt;/a&gt; is a data transfer service that simplifies, automates, and accelerates moving and replicating data between on-premises storage and AWS storage services such as &lt;a href="https://aws.amazon.com/efs/"&gt;Amazon Elastic File System (EFS)&lt;/a&gt; and &lt;a href="https://aws.amazon.com/pm/serv-s3/?trk=ps_a134p000004f2aOAAQ&amp;amp;trkCampaign=acq_paid_search_brand&amp;amp;sc_channel=PS&amp;amp;sc_campaign=acquisition_US&amp;amp;sc_publisher=Google&amp;amp;sc_category=Storage&amp;amp;sc_country=US&amp;amp;sc_geo=NAMER&amp;amp;sc_outcome=acq&amp;amp;sc_detail=amazon%20s3&amp;amp;sc_content=S3_e&amp;amp;sc_matchtype=e&amp;amp;sc_segment=488982706722&amp;amp;sc_medium=ACQ-P%7CPS-GO%7CBrand%7CDesktop%7CSU%7CStorage%7CS3%7CUS%7CEN%7CText&amp;amp;s_kwcid=AL!4422!3!488982706722!e!!g!!amazon%20s3&amp;amp;ef_id=CjwKCAjwmK6IBhBqEiwAocMc8iNYVwzRmfBbxDlaf2E_jmA6daQzxu3mQgey9dr8QGOq-vpHaCnohRoCsVUQAvD_BwE:G:s&amp;amp;s_kwcid=AL!4422!3!488982706722!e!!g!!amazon%20s3"&gt;Amazon S3&lt;/a&gt;.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Data Preparation
&lt;/h3&gt;

&lt;p&gt;AWS Cloud provides three services that provide prepare and organize the data.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;a href="https://aws.amazon.com/glue/"&gt;AWS Glue&lt;/a&gt; is a fully managed ETL service that makes it simple and cost-effective to categorize, clean, enrich, and migrate data from a source system to a data store for ML.&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://docs.aws.amazon.com/glue/latest/ug/what-is-glue-studio.html"&gt;AWS Glue Studio&lt;/a&gt; provides a graphical interface that enables visual composition of data transformation workflows on AWS Glue’s engine.&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://aws.amazon.com/glue/features/databrew/"&gt;AWS Glue DataBrew&lt;/a&gt;, a visual data preparation tool, can be used to simplify the process of cleaning and normalizing the data.&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://aws.amazon.com/sagemaker/"&gt;Amazon SageMaker&lt;/a&gt; is a comprehensive service that provides purpose-built tools for every step of ML development and implementation.&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://aws.amazon.com/sagemaker/data-wrangler/"&gt;Amazon SageMaker Data Wrangler&lt;/a&gt; is a service that enables the aggregation and preparation of data for ML and is directly integrated into Amazon SageMaker Studio.&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://aws.amazon.com/emr/"&gt;Amazon EMR&lt;/a&gt;: Many organizations use Spark for data processing and other purposes such as for a data warehouse.  Amazon EMR, a managed service for Hadoop-ecosystem clusters, can be used to process data.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Data quality
&lt;/h3&gt;

&lt;p&gt;Public sector organizations need to ensure that data ingested and prepared for ML is of the highest quality by establishing a well-defined data quality framework. See &lt;a href="https://aws.amazon.com/blogs/industries/how-to-architect-data-quality-on-the-aws-cloud/"&gt;How to Architect Data Quality on the AWS Cloud&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Model Training and Tuning
&lt;/h2&gt;

&lt;p&gt;One of the major challenges facing the public sector is the ability for team members to apply a consistent pattern or framework for working with multitudes of options that exist in this space.&lt;br&gt;
The AWS Cloud enables public sector customers to overcome challenges in model selection, training, and tuning as described in the following.&lt;/p&gt;

&lt;h3&gt;
  
  
  Model Selection
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Amazon SageMaker&lt;/strong&gt; provides the flexibility to select from a wide number of options using a consistent underlying platform.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Programming Language&lt;/strong&gt;: Amazon SageMaker notebook kernels provide the ability to use both Python, as well as R, natively. To use coding languages such as Stan or Julia, create a Docker image and bring it into SageMaker for model training and inference. To use programming languages like C++ or Java, custom images on Amazon ECS/EKS can be used to perform model training.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Built-in algorithms&lt;/strong&gt;: Amazon SageMaker Built-in Algorithms provides several built-in algorithms covering different types of ML problems. These algorithms are already optimized for speed, scale, and accuracy.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Script Mode&lt;/strong&gt;: For experienced ML programmers who are comfortable with using their own algorithms, Amazon SageMaker provides the option to write your custom code (script) in a text file with a .py extension (see Figure 1).
&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--kToaikO5--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/jh81eh01tn23dqflelc0.jpg" alt="Image description" width="572" height="307"&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Use a custom Docker image&lt;/strong&gt;: ML programmers may use algorithms that aren't supported by SageMaker, a custom Docker image can be used in these cases to train the algorithm and provide the model (see Figure 2 below).
&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--CJwPqqIA--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/o0a09uqwdiywks2d0c29.jpg" alt="Image description" width="531" height="283"&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Model Training
&lt;/h3&gt;

&lt;p&gt;Amazon SageMaker provides a number of built-in options for optimizing model training performance, input data formats, and distributed training.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Data parallel&lt;/strong&gt;: SageMaker’s &lt;a href="https://sagemaker.readthedocs.io/en/stable/api/training/smd_data_parallel.html"&gt;distributed data parallel library&lt;/a&gt; can be considered for running training jobs in parallel in case it is common to have multiple training iterations per epoch.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Pipe mode&lt;/strong&gt;: Pipe mode accelerates the ML training process: instead of downloading data to the local Amazon EBS volume prior to starting the model training, Pipe mode streams data directly from S3 to the training algorithm while it is running.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Incremental training&lt;/strong&gt;: Amazon SageMaker supports &lt;a href="https://docs.aws.amazon.com/sagemaker/latest/dg/incremental-training.html"&gt;incremental training&lt;/a&gt; to train a new model from an existing model artifact, to save both training time and resources.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Model Parallel training&lt;/strong&gt;: Amazon SageMaker’s distributed model parallel library can be used to split a model automatically and efficiently across multiple GPUs and instances and coordinate model training.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Model Tuning
&lt;/h3&gt;

&lt;p&gt;Amazon SageMaker provides automatic hyperparameter tuning to find the best version of a model in an efficient manner. The following best practices ensure a better tuning result:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Limit the number of hyperparameters&lt;/strong&gt;: limiting the search to a much smaller number is likely to give better results, as this can reduce the computational complexity of a hyperparameter tuning job and provides better understanding of how a specific hyperparameter would affect the model performance.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Choose hyperparameter ranges appropriately&lt;/strong&gt;: better results are obtained by limiting the search to a small range of values.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Pay attention to scales for hyperparameters&lt;/strong&gt;: During hyperparameter tuning, SageMaker attempts to figure out if hyperparameters are log-scaled or linear-scaled.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Set the best number of concurrent training jobs&lt;/strong&gt;: a tuning job improves only through successive rounds of experiments. Typically, running one training job at a time achieves the best results with the least amount of compute time.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Report the wanted objective metric for tuning when the training job runs on multiple instances&lt;/strong&gt;: distributed training jobs should be designed such that the objective metric reported is the one that is needed.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Enable early stopping for hypermeter tuning job&lt;/strong&gt;: Early stopping helps reduce compute time and helps avoid overfitting the model.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Run a warm start using previous tuning jobs&lt;/strong&gt;: Use a warm start for fine-tuning previous hyperparameter tuning jobs.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  MLOps
&lt;/h1&gt;

&lt;p&gt;MLOps is the discipline of integrating ML workloads into release management, Continuous Integration / Continuous Delivery (CI/CD), and operations.&lt;br&gt;
Using ML models in software development makes it difficult to achieve own different parts of the process; data engineers might be building pipelines to make ML engineers or developers have to work on integrating the models and releasing them.&lt;br&gt;
AWS Cloud provides a number of different options that solve these challenges, either by building an MLOps pipeline from scratch or by using managed services.&lt;/p&gt;

&lt;h3&gt;
  
  
  Amazon SageMaker Projects
&lt;/h3&gt;

&lt;p&gt;By using a SageMaker project, teams of data scientists and developers can work together on ML business problems. SageMaker projects use MLOps templates that automate the model building and deployment pipelines using CI/CD.&lt;/p&gt;

&lt;h3&gt;
  
  
  Amazon SageMaker Pipelines
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://aws.amazon.com/sagemaker/pipelines/"&gt;SageMaker Pipelines&lt;/a&gt; is a purpose-built, CI/CD service for ML. &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  It brings CI/CD practices to ML, such as maintaining parity between development and production environments, version control, on-demand testing, and end-to-end automation, helping scale ML throughout the organization.&lt;/li&gt;
&lt;li&gt;  With the SageMaker Pipelines model registry, model versions can be stored in a central repository for easy browsing, discovery, and selection of the right model for deployment based on business requirements.&lt;/li&gt;
&lt;li&gt;  Pipelines provide the ability to log each step within the ML workflow for a complete audit trail of model components.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  AWS CodePipeline and AWS Lambda
&lt;/h3&gt;

&lt;p&gt;For AWS programmers, CodePipeline exists to utilize the same workflows for ML. Figure 3 below represents a reference pipeline for deployment on AWS.&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--z5THGf_6--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/b1bogwo29jd79wc60t0j.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--z5THGf_6--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/b1bogwo29jd79wc60t0j.jpg" alt="Image description" width="572" height="390"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  AWS StepFunctions Data Science Software Development Kit (SDK)
&lt;/h3&gt;

&lt;p&gt;The &lt;a href="https://docs.aws.amazon.com/step-functions/latest/dg/concepts-python-sdk.html"&gt;AWS Step Functions Data Science SDK&lt;/a&gt; is an open-source Python library that allows data scientists to create workflows that process and publish ML models using SageMaker and Step Functions.&lt;/p&gt;

&lt;h2&gt;
  
  
  AWS MLOps Framework
&lt;/h2&gt;

&lt;p&gt;Figure 4 below illustrates an AWS solution that provides an extendable framework with a standard interface for managing ML pipelines. The solution provides a ready-made template to upload trained models (also referred to as a bring your own model), configure the orchestration of the pipeline, and monitor the pipeline's operations.&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--4wGP3AZ5--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/j4nk7pr8zi5teulcme4x.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--4wGP3AZ5--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/j4nk7pr8zi5teulcme4x.jpg" alt="Image description" width="629" height="302"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Deploy Custom Deep Learning Models
&lt;/h3&gt;

&lt;p&gt;AWS also provides the option to deploy custom code on virtual machines using the followings:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Amazon EC2, and containers using self-managed Kubernetes on Amazon EC2.&lt;/li&gt;
&lt;li&gt;  Amazon Elastic Container Service (Amazon ECS) and Amazon Elastic Kubernetes Service (Amazon EKS).&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://aws.amazon.com/machine-learning/amis/"&gt;AWS Deep Learning AMIs&lt;/a&gt; can be used to accelerate deep learning by quickly launching Amazon EC2 instances&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://aws.amazon.com/machine-learning/containers/"&gt;AWS Deep Learning Containers&lt;/a&gt; are Docker images pre-installed with deep learning frameworks to deploy optimized ML environments.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Deploy ML at the edge
&lt;/h3&gt;

&lt;p&gt;In some cases, such as with edge devices, inferencing needs to occur even when there is limited or no connectivity to the cloud. Mining fields are an example of this type of use case.&lt;br&gt;
&lt;a href="https://aws.amazon.com/greengrass/"&gt;AWS IoT Greengrass&lt;/a&gt; enables ML inference locally using models that are created, trained, and optimized in the cloud using Amazon SageMaker, AWS Deep Learning AMI, or AWS Deep Learning Containers, and deployed on the edge devices.&lt;/p&gt;

&lt;h2&gt;
  
  
  Management and Governance
&lt;/h2&gt;

&lt;p&gt;ML workloads need to provide increased visibility for monitoring and auditing. This section highlights several AWS services and associated best practices to address these management and governance challenges.&lt;br&gt;
Enable governance and control&lt;br&gt;
AWS Cloud provides several services that enable governance and control. These include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;a href="https://aws.amazon.com/controltower/?control-blogs.sort-by=item.additionalFields.createdDate&amp;amp;control-blogs.sort-order=desc"&gt;&lt;strong&gt;AWS Control Tower&lt;/strong&gt;&lt;/a&gt;: creates a landing zone that consists of a predefined structure of accounts using AWS Organizations, the ability to create accounts using &lt;a href="https://docs.aws.amazon.com/ARG/latest/userguide/"&gt;AWS Service Catalog&lt;/a&gt;, enforcement of compliance rules called guardrails using Service Control Policies, and detection of policy violations using AWS Config.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;AWS License Manager&lt;/strong&gt;: AWS License Manager can be used to track this software obtained from the AWS Marketplace and keep a consolidated view of all licenses and enables sharing of licenses with other accounts in the organization.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Resource Tagging&lt;/strong&gt;: Automated tools such as AWS Resource Groups and the Resource Groups Tagging API enable programmatic control of tags, making it easier to automatically manage, search, and filter tags and resources.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Provision ML resources that meet policies
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;AWS CloudFormation&lt;/strong&gt;: provides a mechanism to model a collection of related AWS and third-party resources, provision them quickly and consistently, and manage them throughout their lifecycles, by treating infrastructure as code.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;AWS Cloud Development Kit (CDK)&lt;/strong&gt;: The AWS CDK allows teams to define cloud infrastructure in code directly in supported programming languages (i.e., TypeScript, JavaScript, Python, Java, and C#).&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://docs.aws.amazon.com/servicecatalog/latest/adminguide/getstarted.html"&gt;&lt;strong&gt;AWS Service Catalog&lt;/strong&gt;&lt;/a&gt;: provides a solution for public sector by enabling the central management of commonly deployed IT services and achieves consistent governance and meets compliance requirements.
&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--1TQnm6uu--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/f8zmdnt2md6et3053q4e.jpg" alt="Image description" width="545" height="257"&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Operate environment with governance
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Amazon CloudWatch&lt;/strong&gt; is a monitoring and observability service used to monitor resources and applications run on AWS in real time.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Amazon EventBridge&lt;/strong&gt; is a serverless event bus service that can monitor status change events in Amazon SageMaker.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;SageMaker Model Monitor&lt;/strong&gt; can be used to continuously monitor the quality of ML models in production. Model Monitor can notify team members when there are deviations in the model quality. see the &lt;a href="https://github.com/awslabs/amazon-sagemaker-examples/blob/master/sagemaker_model_monitor/introduction/SageMaker-ModelMonitoring.ipynb"&gt;Introduction to Amazon SageMaker Model Monitor&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;AWS CloudTrail&lt;/strong&gt;: provides a record of actions taken by a user, role, or an AWS service in SageMaker. CloudTrail captures all API calls for SageMaker.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Security and compliance
&lt;/h1&gt;

&lt;p&gt;Public sector organizations have a number of security challenges and concerns.&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--jDL7Szg6--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/qtzqm00391l8qxi5gf30.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--jDL7Szg6--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/qtzqm00391l8qxi5gf30.jpg" alt="Image description" width="624" height="387"&gt;&lt;/a&gt;&lt;br&gt;
This section provides best practices and guidelines to address some of these security and compliance challenges.&lt;/p&gt;

&lt;h3&gt;
  
  
  Compute and network isolation
&lt;/h3&gt;

&lt;p&gt;In public sector ML projects, it’s required to keep the environments, data and workloads secure and isolated from internet access. These can be achieved using the following methods:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Provision ML components in an isolated VPC with no internet access.&lt;/li&gt;
&lt;li&gt;  Use VPC endpoint and end-point policies to further limit access.&lt;/li&gt;
&lt;li&gt;  Allow access from only within the VPC: An IAM policy can be created to prevent users outside the VPC from accessing SageMaker Studio or SageMaker notebooks over the internet.&lt;/li&gt;
&lt;li&gt;  Intrusion detection and prevention: AWS Gateway Load Balancer (GWLB) can be used to deploy, scale, and manage the availability of third-party virtual appliances.&lt;/li&gt;
&lt;li&gt;  Additional security to allow access to resources outside your VPC.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Data Protection
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Protect data at rest&lt;/strong&gt;: AWS Key Management service (KMS) can be used to encrypt ML data, studio notebooks and SageMaker notebook instances.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Protect data in transit&lt;/strong&gt;: AWS makes extensive use of HTTPS communication for its APIs.&lt;/li&gt;
&lt;li&gt;  Secure shared notebook instances.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Authentication and Authorization
&lt;/h2&gt;

&lt;p&gt;AWS IAM enables control of access to AWS resources. IAM administrators control who can be authenticated (signed in) and authorized (have permissions) to use SageMaker resources. Two common ways to implement least privilege access to the SageMaker environments are identity-based policies and resource-based policies:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;a href="https://docs.aws.amazon.com/IAM/latest/UserGuide/access_policies.html#policies_id-based"&gt;Identity-based policies&lt;/a&gt; are attached to an IAM user, group, or role. These policies specify what that identity can do.&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://docs.aws.amazon.com/IAM/latest/UserGuide/access_policies.html#policies_resource-based"&gt;Resource-based policies&lt;/a&gt; are attached to a resource. These policies specify who has access to the resource, and what actions can be performed on it.
Please refer to &lt;a href="https://aws.amazon.com/blogs/machine-learning/configuring-amazon-sagemaker-studio-for-teams-and-groups-with-complete-resource-isolation/"&gt;Configuring Amazon SageMaker Studio&lt;/a&gt; for teams and groups with complete resource isolation.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Artifact and model management
&lt;/h3&gt;

&lt;p&gt;The recommended best practice is to use version control to track code or other model artifacts. If model artifacts are modified or deleted, either accidentally or deliberately, version control allows you to roll back to a previous stable release.&lt;/p&gt;

&lt;h3&gt;
  
  
  Security compliance
&lt;/h3&gt;

&lt;p&gt;For a list of AWS services in scope of specific compliance programs, see &lt;a href="https://aws.amazon.com/compliance/services-in-scope/"&gt;AWS Services in Scope by Compliance Program&lt;/a&gt;. AWS provides the following resources to help with compliance:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Security and Compliance Quick Start Guides.&lt;/li&gt;
&lt;li&gt;  Architecting for HIPAA Security and Compliance Whitepaper.&lt;/li&gt;
&lt;li&gt;  AWS Compliance Resources.&lt;/li&gt;
&lt;li&gt;  AWS Config.&lt;/li&gt;
&lt;li&gt;  AWS Security Hub.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Cost optimization
&lt;/h1&gt;

&lt;p&gt;Cost management is a primary concern for public sector organizations projects to ensure the best use of public funds while enabling agency missions.&lt;br&gt;
Prepare&lt;br&gt;
Cost control in this phase can be accomplished using the following techniques:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Data Storage&lt;/strong&gt;: it is essential to establish a cost control strategy at the storage level.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Data Labeling&lt;/strong&gt;: Data labeling is a key process of identifying raw data (such as images, text files, and videos) and adding one or more meaningful and informative labels to provide context so that an ML model can learn from it.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Data Wrangling&lt;/strong&gt;: Amazon SageMaker Data Wrangler can be used to reduce this time spent, lowering the costs of the project. With Data Wrangler, data can be imported from various data sources, and transformed without requiring coding.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Build
&lt;/h3&gt;

&lt;p&gt;Cost control in this phase can be accomplished using the following techniques:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Notebook Utilization&lt;/strong&gt;: It helps prepare and process data, write code to train models, deploy models to SageMaker hosting, and test or validate models.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Test code locally&lt;/strong&gt;: Before a training job is submitted, running the fit function in local mode enables early feedback prior to running in SageMaker’s managed training or hosting environments.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Use Pipe mode&lt;/strong&gt; (where applicable) to reduce training time.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Find the right balance&lt;/strong&gt;: Performance vs. accuracy.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Jumpstart&lt;/strong&gt;: SageMaker JumpStart accelerates time-to-deploy over 150 open-source models and provides pre-built solutions, preconfigured with all necessary AWS services required to launch the solution into production, including CloudFormation templates and reference architecture.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;AWS Marketplace&lt;/strong&gt;: AWS Marketplace is a digital catalog with listings from independent software vendors to find, test, buy, and deploy software that runs on AWS.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Train and Tune
&lt;/h3&gt;

&lt;p&gt;Cost control in this phase can be accomplished using the following techniques:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Use Spot Instances&lt;/strong&gt;: Training jobs can be configured to use Spot Instances and a stopping condition can be used to specify how long Amazon SageMaker waits for a job to run using EC2 Spot Instances.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Hyperparameter optimization (HPO)&lt;/strong&gt;: it’s automatically adjusting hundreds of different combinations of parameters to quickly arrive at the best solution for your ML problem.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;CPU vs GPU&lt;/strong&gt;: CPUs are best at handling single, more complex calculations sequentially, whereas GPUs are better at handling multiple but simple calculations in parallel.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Distributed Training&lt;/strong&gt;: the training process can be sped up by distributing training on multiple machines or processes in a cluster as described earlier&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Monitor the performance&lt;/strong&gt; of your training jobs to identify waste.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Deploy and Manage
&lt;/h3&gt;

&lt;p&gt;This step of the ML lifecycle involves deployment of the model to get predictions, and managing the model to ensure it meets functional and non-functional requirements of the application.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;strong&gt;Endpoint deployment&lt;/strong&gt;: Amazon SageMaker enables testing of new models using A/B testing. Endpoints need to be deleted when testing is completed to reduce costs.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Multi-model endpoints&lt;/strong&gt;: reduce hosting costs by improving endpoint utilization and provide a scalable and cost-effective solution to deploying many models.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Auto Scaling&lt;/strong&gt;: optimizes the cost of model endpoints.&lt;/li&gt;
&lt;li&gt;  Amazon Elastic Inference for deep learning.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;Analyzing costs with Cost Explorer&lt;/strong&gt;: Cost Explorer is a tool that enables viewing and analyzing AWS service-related costs and usage including SageMaker.&lt;/li&gt;
&lt;li&gt;  &lt;strong&gt;AWS Budgets&lt;/strong&gt;: help you manage Amazon SageMaker costs, including development, training, and hosting, by setting alerts and notifications when cost or usage exceeds (or is forecasted to exceed) the budgeted amount.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Bias and Explainability
&lt;/h1&gt;

&lt;p&gt;Demonstrating explainability is a significant challenge because complex ML models are hard to understand and even harder to interpret and debug.&lt;br&gt;
AWS Cloud provides the following capabilities and services to assist public sector organizations in resolving these challenges.&lt;/p&gt;

&lt;h3&gt;
  
  
  Amazon SageMaker Debugger
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://docs.aws.amazon.com/sagemaker/latest/dg/train-debugger.html"&gt;Amazon SageMaker Debugger&lt;/a&gt; provides visibility into the model training process for real-time and offline analysis. In the existing training code for TensorFlow, Keras, Apache MXNet, PyTorch, and XGBoost.&lt;br&gt;
It provides three built-in tensor collections called feature importance, average_shap, and full_shap, to visualize and analyze captured tensors specifically for model explanation.&lt;/p&gt;

&lt;h3&gt;
  
  
  Amazon SageMaker Clarify
&lt;/h3&gt;

&lt;p&gt;Amazon SageMaker Clarify is a service that is integrated into SageMaker Studio and detects potential bias during data preparation, model training, and in deployed models, by examining specified attributes.&lt;br&gt;
Bias in attributes related to age can be examined in the initial dataset, in the trained as well as the deployed model, and quantified in a detailed report.&lt;br&gt;
SageMaker Clarify also enables explainability by including feature importance graphs using SHAP to help explain model predictions.&lt;/p&gt;

&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;Public sector organizations have complex mission objectives and are increasingly adopting ML services to help with their initiatives. ML can transform the way government agencies operate, and enable them to provide improved citizen services. However, several barriers remain for these organizations to implement ML. This whitepaper outlined some of the challenges and provided best practices that can help address these challenges using AWS Cloud.&lt;/p&gt;

&lt;h1&gt;
  
  
  Next Steps
&lt;/h1&gt;

&lt;p&gt;Adopting the AWS Cloud can provide you with sustainable advantages for telehealth systems. Your AWS account team can work together with your team and/or your chosen member of the AWS Partner Network (APN) to implement your enterprise cloud computing initiatives. You can reach out to an AWS partner through the AWS Partner Network. Get started on AI and ML by visiting &lt;a href="https://aws.amazon.com/machine-learning/containers/"&gt;AWS ML&lt;/a&gt;, AWS ML Embark Program, or the &lt;a href="https://aws.amazon.com/ml-solutions-lab/"&gt;ML Solutions Lab&lt;/a&gt;.&lt;/p&gt;

</description>
      <category>aws</category>
      <category>machinelearning</category>
      <category>cloud</category>
    </item>
    <item>
      <title>Hybrid Machine Learning</title>
      <dc:creator>Heba K. Ahmed</dc:creator>
      <pubDate>Sun, 15 Aug 2021 11:58:06 +0000</pubDate>
      <link>https://dev.to/aws-builders/hybrid-machine-learning-24b</link>
      <guid>https://dev.to/aws-builders/hybrid-machine-learning-24b</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;Companies who adopt a cloud-native approach realize its value when they marry compute capacity.&lt;br&gt;
But for those companies born before the cloud, in many cases even before generalized computers,&lt;br&gt;
to build on-premises, how can they realize the value of newly launched cloud services when the early.&lt;br&gt;
AWS proposes a series of tenets to guide our discussion of a world-class hybrid machine learning experience: &lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; Seamless management experience. &lt;/li&gt;
&lt;li&gt; Tunable latency. &lt;/li&gt;
&lt;li&gt; Fast time-to-value. &lt;/li&gt;
&lt;li&gt; Flexible. &lt;/li&gt;
&lt;li&gt; Low cost. &lt;/li&gt;
&lt;li&gt; End in the cloud&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;This document follows the machine learning lifecycle, from development to training and deployment. &lt;/p&gt;

&lt;h3&gt;
  
  
  What is Hybrid?
&lt;/h3&gt;

&lt;p&gt;We look at hybrid architectures as having a minimum of two compute environments, what we will call “primary” and “secondary” environments. Generally speaking, we see the primary environment as where the workload begins, and the secondary environment is where the workload ends.&lt;/p&gt;

&lt;h3&gt;
  
  
  What “Hybrid” is not?
&lt;/h3&gt;

&lt;p&gt;While either your primary or secondary environment may be, in fact, another cloud provider, we prefer not to explicitly discuss those patterns here. That’s because the technical details and capabilities around specific services, features, and global deployment options are not necessarily the same across all cloud providers, and we consider that type of analysis outside the scope for this document. &lt;/p&gt;

&lt;h2&gt;
  
  
  Hybrid patterns for development
&lt;/h2&gt;

&lt;p&gt;Development refers to the phase in machine learning when customers are iteratively building models. This may or may not include exploratory data analysis, deep learning model development, &lt;/p&gt;

&lt;p&gt;Generally speaking, there are two major options for hybrid development:&lt;/p&gt;

&lt;p&gt;1-  laptop and desktop personal computers.&lt;br&gt;
2-  Self-managed local servers utilizing specialized GPUs, colocations, self-managed racks, or corporate data centers. Customers can develop in one or both of these compute environments, and below we’ll describe hybrid patterns for model deployment using both of these.&lt;/p&gt;

&lt;h3&gt;
  
  
  Develop on personal computers, to train and host in the cloud
&lt;/h3&gt;

&lt;p&gt;Customers can use local development environments, such as PyCharm or Jupyter installations on their laptops or personal computer, and then connect to the cloud via AWS Identity and Access Management (IAM) permissions and interface with AWS service API’s through the AWS CLI or an AWS SDK (ex boto3).&lt;br&gt;
&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fw957knb0hp3it64bgame.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fw957knb0hp3it64bgame.jpg" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Advantages&lt;/strong&gt;&lt;br&gt;
1-  You have full control of your IDE in this scenario. You just have to open up your computer to get started.&lt;br&gt;
2-  You can easily manage what’s sitting in your S3 bucket, vs what’s running on your local laptop.&lt;br&gt;
3-  You iteratively write a few lines of code in your complex models, you check them locally, and you only land in the cloud to scale/track/deploy.&lt;br&gt;
&lt;strong&gt;Disadvantages&lt;/strong&gt;&lt;br&gt;
1-  Inability to scale beyond the compute resources of your laptop.&lt;br&gt;
2-  Lack of access to GUI-centric features like Autopilot, Data Wrangler, Pipelines.&lt;br&gt;
3-  If your laptop dies and you didn’t back up externally, your work is gone! Difficulty in onboarding non-super user employees can increase over time as software, OS, and hardware versions change.&lt;br&gt;
4-  This onboarding difficulty gets more and more painful as time goes by, in some worst-case scenarios, it leads to highly valued employees not getting access to Python or Pandas for multiple months!&lt;br&gt;
&lt;strong&gt;When to move&lt;/strong&gt;: &lt;br&gt;
If you find yourself spending a significant portion of your time managing local compute environments, it’s time to move to the cloud.&lt;/p&gt;

&lt;h3&gt;
  
  
  Develop on local servers, to train and host in the cloud
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbj3pwdqzixp8dzrh0sln.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbj3pwdqzixp8dzrh0sln.jpg" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Advantages&lt;/strong&gt;&lt;br&gt;
1-  Ability to capitalize on previous investments in local compute.&lt;br&gt;
2-  Simplifies infrastructure setup required to meet some regulatory requirements, such as those for specialized industries (health, gaming, financial). &lt;br&gt;
3-  Lower price point per compute on some compute environments, including some advanced GPU’s &lt;br&gt;
4-  Ideal for non-bursting, stable computes with high precision in the 6+ month forecast.&lt;br&gt;
&lt;strong&gt;Disadvantages&lt;/strong&gt;&lt;br&gt;
1-   A fundamental challenge to dynamically provision compute resources with the needs of your business, leading teams to frequently be stuck in either overutilization of local compute resources or under-utilization.&lt;br&gt;
2-  Expensive GPUs can take months to ship, which leads to a larger total cost of ownership. New ideas and features can take longer to launch, because of extra effort to develop.&lt;br&gt;
&lt;strong&gt;When to move&lt;/strong&gt;&lt;br&gt;
1-  When you spend more time managing your local development than you do working on new data science projects.&lt;br&gt;
2-  When the multiple months it takes to procure, wait for, and provision additional compute resources leave your development teams sitting idle.&lt;/p&gt;

&lt;h2&gt;
  
  
  Hybrid patterns for training
&lt;/h2&gt;

&lt;p&gt;Usually, a hybrid pattern for training comes down to one of two paths. &lt;br&gt;
1-  Either you are training locally, and you want to deploy in the cloud.&lt;br&gt;
2-   Or you have all of your data sitting on local resources, and you want to select from that to move into the cloud to train.&lt;/p&gt;

&lt;h3&gt;
  
  
  Training locally, to deploy in the cloud
&lt;/h3&gt;

&lt;p&gt;1-  First, if you are training locally then you will need to acquire the compute capacity to train a model.&lt;br&gt;
2-  After your model is trained, there are two common approaches for packaging and hosting it in the cloud.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Docker – using a Docker file you can build your custom image that hosts your inference script, model artifact, and packages. Register this image in the Elastic Container Registry (ECR), and point to it from your SageMaker estimator.&lt;/li&gt;
&lt;li&gt;  Another option is using the pre-built containers within the SageMaker Python SDK, also known as the deep learning containers (or DL AMI’s). Bring your inference script and custom packages, upload your model artifact to Amazon S3, and import an estimator for your framework of choice. Define the version of the framework you need in the estimator, or install it directly with a requirements.txt file or a custom bash script.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo7s7n6u5cz32a6mbuab4.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo7s7n6u5cz32a6mbuab4.jpg" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  How to monitor your model in the cloud
&lt;/h2&gt;

&lt;p&gt;A key feature for hosting is model monitor, or the ability to detect data, bias, feature, and model thresholds, trigger a re-training pipeline.&lt;br&gt;
1-  Upload your training data to an Amazon S3 bucket, and use our pre-built image to learn the upper and lower bounds of your training data.&lt;br&gt;
2-  you will receive a JSON file with the upper and lower statistically recommended bounds for each feature. You can modify these thresholds.&lt;br&gt;
3-  After confirming your thresholds, schedule monitoring jobs in your production environment. These jobs run automatically, comparing your captured inference requests in Amazon S3 with your thresholds.&lt;br&gt;
4-  You will receive CloudWatch alerts when your inference data is outside of your pre-determined thresholds, and you can use those alerts to trigger a re-training pipeline.&lt;/p&gt;

&lt;h3&gt;
  
  
  How to handle retraining/retuning
&lt;/h3&gt;

&lt;p&gt;1-  It is easy to run a retraining and retuning job in the cloud without the overhead of provision, scheduling, and managing your physical resources around this job.&lt;br&gt;
2-  SageMaker makes train and tuning jobs easy to manage because all you need to bring is your training script and dataset.&lt;br&gt;
3-  Follow best practices for training on SageMaker, ensuring your new dataset is loaded into an Amazon S3 bucket or other supported data source.&lt;br&gt;
4-  Another key feature of hosting models in SageMaker is multi-model endpoints.&lt;br&gt;
5-  Define your inference script, ensuring the framework is supported by SageMaker multi-model.&lt;br&gt;
6-  Create the multi-model endpoint, pointing to Amazon S3, and load your model artifacts into the SageMaker endpoint, invoking the endpoint with the name of the model you want to use.&lt;/p&gt;

&lt;h3&gt;
  
  
  How to serve thousands of models in the cloud at a low cost
&lt;/h3&gt;

&lt;p&gt;Another key feature of hosting models in SageMaker is multi-model endpoints. SageMaker endpoint, invoking the endpoint with the name of the model you want to use.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Advantages:&lt;/strong&gt;&lt;br&gt;
1-  have more control over your training environment.&lt;br&gt;
2-  the cloud not only provides greater flexibility, but can increase a firm’s security posture, by freeing up resources from physical security, patching, and procurement.&lt;br&gt;
&lt;strong&gt;Disadvantages:&lt;/strong&gt;&lt;br&gt;
1-  Not taking advantage of cost savings on spot instances.&lt;br&gt;
2-  Not using pre-built Docker images, but potentially wasting engineering effort developing these from scratch.&lt;br&gt;
3-  Not using advanced distributed training toolkits or custom hardware like the upcoming training.&lt;br&gt;
4-  Not using prebuilt tuning packages, but need to build or buy your tuning package.&lt;br&gt;
5-  Not using the debugger, profiler, feature store, and other training benefits.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;When to move:&lt;/strong&gt;&lt;br&gt;
1-  When the cost of developing your local training platform exceeds its use, &lt;br&gt;
2-  Also when the time to provision additional computes resources is far outstripped by the demand for training by your data science resources or business needs.&lt;/p&gt;

&lt;h2&gt;
  
  
  Storing data locally, to train and deploy in the cloud
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Schedule data transfer jobs with AWS DataSync
&lt;/h3&gt;

&lt;p&gt;1-  AWS DataSync is a data transfer service that simplifies, automates, and accelerates moving data between on-premises storage systems and AWS storage services, as well as between AWS storage services.&lt;br&gt;
2-  Using AWS DataSync you can easily move petabytes of data from your local on-premises servers up to the AWS cloud.&lt;br&gt;
3-  AWS DataSync connects to your local NFS resources, looks for any changes, and handles populating your cloud environment.&lt;/p&gt;

&lt;h2&gt;
  
  
  Migrating from Local HDFS
&lt;/h2&gt;

&lt;p&gt;you might wholly embrace HDFS as your center and move towards hosting it within a managed service, Amazon Elastic Map Reduce (EMR).&lt;/p&gt;

&lt;p&gt;If you are interested in learning how to migrate from local HDFS clusters to Amazon EMR, please see this migration guide: &lt;a href="https://d1.awsstatic.com/whitepapers/amazon_emr_migration_guide.pdf" rel="noopener noreferrer"&gt;https://d1.awsstatic.com/whitepapers/amazon_emr_migration_guide.pdf&lt;/a&gt; &lt;/p&gt;

&lt;h3&gt;
  
  
  Best practices
&lt;/h3&gt;

&lt;p&gt;1-  Use Amazon S3 intelligent tiering for objects over 128 KB.&lt;br&gt;
2-  Use multiple AWS accounts, and connect them with Organizations.&lt;br&gt;
3-  Set billing alerts.&lt;br&gt;
4-  Enable SSO with your current Active Directory provider.&lt;br&gt;
5-  Turn on Studio!&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Advantages:&lt;/strong&gt;&lt;br&gt;
1-  This is a fast way to realize the value of your locally stored datasets, particularly during a cloud migration.&lt;br&gt;
2-  Training and developing in the cloud gives you access to fully managed features within Amazon SageMaker and the entire AWS cloud.&lt;br&gt;
3-  You can offload your on-premises resources more easily by leveraging capabilities in the cloud.&lt;br&gt;
4-  This frees up your teams from procuring, provisioning, securing, and patching local compute resources, enabling them to dynamically scale these with the needs of your business.&lt;br&gt;
5-  Generally speaking, you can deploy more models more quickly by training and hosting in the cloud.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Disadvantages:&lt;/strong&gt;&lt;br&gt;
1-  Expending more resources storing data locally than potentially necessary.&lt;br&gt;
2-  If you intend to train your ML models locally, you should anticipate a high volume of node drop-outs in your data centers. One large job can easily consume 1TB of RAM while another can require smaller memory, but execute for potentially days.&lt;br&gt;
3-  Cost mitigation can be important here. Customers should be aware of any data duplication across environments and take action to reduce costs aggressively.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;When to move&lt;/strong&gt;&lt;br&gt;
1-  When the cost of managing, securing, and storing your data on-premises exceeds the cost of archiving and storing it in the cloud.&lt;/p&gt;

&lt;h2&gt;
  
  
  Develop in the cloud while connecting to data hosted on-premises
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fj26g9kv6o14x3fuy7bok.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fj26g9kv6o14x3fuy7bok.jpg" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Data Wrangler &amp;amp; Snowflake
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Data Wrangler enables customers to browse and access data stores across Amazon S3, Amazon Athena, Amazon Redshift, and 3rd party data warehouses like Snowflake. &lt;/li&gt;
&lt;li&gt;This hybrid ML pattern provides customers the ability to develop in the cloud while accessing data stored on-premises, as organizations develop their migration plans.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Train in the cloud, to deploy ML models on-premises
&lt;/h3&gt;

&lt;p&gt;If you are deploying on-premises, you need to develop and host your local webserver. We strongly recommend you decouple* hosting your model artifact from your application.&lt;/p&gt;

&lt;p&gt;1-  Amazon SageMaker lets you specify any type of model framework, version, or output artifact you need to.&lt;br&gt;
2-  You’ll find all model artifacts wrapped as tar.gz archives after training jobs, as this compressed file format saves on job time and data costs.&lt;br&gt;
3-  If you are using your own image, you will need to own updating that image as the software version, such as TensorFlow or PyTorch, undergoes potentially major changes over time.&lt;br&gt;
4-  Lastly, keep in mind that it is an unequivocal best practice to decouple hosting your ML model from hosting your application.&lt;br&gt;
A key step in your innovation flywheel is that once you use dedicated resources to host your ML model, specifically ones that are separated from your application, this greatly simplifies your process to push better models. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Advantages&lt;/strong&gt;&lt;br&gt;
1-  Can use SageMaker Neo to compile your model for a target device • Feels like you have more control up front.&lt;br&gt;
2-  Taking advantage of cloud-based training and tuning features, like spot, debugger, model and data-parallel, Bayesian optimization, etc.&lt;br&gt;
3-  Can enable your organization to progress on their cloud migration plan while the application development moves to the cloud.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Disadvantages&lt;/strong&gt;&lt;br&gt;
1-  Need to develop, manage, maintain, and respond to operational issues with locally managed web servers.&lt;br&gt;
2-  Own the burden of building and maintaining up-to-date versions of model software frameworks, such as TensorFlow or PyTorch.&lt;br&gt;
3-  Bigger risk of tightly coupling compute for your model with computing for your application, making it more challenging for you to deploy new models and new features to your application over time.&lt;br&gt;
4-  Need to develop your own data-drift detection and model monitor capabilities.&lt;br&gt;
5-  Not taking advantage of cloud-based features for global deployment, see next section for more details.&lt;br&gt;
6-  Need to develop your own application monitoring pipeline that extracts key metrics, business details, and model responses, to share with business and data science stakeholders.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F55vh3j8hn89gijmruxm5.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F55vh3j8hn89gijmruxm5.jpg" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;When to move&lt;/strong&gt;&lt;br&gt;
1-  When your ability to deploy new applications on-premises is hindered by your need to procure, provision, and manage local infrastructure.&lt;br&gt;
2-  When your model is losing accuracy and/or performance over time, due to your inability to quickly retrain and redeploy.&lt;br&gt;
3-  When the cost of monitoring, updating, maintaining, and troubleshooting.&lt;/p&gt;

&lt;h3&gt;
  
  
  Monitor ML models deployed on-premises with SageMaker Edge Manager
&lt;/h3&gt;

&lt;ol&gt;
&lt;li&gt;  SageMaker Edge Manage makes it easy for customers to manage ML models deployed on Windows, Linux, or ARM-based compute environments.&lt;/li&gt;
&lt;li&gt; Install the edge manager agent onto the CPU of your intended device, and leverage AWS IOT Core or another transfer method to download the model to the device, and execute local inferencing.&lt;/li&gt;
&lt;li&gt; Edge Manage simplifies the monitoring and updating of these models by bringing the control plane up to the cloud.&lt;/li&gt;
&lt;li&gt; You can bring your own monitoring algorithm to the service and trigger retraining pipelines as necessary, using the service the redeploy that model back down to the local device.&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Hybrid patterns for deployment
&lt;/h2&gt;

&lt;p&gt;In this pattern, we focus mostly on hosting the model in the cloud but interacting with applications that may be hosted on-premises.&lt;br&gt;
Hosting in the cloud to applications on-premises can enable the data scientists, while patterns for hosting ML models via Lambda at the Edge, Outposts, Local Zones, and Wavelength.&lt;/p&gt;

&lt;h3&gt;
  
  
  Serve models in the cloud to applications hosted on-premises
&lt;/h3&gt;

&lt;p&gt;The most common use case for a hybrid pattern like this is enterprise migrations.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv45l8x56p19t6joze3ug.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv45l8x56p19t6joze3ug.jpg" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Advantages&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; Are deploying ML models to application consumers.&lt;/li&gt;
&lt;li&gt; Can use custom hardware for ultra-low response times with AWS Inferentia.&lt;/li&gt;
&lt;li&gt; Can serve thousands of models on the cheap with multi-model endpoints.&lt;/li&gt;
&lt;li&gt; Can deploy complex feature transforms with inference pipelines.&lt;/li&gt;
&lt;li&gt; Can use built-in autoscaling and model monitor.&lt;/li&gt;
&lt;li&gt; Can easily develop retrain and tune pipelines.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Disadvantages&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; Risk of your local application infrastructure maintenance hindering the speed of your model development.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;When to move&lt;/strong&gt;: When your ability to deploy new applications on-premises is hindered by your need to procure, provision, and manage local infrastructure.&lt;/p&gt;

&lt;h2&gt;
  
  
  Host ML Models with Lambda at Edge to applications on-premises
&lt;/h2&gt;

&lt;p&gt;1-  This pattern takes advantage of a key capability of the AWS global network – the content delivery network known as Amazon CloudFront.&lt;br&gt;
2-  Once you’ve set your Lambda function to trigger off of CloudFront, you’re telling the service to replicate that function across all available regions and points of presence. This can take up to 8 minutes to replicate and become available.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Advantages&lt;/strong&gt;&lt;br&gt;
1-  Can use CloudFront, opening you up to serving on hundreds of points of presence around the world, and saving you from having to manage these. &lt;br&gt;
2-  Works nicely with Docker images on SageMaker, because you can create from ECR.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Disadvantages&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; Can’t use GPU’s, so you may introduce a bit of latency in some cases, particularly where customers may be better served by an ML model on Inferentia hosted in a nearby AWS Region.&lt;/li&gt;
&lt;li&gt; Lambda has a hard limit on the largest amount of memory you can allocate to a function, which is 10.24GB. For many “classical” ML models, such as XGBoost or linear regressors, 10GB is more than sufficient. However, for some more complex deep learning models, especially those in the 10’s to 100’s of billions of parameters, 10GB is woefully short in terms of RAM.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;When to move&lt;/strong&gt;&lt;br&gt;
1-  When you need more advanced drift and monitoring capabilities.&lt;br&gt;
2-  When you want to introduce complex feature transforms, such as with inference pipelines on SageMaker.&lt;br&gt;
3-  When you want to serve thousands of models per use case.&lt;/p&gt;

&lt;h3&gt;
  
  
  Training with a 3rd party SaaS provider to host in the cloud
&lt;/h3&gt;

&lt;p&gt;1-  Ensure your provider allows the export of proprietary software frameworks, such as jars, bundles, images, etc. Follow w steps to create a Docker file using that software framework , port into the Elastic Container Registry, and host on SageMaker.&lt;/p&gt;

&lt;h2&gt;
  
  
  Control plane patterns for hybrid ML
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt; One such common control plane is Kubeflow in conjunction with EKS Anywhere. EKS Anywhere is currently in private preview, anticipated to come online in 2021.&lt;/li&gt;
&lt;li&gt; SageMaker offers a native approach for workflow orchestration, known as SageMaker Pipelines. SageMaker Pipelines is ideal for advanced SageMaker users, especially those who are already onboarded to the IDE SageMaker Studio. &lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Auxiliary services for hybrid ML patterns
&lt;/h2&gt;

&lt;h3&gt;
  
  
  AWS Outposts
&lt;/h3&gt;

&lt;p&gt;1-  Order AWS Outposts and Amazon will ship, install, and manage these resources for you. You can connect to these resources however you prefer, and manage them from the cloud.&lt;br&gt;
2-  You can deploy ML models via ECS to serve inference with ultra-low latency within your data centers, using AWS Outposts. You can also use ECS for model training, integrating with SageMaker in the cloud and ECS on Outposts.&lt;br&gt;
3-  Outposts help solve cases where customers want to build applications in countries where there is not currently an AWS Region, or for regulations that have strict data residency requirements, like online gambling and sports betting.&lt;/p&gt;

&lt;h3&gt;
  
  
  AWS Inferentia
&lt;/h3&gt;

&lt;ol&gt;
&lt;li&gt; AWS Inferentia provides ease of accessing custom hardware for ML inferencing.&lt;/li&gt;
&lt;li&gt; The majority of Alexa operates in a hybrid ML pattern, hosting models on AWS Inferentia and serving hundreds of millions of Alexa-enabled devices around the world. Using AWS Inferentia, Alexa was able to reduce its cost of hosting by 25%.&lt;/li&gt;
&lt;li&gt; You can use SageMaker’s managed deep learning containers to train your ML models, compile them for Inferentia with Neo, host on the cloud, and develop your retrain and tune pipeline as usual.&lt;/li&gt;
&lt;/ol&gt;

&lt;h3&gt;
  
  
  AWS Direct Connect
&lt;/h3&gt;

&lt;p&gt;Ability to establish a private connection between your on-premises resources and your data center. Remember to establish a redundant link, as wires do go south!&lt;/p&gt;

&lt;h2&gt;
  
  
  Hybrid ML Use Cases
&lt;/h2&gt;

&lt;p&gt;1-  Enterprise Migrations&lt;br&gt;
2-  Manufacturing&lt;br&gt;
3-  Gaming&lt;br&gt;
4-  Mobile application development&lt;br&gt;
5-  AI-enhanced media and content creation&lt;br&gt;
6-  Autonomous Vehicles&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;In this document, we explored hybrid ML patterns across the entire ML lifecycle. We looked at developing locally while training and deploying in the cloud. We discussed patterns for training locally to deploy on the cloud and even to host ML models in the cloud to serve applications on-premises.&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>aws</category>
      <category>sagemaker</category>
      <category>awscommunitybuilder</category>
    </item>
    <item>
      <title>Will Amazon Sagemaker ever rule the world?</title>
      <dc:creator>Heba K. Ahmed</dc:creator>
      <pubDate>Sun, 11 Jul 2021 19:39:35 +0000</pubDate>
      <link>https://dev.to/aws-builders/will-amazon-sagemaker-ever-rule-the-world-1b90</link>
      <guid>https://dev.to/aws-builders/will-amazon-sagemaker-ever-rule-the-world-1b90</guid>
      <description>&lt;p&gt;The phrase "never normal" became commonplace in 2020. And, as with most dramatic upheavals, some leaders' first reaction is to focus on survival. Customers may learn how AI and machine learning spurs corporate innovation, enhances consumer experiences, and boosts profits. Whether it's improving consumer experiences, making sophisticated real-time recommendations, speeding up new product development, increasing employee productivity, or saving expenses and decreasing fraud, businesses are increasingly turning to AI and machine learning to solve problems.&lt;br&gt;
AI and ML hold the promise of transforming industries, increasing efficiencies, and driving innovation. The key to machine learning success is scale.&lt;br&gt;
In 2018, Amazon Sagemaker was a highly scalable machine learning and deep learning service that supported 11 of its own algorithms as well as any others you provided. You had to conduct your own ETL and feature engineering because hyperparameter optimization was still in preview.&lt;/p&gt;

&lt;p&gt;Since then, SageMaker has grown in scope, adding IDEs (SageMaker Studio) and automatic machine learning (SageMaker Autopilot) to the core notebooks, as well as a slew of new services to the whole ecosystem, as illustrated in the diagram below.&lt;/p&gt;

&lt;p&gt;This ecosystem helps with machine learning from start to finish, from model development through training and tuning to deployment and management.&lt;br&gt;
&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F06quo0kb972ocdyiir9m.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F06quo0kb972ocdyiir9m.jpg" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;In a nutshell, Amazon SageMaker is a set of libraries and interfaces that make building and deploying machine learning models easier. It's also important to remember that the SageMaker platform is made up of a variety of products and services that may be customized to meet your specific needs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Benefits of Amazon SageMaker&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;There are several benefits to using Amazon SageMaker, but here are four areas that can be used as examples to show the strength of the platform.&lt;/p&gt;

&lt;p&gt;Those areas are accessibility, customization, scalability, and efficiency.&lt;/p&gt;

&lt;p&gt;The Dashboard on the AWS SageMaker console shows all the different areas and all the different services that can be accessed through the web interface. The tools and services can also be managed from an external machine using the Amazon SageMaker CLI or command-line interface.&lt;/p&gt;

&lt;p&gt;There are many ways to customize this process. And the approach might seem daunting, but once you run through a basic pattern, the opportunities for creation are pretty much endless.&lt;/p&gt;

&lt;p&gt;For your hosted models, Amazon SageMaker enables automated scaling (autoscaling). In reaction to changes in your workload, autoscaling dynamically alters the number of instances supplied for a model. Autoscaling brings extra instances online as the workload grows.&lt;br&gt;
Autoscaling removes the unnecessary instances as the workload drops, so you don't have to pay for instances you're not using.&lt;/p&gt;

&lt;p&gt;Amazon Sagemake provides a compute instance with a jupyter notebook running the r/python kernel that we may choose based on our Data engineering requirements on demand. Using the traditional way, we may display, process, clean, and transform the data into our necessary forms using libraries such as Pandas or Matplotlib.&lt;br&gt;
We can train the models utilizing a different compute instance dependent on the model's computing needs, such as memory-optimized or GPU enabled, after data engineering.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsq9gu4g4hpxbtmm1f7x4.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsq9gu4g4hpxbtmm1f7x4.gif" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;For a range of models, take advantage of sensible default high-performance hyperparameter tweaking settings. Use performance-optimized or algorisms from AWS's extensive library, or bring our own algorisms in using industry-standard containers. Also, make the trained model available as an API, deploying it on a new computer instance to fulfill business needs and scalability.&lt;/p&gt;

&lt;p&gt;And the entire process of providing hardware instances, performing high-capacity data jobs, coordinating the entire flow with simple commands while removing manual complications, and eventually enabling serverless elastic deployment works with only a few lines of code while being cost-effective. Sagemaker is a game-changing enterprise solution.&lt;/p&gt;

&lt;p&gt;For most data scientists who want to achieve a genuinely end-to-end solution, Amazon Sagemaker is a wonderful offer. It handles the abstraction as well as the numerous software development talents required to complete the assignment while remaining extremely effective, versatile, and cost-efficient. Most significantly, it allows you to concentrate on the core ML experiments while supplementing the necessary abilities with simple abstractions that are similar to our current methodology.&lt;br&gt;
&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fex36sgs5wa4kc23bfu7z.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fex36sgs5wa4kc23bfu7z.jpg" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>aws</category>
      <category>machinelearning</category>
    </item>
  </channel>
</rss>
