{"id":1958,"date":"2022-03-10T19:51:54","date_gmt":"2022-03-10T19:51:54","guid":{"rendered":"https:\/\/salarydistribution.com\/machine-learning\/2022\/03\/10\/build-share-deploy-how-business-analysts-and-data-scientists-achieve-faster-time-to-market-using-no-code-ml-and-amazon-sagemaker-canvas\/"},"modified":"2022-03-10T19:51:54","modified_gmt":"2022-03-10T19:51:54","slug":"build-share-deploy-how-business-analysts-and-data-scientists-achieve-faster-time-to-market-using-no-code-ml-and-amazon-sagemaker-canvas","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2022\/03\/10\/build-share-deploy-how-business-analysts-and-data-scientists-achieve-faster-time-to-market-using-no-code-ml-and-amazon-sagemaker-canvas\/","title":{"rendered":"Build, Share, Deploy: how business analysts and data scientists achieve faster time-to-market using no-code ML and Amazon SageMaker Canvas"},"content":{"rendered":"<div id=\"\">\n<p>Machine learning (ML) helps organizations increase revenue, drive business growth, and reduce cost by optimizing core business functions across multiple verticals, such as demand forecasting, credit scoring, pricing, predicting customer churn, identifying next best offers, predicting late shipments, and improving manufacturing quality. Traditional ML development cycles take months and require scarce data science and ML engineering skills. Analysts\u2019 ideas for ML models often sit in long backlogs awaiting data science team bandwidth, while data scientists focus on more complex ML projects requiring their full skillset.<\/p>\n<p>To help break this stalemate, we\u2019ve <a href=\"https:\/\/aws.amazon.com\/blogs\/aws\/announcing-amazon-sagemaker-canvas-a-visual-no-code-machine-learning-capability-for-business-analysts\/\" target=\"_blank\" rel=\"noopener noreferrer\">introduced Amazon SageMaker Canvas<\/a>, a no-code ML solution that can help companies accelerate delivery of ML solutions down to hours or days. SageMaker Canvas enables analysts to easily use available data in data lakes, data warehouses, and operational data stores; build ML models; and use them to make predictions interactively and for batch scoring on bulk datasets\u2014all without writing a single line of code.<\/p>\n<p>In this post, we show how SageMaker Canvas enables collaboration between data scientists and business analysts, achieving faster time to market and accelerating the development of ML solutions. Analysts get their own no-code ML workspace in SageMaker Canvas, without having to become an ML expert. Analysts can then share their models from Canvas with a few clicks, which data scientists will be able to work with in <a href=\"https:\/\/aws.amazon.com\/sagemaker\/studio\/\" target=\"_blank\" rel=\"noopener noreferrer\">Amazon SageMaker Studio<\/a>, an end-to-end ML integrated development environment (IDE). By working together, business analysts can bring their domain knowledge and the results of the experimentation, while data scientists can effectively create pipelines and streamline the process.<\/p>\n<p>Let\u2019s deep dive on what the workflow would look like.<\/p>\n<h2>Business analysts build a model, then share it<\/h2>\n<p>To understand how SageMaker Canvas simplifies collaboration between business analysts and data scientists (or ML engineers), we first approach the process as a business analyst. Before you get started, refer to <a href=\"https:\/\/aws.amazon.com\/blogs\/aws\/announcing-amazon-sagemaker-canvas-a-visual-no-code-machine-learning-capability-for-business-analysts\/\" target=\"_blank\" rel=\"noopener noreferrer\">Announcing Amazon SageMaker Canvas \u2013 a Visual, No Code Machine Learning Capability for Business Analysts<\/a> for instructions on building and testing the model with SageMaker Canvas.<\/p>\n<p>For this post, we use a modified version of the <a href=\"https:\/\/www.kaggle.com\/mlg-ulb\/creditcardfraud\" target=\"_blank\" rel=\"noopener noreferrer\">Credit Card Fraud Detection dataset<\/a> from Kaggle, a well-known dataset for a binary classification problem. The dataset is originally highly unbalanced\u2014it has very few entries classified as a negative class (anomalous transactions). Regardless of the target feature distribution, we can still use this dataset, because SageMaker Canvas handles this imbalance as it trains and tunes a model automatically. This dataset consists of about 9 million cells. You can also download a <a href=\"https:\/\/static.us-east-1.prod.workshops.aws\/public\/76c8599c-2862-4690-9120-e759b53f8636\/static\/datasets\/creditcard_under_smote.csv\" target=\"_blank\" rel=\"noopener noreferrer\">reduced version of this dataset<\/a>. The dataset size is much smaller, at around 500,000 cells, because it has been randomly under-sampled and then over-sampled with the SMOTE technique to ensure that as little information as possible is lost during this process. Running an entire experiment with this reduced dataset costs you $0 under the SageMaker Canvas Free Tier.<\/p>\n<p>After the model is built, analysts can use it to make predictions directly in Canvas for either individual requests, or for an entire input dataset in bulk.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-33534\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2022\/02\/28\/predict.png\" alt=\"Use the trained model to generate predictions\" width=\"512\" height=\"275\" data-darkreader-inline-border-top=\"\" data-darkreader-inline-border-right=\"\" data-darkreader-inline-border-bottom=\"\" data-darkreader-inline-border-left=\"\"><\/p>\n<p>Models built with Canvas Standard Build can also be easily shared at a click of a button with data scientists and ML engineers that use SageMaker Studio. This allows a data scientist to validate the performance of the model you\u2019ve built and provide feedback. ML engineers can pick up your model and integrate it with existing workflows and products available to your company and your customers. Note that, at the time of writing, it\u2019s not possible to share a model built with Canvas Quick Build, or a time series forecasting model.<\/p>\n<p>Sharing a model via the Canvas UI is straightforward:<\/p>\n<ol>\n<li>On the page showing the models that you\u2019ve created, choose a model.<\/li>\n<li>Choose <strong>Share<\/strong>.<img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-33533\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2022\/02\/28\/model-trained.png\" alt=\"Share the trained model from the Analyze tab\" width=\"635\" height=\"212\" data-darkreader-inline-border-top=\"\" data-darkreader-inline-border-right=\"\" data-darkreader-inline-border-bottom=\"\" data-darkreader-inline-border-left=\"\"><\/li>\n<li>Choose one or more versions of the model that you want to share.<\/li>\n<li>Optionally, include a note giving more context about the model or the help you\u2019re looking for.<\/li>\n<li>Choose <strong>Create SageMaker Studio Link<\/strong>.<img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-33539\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2022\/02\/28\/share-model.png\" alt=\"Share the model with SageMaker Studio\" width=\"568\" height=\"589\" data-darkreader-inline-border-top=\"\" data-darkreader-inline-border-right=\"\" data-darkreader-inline-border-bottom=\"\" data-darkreader-inline-border-left=\"\"><\/li>\n<li>Copy the generated link.<img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-33537\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2022\/02\/28\/share-link.png\" alt=\"Copy the generated link\" width=\"791\" height=\"100\" data-darkreader-inline-border-top=\"\" data-darkreader-inline-border-right=\"\" data-darkreader-inline-border-bottom=\"\" data-darkreader-inline-border-left=\"\"><\/li>\n<\/ol>\n<p>And that\u2019s it! You can now share the link with your colleagues via Slack, email, or any other method of your preference. The data scientist needs to be in the same SageMaker Studio domain in order to access your model, so make sure this is the case with your organization admin.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-33538\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2022\/02\/28\/share-message.png\" alt=\"Share the model by sending a Slack message or an email\" width=\"797\" height=\"168\" data-darkreader-inline-border-top=\"\" data-darkreader-inline-border-right=\"\" data-darkreader-inline-border-bottom=\"\" data-darkreader-inline-border-left=\"\"><\/p>\n<h2>Data scientists access the model information from SageMaker Studio<\/h2>\n<p>Now, let\u2019s play the role of a data scientist or ML engineer, and see things from their point of view using SageMaker Studio.<\/p>\n<p>The link shared by the analyst takes us into SageMaker Studio, the first cloud-based IDE for the end-to-end ML workflow.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-33540\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2022\/02\/28\/studio-tab-model.png\" alt=\"Show the model overview as seen in SageMaker Studio\" width=\"800\" height=\"529\" data-darkreader-inline-border-top=\"\" data-darkreader-inline-border-right=\"\" data-darkreader-inline-border-bottom=\"\" data-darkreader-inline-border-left=\"\"><\/p>\n<p>The tab opens automatically, and shows an overview of the model created by the analyst in SageMaker Canvas. You can quickly see the model\u2019s name, the ML problem type, the model version, and which user created the model (under the field Canvas user ID). You also have access to details about the input dataset and the best model that SageMaker was able to produce. We will dive into that later in the post.<\/p>\n<p>On the <strong>Input Dataset<\/strong> tab, you can also see the data flow from the source to the input dataset. In this case, only one data source is used and no join operations have been applied, so a single source is shown. You can analyze statistics and details about the dataset by choosing <strong>Open data exploration notebook<\/strong>. This notebook lets you explore the data that was available before training the model, and contains an analysis of the target variable, a sample of the input data, statistics and descriptions of columns and rows, as well as other useful information for the data scientist to know more about the dataset. To learn more about this report, refer to <a href=\"https:\/\/docs.aws.amazon.com\/sagemaker\/latest\/dg\/autopilot-data-exploration-report.html\" target=\"_blank\" rel=\"noopener noreferrer\">Data exploration report<\/a>.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-33532\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2022\/02\/28\/model-overview.png\" alt=\"Show the model overview, with the completed jobs and the job information\" width=\"847\" height=\"333\" data-darkreader-inline-border-top=\"\" data-darkreader-inline-border-right=\"\" data-darkreader-inline-border-bottom=\"\" data-darkreader-inline-border-left=\"\"><\/p>\n<p>After analyzing the input dataset, let\u2019s move on to the second tab of the model overview, <strong>AutoML Job<\/strong>. This tab contains a description of the AutoML job when you selected the Standard Build option in SageMaker Canvas.<\/p>\n<p>The AutoML technology underneath SageMaker Canvas eliminates the heavy lifting of building ML models. It automatically builds, trains, and tunes the best ML model based on your data by using an automated approach, while allowing you to maintain full control and visibility. This visibility on the generated candidate models as well as the hyper-parameters used during the AutoML process is contained in the <a href=\"https:\/\/docs.aws.amazon.com\/sagemaker\/latest\/dg\/autopilot-candidate-generation-notebook.html\" target=\"_blank\" rel=\"noopener noreferrer\">candidate generation notebook<\/a>, which is available on this tab.<\/p>\n<p>The <strong>AutoML Job <\/strong>tab also contains a list of every model built as part of the AutoML process, sorted by the F1 objective metric. To highlight the best model out of the training jobs launched, a tag with a green circle is used in the <strong>Best Model<\/strong> column. You can also easily visualize other metrics used during the training and evaluation phase, such as the accuracy score and the Area Under the Curve (AUC). To learn more about the models that you can train during an AutoML job and the metrics used for evaluating the performances of the trained model, refer to <a href=\"https:\/\/docs.aws.amazon.com\/sagemaker\/latest\/dg\/autopilot-model-support-validation.html\" target=\"_blank\" rel=\"noopener noreferrer\">Model support, metrics, and validation<\/a>.<\/p>\n<p>To learn more about the model, you can now right-click the best model and choose <strong>Open in model details<\/strong>. Alternatively, you can choose the <strong>Best model<\/strong> link at the top of the <strong>Model overview<\/strong> section you first visited.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-33531\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2022\/02\/28\/model-details.png\" alt=\"Model details with feature importances and metrics\" width=\"804\" height=\"727\" data-darkreader-inline-border-top=\"\" data-darkreader-inline-border-right=\"\" data-darkreader-inline-border-bottom=\"\" data-darkreader-inline-border-left=\"\"><\/p>\n<p>The model details page contains a plethora of useful information regarding the model that performed best with this input data. Let\u2019s first focus on the summary at the top of the page. The preceding example screenshot shows that, out of hundreds of model training runs, an XGBoost model performed best on the input dataset. At the time of this writing, SageMaker Canvas can train three types of ML algorithms: linear learner, XGBoost, and a multilayer perceptron (MLP), each with a wide variety of preprocessing pipelines and hyper-parameters. To learn more about each algorithm, refer to <a href=\"https:\/\/docs.aws.amazon.com\/sagemaker\/latest\/dg\/autopilot-model-support-validation.html#autopilot-algorithm-suppprt\" target=\"_blank\" rel=\"noopener noreferrer\">supported algorithms page<\/a>.<\/p>\n<p>SageMaker also includes an explanatory functionality thanks to a scalable and efficient implementation of <a href=\"https:\/\/papers.nips.cc\/paper\/2017\/file\/8a20a8621978632d76c43dfd28b67767-Paper.pdf\" target=\"_blank\" rel=\"noopener noreferrer\">KernelSHAP<\/a>, based on the concept of a Shapley value from the field of cooperative game theory that assigns to each feature an importance value for a particular prediction. This allows for transparency about how the model arrived at its predictions, and it\u2019s very useful to define feature importance. A complete explainability report including feature importance is downloadable in PDF, notebook, or raw data format. In that report, a wider set of metrics are shown as well as a full list of hyper-parameters used during the AutoML job. To learn more about how SageMaker provides integrated explainability tools for AutoML solutions and standard ML algorithms, see <a href=\"https:\/\/aws.amazon.com\/blogs\/machine-learning\/use-integrated-explainability-tools-and-improve-model-quality-using-amazon-sagemaker-autopilot\/\" target=\"_blank\" rel=\"noopener noreferrer\">Use integrated explainability tools and improve model quality using Amazon SageMaker Autopilot<\/a>.<\/p>\n<p>Finally, the other tabs in this view show information about performance details (confusion matrix, precision recall curve, ROC curve), artifacts used for inputs and generated during the AutoML job, and network details.<\/p>\n<p>At this point, the data scientist has two choices: directly deploy the model, or create a training pipeline that can be scheduled or triggered manually or automatically. The following sections provide some insights into both options.<\/p>\n<h3>Deploy the model directly<\/h3>\n<p>If the data scientist is satisfied with the results obtained by the AutoML job, they can directly deploy the model from the <strong>Model Details<\/strong> page. It\u2019s as simple as choosing <strong>Deploy model<\/strong> next to the model name.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter wp-image-33530 size-full\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2022\/02\/28\/model-details-2.png\" alt=\"Additional model details, from where to deploy the model\" width=\"675\" height=\"266\" data-darkreader-inline-border-top=\"\" data-darkreader-inline-border-right=\"\" data-darkreader-inline-border-bottom=\"\" data-darkreader-inline-border-left=\"\"><\/p>\n<p>SageMaker shows you two options for deployment: a real-time endpoint, powered by <a href=\"https:\/\/docs.aws.amazon.com\/sagemaker\/latest\/dg\/how-it-works-deployment.html\" target=\"_blank\" rel=\"noopener noreferrer\">Amazon SageMaker endpoints<\/a>, and batch inference, powered by <a href=\"https:\/\/docs.aws.amazon.com\/sagemaker\/latest\/dg\/how-it-works-batch.html\" target=\"_blank\" rel=\"noopener noreferrer\">Amazon SageMaker batch transform<\/a>.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-33536\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2022\/02\/28\/predictors.png\" alt=\"Option to launch prediction from AutoML\" width=\"1512\" height=\"629\" data-darkreader-inline-border-top=\"\" data-darkreader-inline-border-right=\"\" data-darkreader-inline-border-bottom=\"\" data-darkreader-inline-border-left=\"\"><\/p>\n<p>SageMaker also provides other modes of inference. To learn more, see <a href=\"https:\/\/docs.aws.amazon.com\/sagemaker\/latest\/dg\/deploy-model.html\" target=\"_blank\" rel=\"noopener noreferrer\">Deploy Models for Inference<\/a>.<\/p>\n<p>To enable the real-time predictions mode, you simply give the endpoint a name, an instance type, and an instance count. Because this model doesn\u2019t require heavy compute resources, you can use a CPU-based instance with an initial count of 1. You can learn more about the different kind of instances available and their specs on the <a href=\"https:\/\/aws.amazon.com\/sagemaker\/pricing\/\" target=\"_blank\" rel=\"noopener noreferrer\">Amazon SageMaker Pricing page<\/a> (in the <strong>On-Demand Pricing<\/strong> section, choose the <strong>Real-Time Inference<\/strong> tab). If you don\u2019t know which instance you should choose for your deployment, you can also ask SageMaker to find the best one for you according to your KPIs by using the <a href=\"https:\/\/docs.aws.amazon.com\/sagemaker\/latest\/dg\/inference-recommender.html\" target=\"_blank\" rel=\"noopener noreferrer\">SageMaker Inference Recommender<\/a>. You can also provide additional optional parameters, regarding whether or not you want to capture request and response data to or from the endpoint. This can prove useful if you\u2019re planning on <a href=\"https:\/\/docs.aws.amazon.com\/sagemaker\/latest\/dg\/how-it-works-model-monitor.html\" target=\"_blank\" rel=\"noopener noreferrer\">monitoring your model<\/a>. You can also choose which content you wish to provide as part of your response\u2014whether it\u2019s just the prediction or the prediction probability, the probability of all classes, and the target labels.<\/p>\n<p>To run a batch scoring job getting predictions for an entire set of inputs at one time, you can launch the batch transform job from the <a href=\"http:\/\/aws.amazon.com\/console\" target=\"_blank\" rel=\"noopener noreferrer\">AWS Management Console<\/a> or via the SageMaker Python SDK. To learn more about batch transform, refer to <a href=\"https:\/\/docs.aws.amazon.com\/sagemaker\/latest\/dg\/batch-transform.html\" target=\"_blank\" rel=\"noopener noreferrer\">Use Batch Transform<\/a> and the example notebooks.<\/p>\n<h3>Define a training pipeline<\/h3>\n<p>ML models can very rarely, if ever, be considered static and unchanging, because they drift from the baseline they\u2019ve been trained on. Real-world data evolves over time, and more patterns and insights emerge from it, which may or may not be captured by the original model trained on historical data. To solve this problem, you can set up a training pipeline that automatically retrains your models with the latest data available.<\/p>\n<p>In defining this pipeline, one of the options of the data scientist is to once again use AutoML for the training pipeline. You can launch an AutoML job programmatically by invoking the create_auto_ml_job() API from the <a href=\"https:\/\/bit.ly\/3HhiDsM\" target=\"_blank\" rel=\"noopener noreferrer\">AWS Boto3 SDK<\/a>. You can call this operation from an <a href=\"http:\/\/aws.amazon.com\/lambda\" target=\"_blank\" rel=\"noopener noreferrer\">AWS Lambda<\/a> function within an <a href=\"http:\/\/aws.amazon.com\/step-functions\" target=\"_blank\" rel=\"noopener noreferrer\">AWS Step Functions<\/a> workflow, or from a LambdaStep in <a href=\"https:\/\/aws.amazon.com\/sagemaker\/pipelines\/\" target=\"_blank\" rel=\"noopener noreferrer\">Amazon SageMaker Pipelines<\/a>.<\/p>\n<p>Alternatively, the data scientist can use the knowledge, artifacts, and hyper-parameters obtained from the AutoML job to define a complete training pipeline. You need the following resources:<\/p>\n<ul>\n<li><strong>The algorithm that worked best for the use case<\/strong> \u2013 You already obtained this information from the summary of the Canvas-generated model. For this use case, it\u2019s the XGBoost built-in algorithm. For instructions on how to use the SageMaker Python SDK to train the XGBoost algorithm with SageMaker, refer to <a href=\"https:\/\/sagemaker.readthedocs.io\/en\/stable\/frameworks\/xgboost\/using_xgboost.html\" target=\"_blank\" rel=\"noopener noreferrer\">Use XGBoost with the SageMaker Python SDK<\/a>.<br \/><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-33526\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2022\/02\/28\/completed-job.png\" alt=\"Information about the algorithm that was trained with the Canvas job\" width=\"477\" height=\"206\" data-darkreader-inline-border-top=\"\" data-darkreader-inline-border-right=\"\" data-darkreader-inline-border-bottom=\"\" data-darkreader-inline-border-left=\"\"><\/li>\n<\/ul>\n<ul>\n<li><strong>The hyperparameters derived by the AutoML job<\/strong> \u2013 These are available in the <strong>Explainability<\/strong> section. You can use them as inputs when defining the training job with the SageMaker Python SDK.<br \/><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-33529\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2022\/02\/28\/hyperparameters.png\" alt=\"The model hyperparameters\" width=\"814\" height=\"397\" data-darkreader-inline-border-top=\"\" data-darkreader-inline-border-right=\"\" data-darkreader-inline-border-bottom=\"\" data-darkreader-inline-border-left=\"\"><\/li>\n<\/ul>\n<ul>\n<li><strong>The feature engineering code provided in the Artifacts section<\/strong> \u2013 You can use this code both for preprocessing the data before training (for example, via Amazon SageMaker Processing), or before inference (for example, as part of a SageMaker inference pipeline).<br \/><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-33528\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2022\/02\/28\/feature-eng-code.png\" alt=\"The S3 URI of the Feature Engineering Code\" width=\"977\" height=\"143\" data-darkreader-inline-border-top=\"\" data-darkreader-inline-border-right=\"\" data-darkreader-inline-border-bottom=\"\" data-darkreader-inline-border-left=\"\"><\/li>\n<\/ul>\n<p>You can combine these resources as part of a SageMaker pipeline. We omit the implementation details in this post\u2014stay tuned for more content coming on this topic.<\/p>\n<h2>Conclusion<\/h2>\n<p>SageMaker Canvas lets you use ML to generate predictions without needing to write any code. A business analyst can autonomously start using it with local datasets, as well as data already stored on <a href=\"http:\/\/aws.amazon.com\/s3\" target=\"_blank\" rel=\"noopener noreferrer\">Amazon Simple Storage Service<\/a> (Amazon S3), <a href=\"http:\/\/aws.amazon.com\/redshift\" target=\"_blank\" rel=\"noopener noreferrer\">Amazon Redshift<\/a>, or Snowflake. With just a few clicks, they can prepare and join their datasets, analyze estimated accuracy, verify which columns are impactful, train the best performing model, and generate new individual or batch predictions, all without any need for pulling in an expert data scientist. Then, as needed, they can share the model with a team of data scientists or MLOps engineers, who import the models into SageMaker Studio, and work alongside the analyst to deliver a production solution.<\/p>\n<p>Business analysts can independently gain insights from their data without having a degree in ML, and without having to write a single line of code. Data scientists can now have additional time to work on more challenging projects that can better use their extensive knowledge of AI and ML.<\/p>\n<p>We believe this new collaboration opens the door to building many more powerful ML solutions for your business. You now have analysts producing valuable business insights, while letting data scientists and ML engineers help refine, tune, and extend as needed.<\/p>\n<h2>Additional Resources<\/h2>\n<hr>\n<h3>About the Authors<\/h3>\n<p><a href=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2021\/12\/10\/David-G.png\"><img decoding=\"async\" loading=\"lazy\" class=\"wp-image-31675 size-full alignleft\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2021\/12\/10\/David-G.png\" alt=\"\" width=\"100\" height=\"105\"><\/a><strong>Davide Gallitelli<\/strong> is a Specialist Solutions Architect for AI\/ML in the EMEA region. He is based in Brussels and works closely with customers throughout Benelux. He has been a developer since he was very young, starting to code at the age of 7. He started learning AI\/ML at university, and has fallen in love with it since then.<\/p>\n<p><strong><a href=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2019\/02\/21\/mark-roy-100.jpg\"><img decoding=\"async\" loading=\"lazy\" class=\"size-full wp-image-7780 alignleft\" src=\"https:\/\/d2908q01vomqb2.cloudfront.net\/f1f836cb4ea6efb2a0b1b99f41ad8b103eff4b59\/2019\/02\/21\/mark-roy-100.jpg\" alt=\"\" width=\"100\" height=\"121\"><\/a>Mark Roy<\/strong>\u00a0is a Principal Machine Learning Architect for AWS, helping customers design and build AI\/ML solutions. Mark\u2019s work covers a wide range of ML use cases, with a primary interest in computer vision, deep learning, and scaling ML across the enterprise. He has helped companies in many industries, including insurance, financial services, media and entertainment, healthcare, utilities, and manufacturing. Mark holds six AWS certifications, including the ML Specialty Certification. Prior to joining AWS, Mark was an architect, developer, and technology leader for over 25 years, including 19 years in financial services.<\/p>\n<p>       <!-- '\"` -->\n      <\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/aws.amazon.com\/blogs\/machine-learning\/build-share-deploy-how-business-analysts-and-data-scientists-achieve-faster-time-to-market-using-no-code-ml-and-amazon-sagemaker-canvas\/<\/p>\n","protected":false},"author":0,"featured_media":1959,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/1958"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=1958"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/1958\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/1959"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=1958"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=1958"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=1958"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}