Skip to main content
Version: v3.0 print this page

SageMaker Notebooks

Amorphic platform provides a way to host Jupyter/IPython notebooks, which are interactive, web-based environments that allow users to create and share documents that contain live code, equations, visualizations, and narrative text.

Notebook Operations

Amorphic provides the following operations for Notebook Data Labs :

OperationDescription
Create NotebookCreates a notebook in AWS SageMaker and other necessary AWS resources.
View NotebookView the details of an existing notebook.
Start NotebookStart a stopped notebook.
Stop NotebookStop a running notebook.
Clone NotebookClone a new notebook from an existing notebook.
Delete NotebookDelete an existing notebook.
View/Download Notebook LogsView logs for a notebook instance.

How to Create a Notebook?

Create Notebook

To create an Notebook Data Lab:

  1. Click on + Create Data Lab.
  2. Users will now have an option to either select/upload a template or create from scratch.
  3. Select the Data Lab Type as Notebook.
  4. Fill in the details shown in the table:
AttributeDescription
Data Lab NameGive your notebook data lab a unique name.
DescriptionDescribe the notebook's purpose and relevant details.
KeywordsAdd relevant keywords to the notebook.
Cost TagsSelect the cost tags that need to be attached to the newly created Data Lab for cost monitoring.
Instance TypeChoose the type of ML compute instance to launch the notebook. Users can select from the list of allowed notebook instance types. By default, the instance type used will be ml.t2.medium.
Volume Size (In GB)ML notebook storage volume size in GB. Value should be between 5 GB and 16000 GB. By default, the storage allocated will be 10 GB.
Root AccessSelect this option to enable root access to the notebook instance. By default, this option will be disabled.
Interactive SessionsUser can select this option to enable or disable Glue sessions for the notebook instance. By default, this option will be disabled. When enabling the feature, ensure that you select the system-generated Lifecycle Configuration "enable-glue-session-v2"
Internet AccessThis setting controls whether the notebook instance can access the internet. If you disable this option, the notebook instance can only access resources inside your VPC and will not be able to use Amazon SageMaker training and endpoint services (unless you set up a NAT Gateway in your VPC). By default, this option will be disabled.
Auto StopThis option allows you to save on resource costs by providing a stop time value. The auto stop process will be triggered every hour, looking for any ML Notebooks that need to be notified or stopped, and sends an email when one of the following criteria is met. By default, this option will be disabled.
Shared Resources AccessSelect the shared resources (parameters, shared libraries, domains, etc.) required for the notebook using this option.
Lifecycle ConfigurationName of the Lifecycle Configuration to use for the notebook instance. By default, this option will be set as N/A. For sessions enabled notebooks, Please manually select the system-generated LCC named "enable-glue-sessions-v2"
Datasets AccessSelect datasets with read/write access required for the notebook.
Code RepositorySelect the code repositories required for the notebook. Here, the user can provide access to one Default Code Repository along with multiple Additional Code Repositories.
Note
  • For getting AI powered code suggestions, users can attach the enable-code-whisperer lifecycle configuration to the notebook. Users can then navigate to the Data Lab URL and select the "Resume Suggestions" option from the Code Whisperer extension available in the bottom left corner of the Jupyter Lab UI.
  • While providing Domain Access to the notebook, users will also need to have read/write access to all the datasets (existing and newly created if any) under the selected domains.
  • If you grant read-only access to domains, you won't be able to query Lake Formation datasets within that domain. However, you'll still have access to the files via S3 APIs.
  • To utilize GitLab, GitHub or Bitbucket code repositories in an internet-disabled notebook, it is necessary to whitelist the application proxy with the corresponding repository domains.
  • View type of datasets can be attached only under the Datasets Read Access section.
warning

Enabling root access could compromise NIST compliance. It is recommended to disable root access unless absolutely necessary.

You can set up or create a new notebook instance and use your IPython notebook to perform model training. You can call Python SageMaker SDK to create a training job.

Once a training job is created, you can use the S3 model location information to create a model in the Amorphic portal. For accessing the datasets inside the IPython notebooks, you can check the dataset details for the S3 location information.

For the purpose of creating a SageMaker model in the notebook, the user can use the ml-temp S3 bucket. Amorphic Notebooks have write access to the ml-temp bucket (for example, s3://cdap-us-west-2-484084523624-develop-ml-temp). Please note that this S3 bucket is almost the same as the dataset S3 path, except for the ml-temp at the end. This ml-temp bucket can be used to create a training job and upload a model tar file. This model file location can then be used to create a model using the "Artifact Location" of Amorphic model (see model creation section).

Note

Your data will be stored under ml-temp/<notebook id>/ in the S3 bucket.

You can use the S3 location mentioned here to read the files related to the training dataset and save the output SageMaker model tar file for Amorphic model object creation purpose.

Notebook Details

Amorphic Notebook Data Labs contain all the following information:

TypeDescription
Data Lab NameResource name which uniquely identifies a notebook.
DescriptionA brief description of the notebook.
StatusStatus of the notebook. List of possible statuses include : Pending, InService, Stopping, Stopped and Deleting.
Volume Size (in GB)The size of the ML storage volume attached to the notebook instance (in GB).
Instance TypeInstance type of the SageMaker Notebook instance.
SessionsFlag to identify whether glue sessions are enabled for the notebook instance.
KeywordsKeywords associated with the notebook.
Auto StopStatus of the auto-stop. Ex: Enabled, Disabled.
Remaining TimeAmount of time left for auto-stop (in hr).
Lifecycle ConfigurationName of the lifecycle configuration attached to the notebook instance.
Direct Internet AccessSets whether SageMaker provides internet access to the notebook instance.
Root AccessShows whether root access is enabled for the notebook instance.
Estimated CostApproximate cost incurred since the creation/last modified time.
Linked ResourcesList of all resources linked to the notebook.
Extra Resource AccessList of all resources linked to the notebook.
MessageThe Message field displays information based on the notebook's status.
Data Lab URL (Go to Data Lab)URL to connect to the Jupyter server from notebook instance.
Activity LogsList of all activities performed on the notebook.
Note
  • If the notebook status is failed, the Message field displays failure information.
  • If you do not have all the datasets, code repositories and views access required for the notebook, the Notebook URL will not be displayed and the Message field will show missing resource access information.

Following details are displayed for a Notebook :

Notebook Details

Note
  • Starting from version 1.9, Auto Stop will be replacing Auto terminate. This process will only stop the notebook instance and will not delete the notebook instance.

In the details page, Estimated Cost of the notebook is also displayed to show the approximate cost incurred since the creation/last modified time.

Edit Notebook

The Edit Notebook page follows the same sequence as the create page. Users can choose to update the following configurations :

  • Basic Configuration: You can use this section to update all the basic details of the notebook (description, volume size, lifecycle configuration, etc.) .
  • Auto Stop: You can use this section to auto-stop time or to disable it entirely.
  • Resource Access: You can use this section to update the resources linked to the notebook.
Note
  1. Interactive Sessions and Internet Access cannot be modified once the notebook is created.
  2. The Notebook must be in Stopped state in order to edit compute configurations (Volume Size, Instance Type, Root Access, etc).

Start Notebook

When a notebook is in the stopped state (stopped either manually or through auto stop), user has an option to resume or start the notebook again. The option is available in the UI and can be triggered by pressing the Start Data Lab play button icon situated beside the share button.

When resuming a stopped notebook, if it was created with Auto Stop and the Auto Stop time has already elapsed, users will have two options; either to disable auto stop and resume the notebook, or to update the auto stop time with a newer one.

Note
  • If the auto stop time is set to 10 July, 2024 7:30 PM, the notebook will be stopped at 10 July, 2024 8:00 PM because the stoppage process is scheduled to run at whole hour (UTC)

Stop Notebook

If an notebook is in running state, user is provided with the option to stop it. This feature is useful to reduce incurring costs on running notebooks. The option is available in the UI and can be triggered by pressing the Stop Data Lab button situated beside the share button.

Clone Notebook

To help create notebooks with similar configurations, an option to clone an existing notebook is available. Be sure that the name is different while going through the metadata. The option is available in the UI and can be triggered by pressing the clone option from the drop down that appears after clicking on the three dots situated at the extreme right of the notebook menu bar.

Delete Notebook

If you have sufficient permissions, you can delete the notebook. Deleting a notebook is an asynchronous operation. When triggered, the status will change to Deleting and the notebook will be deleted from AWS SageMaker. Once the notebook is deleted from AWS SageMaker, the associated metadata will also be removed.

Note

The notebook must be in Stopped state in order to perform the delete operation on it.

Auto Stop Notebook Feature

From version 2.7, when a notebook is stopped with the auto-stop feature enabled, users have the option to set a custom auto-stop timer when restarting the notebook. This allows users to choose predefined options, such as 1 hour or 2 hours, or set a custom time after which the notebook will automatically stop. This feature helps users manage the cost of running notebooks, especially during business hours, by tailoring the auto-stop time to their specific needs.

Notebook Start

Auto Stop Time: You can set the maximum auto stop time for the notebook to be less than 168 hours (7 days). Once the current time is found to be greater than the stop time, the notebook will be stopped at the next whole hour. You can also modify the stop time with the maximum time set to less than 168 hours (7 days).

You will receive a notification email when:

  • The auto-stop process trigger runs every hour, and the stoppage time is less than 30 minutes.
  • The auto-stop process was successfully able to stop the notebook after the stop time.
  • The auto-stop process wasn't able to stop the notebook due to some fatal errors.
Note
  • Auto-stop process is scheduled to run every hour on the hour (e.g: 06:00, 07:00, 08:00, 09:00).
  • You will receive an email notification only if you are subscribed to alerts. To enable alerts, refer to Alert Preferences.
  • When the stoppage time elapses, auto stop process will stop the notebook. You need to manually delete the notebook if needed.
warning

Stopped Notebooks will still incur costs. It is recommended to delete the notebook if it is no longer required.

Notebook Logs

Downloadable logs are now available for notebooks. There can be one or three different types of logs available, depending on the type of notebook being used.

For Interactive Sessions enabled notebooks, there are three types of logs: Creation Logs, Start Logs and Jupyter Logs. For Interactive Sessions disabled notebooks, there is only one type of logs, Jupyter Logs.

The following example shows how to access the creation logs for a Interactive Sessions enabled notebook.

Notebook Logs

Update Extra Resource Access

To provide parameter or dataset access to a notebook in large numbers, refer to the documentation on How to provide large number of resources access to an ETL Entity in Amorphic

Glue Session Operations

Amorphic Notebook Data Lab provides operations stated below for an interactive sessions enabled notebook.

OperationDescription
Create Glue SessionCreate a Glue session for a notebook.
Stop Glue SessionStop an existing Glue session for a notebook.
Delete Glue SessionDelete an existing Glue session for a notebook.
View & Download Logs for Glue SessionView and download logs of a Glue session for a notebook.

Before initiating a Glue session, users also have the ability to configure spark settings using imperative magic commands provided by AWS. Below are some beneficial magic commands provided for user reference. For more information visit the link

Magic CommandTypeDescription
%helpshows generic help message with many possible commands and their explanation.
%list_sessionslists all the glue sessions.
%statusshow the current glue session status and configuration.
%stop_sessionstops the current glue sessions straight from the notebook.
%number_of_workersintThe number of workers of a defined worker_type that are allocated when a job runs. worker_type must be set too. The default number_of_workers is 5.
%worker_typeintStandard, G.1X, or G.2X. number_of_workers must be set too. The default worker_type is G.1X.
%iam_roleStringSpecify an IAM role ARN to execute your session with. Default from ~/.aws/configure
%additional_python_modulesListComma separated list of additional Python modules to include in your cluster (can be from PyPI or S3).
%extra_py_filesListComma separated list of additional Python files from Amazon S3.
%extra_jarsListComma-separated list of additional jars to include in the cluster.

Create Glue Session

Create Session

The user must create a notebook instance with Interactive Sessions enabled as described in the steps for creating a notebook.

Once the notebook instance is active (if the notebook status is stopped, the user must start the notebook), the user can find the notebook URL link on the details page (Go to Data Lab) and on opening one the link, the user gets redirected to the Jupyter server. The user then has to create a new Jupyter notebook with Glue Pyspark kernel.

If the user wants to import external libraries or shared libraries into the notebook, use the below Jupyter magics before starting a glue session:

  • %extra_py_files followed by comma separated list of s3 locations for python files.
  • %extra_jars followed by comma separated list of s3 locations for jar files.
  • %additional_python_modules followed by comma separated list such as "awswrangler,pandas==1.5.1,pyarrow==10.0.0" for external python packages.

For additional details on the magic commands available the user can run the magic command %help.

Users can copy the S3 locations from the notebook details page and ETL libraries page for external and shared libraries respectively for glue enabled notebooks.

An example for adding extra python shared library to the session is shown below, for the amorphicutils library.

# copy the path from "Home -> Transformation -> ETL Library -> Details -> (hover over package path and click copy - use the version you need)"
%extra_py_files s3://<amorphic-etl-bucket>/common-libs/<library-id>/libs/python/amorphicutils.zip

Amorphic also supports external libraries which follow the following pattern, again for the amorphicutils library.

%extra_py_files s3://etl-bucket/<notebook-id>/libs/amorphicutils.zip

The following code example will help the user create a glue session for the notebook.

import sys
from awsglue.transforms import *
from awsglue.utils import getResolvedOptions
from pyspark.context import SparkContext
from awsglue.context import GlueContext
from awsglue.job import Job

glueContext = GlueContext(SparkContext.getOrCreate())

With interactive sessions enabled, notebooks offer the capability to execute SQL commands using Magic commands

%%sql
select * from table

Once the glue session is created, all active, stopped and failed sessions of the notebook can be viewed in the sessions tab of the notebook details page.

Stop Glue Session

Stop Session

Delete Glue Session

Delete Session

Note

If an interactive sessions enabled notebook is stopped, all the glue sessions associated with the notebook will also be deleted.

View & Download Logs for Glue Session

Session Logs

Notebook use case

A use case for Amorphic Notebook Data Lab could be, a company that wants to use machine learning to predict customer churn.

The company can set up a new notebook instance on the Amorphic platform and use IPython notebooks to perform model training. They can call the Python SageMaker SDK to create a training job using the customer churn data stored in the S3 bucket.

Once the training job is complete, the company can use the S3 model location information to create a model in the Amorphic portal. They can access the customer churn dataset inside the IPython notebooks using the dataset details and S3 location information. The ml-temp bucket can be used to create the SageMaker model and upload the model tar file, which can then be used to create a model object in the Amorphic portal.

The company can use the S3 location mentioned in the use case to read the files related to the customer churn dataset and save the output SageMaker model tar file for Amorphic model object creation. This allows the company to effectively train a machine learning model to predict customer churn and use it in their business processes.

Note
  • Recently, AWS announced that users of SageMaker service can only access the commercial Anaconda repository without requiring a commercial license until February 1, 2024. After this date, customers will need to determine their own Anaconda license requirements for continued use (refer official AWS documentation). If you are making use of Anaconda channels within your notebooks, you may need to evaluate your specific needs concerning the Anaconda license to ensure compliance and prevent any interruptions in your code (refer Anaconda's Terms of Service).
  • Should you have already procured the necessary licenses, you can make use of the command conda config --add channels defaults to add Anaconda's commercial channels.
  • If you wish to continue using Anaconda without having to procure a license, there are a few free channels that you can use that are run by volunteer communities and offer best effort security (Ex: conda, conda-forge, Bioconda). These can work well for research projects, prototypes, and education. However, they are not recommended for use in sensitive environments.

Query View in Notebook

A view can be queried in notebook using two different methods

  • Using python 'awswrangler' package : If the notebook has internet enabled, then the user can download the awswrangler python package into the notebook using the below command. Do note that any other python packages can be uploaded this way (or via requirements.txt).
# This example sets the packages up for a Glue session with version 4
%additional_python_modules awswrangler==3.4.0,pandas==1.5.3,numpy==1.22.0,pyarrow==8.0.0
%glue_version 4.0

Once the package is installed, the user can query the view by using the following code snippet

#imports
import pandas as pd
import awswrangler as wr
import boto3

# get the aws region
boto3_session = boto3.session.Session()
region_name = boto3_session.region_name

ssm_client = boto3.client('ssm', region_name)
athena_bucket_name = ssm_client.get_parameter(Name="SYSTEM.S3BUCKET.ATHENA")["Parameter"]["Value"]

# To set notebook_id user are required to run the following in a new cell and retrieve the id(ResourceName) from there.
# Or can retrieve it from Amorphic's Datalab page.
# ```!cat /opt/ml/metadata/resource-metadata.json``` --> {"ResourceArn": "arn:aws:sagemaker:::notebook-instance/<notebook_id>", "ResourceName": "<notebook_id>"}
notebook_id = "<>"

# glue sessions disabled notebook has notebook_type -> sagemaker
# glue sessions enabled notebook has notebook_type -> sagemaker-glue-session
notebook_type = "<>"

# set the database name & view name to query the data from
domain_name = "<>"
view_name = "<>"

# set the quey output loc -> s3://AthenaBucket/NotebookType/NotebookId/
df_athena = wr.athena.read_sql_query(f"SELECT * FROM {domain_name}.{view_name}", database=domain_name, ctas_approach=False, s3_output=f"s3://{athena_bucket_name}/{notebook_type}/{notebook_id}/")
  • Using boto3 S3-Athena client : If the notebook has internet disabled or a view is queried inside a conda_glue_pyspark kernel in glue sessions enabled notebook, then the user should prefer using the below script.
#imports
import boto3

# get the aws region
boto3_session = boto3.session.Session()
region_name = boto3_session.region_name

#boto3 clients
athena_client = boto3.client("athena")
ssm_client = boto3.client('ssm', region_name)
athena_bucket_name = ssm_client.get_parameter(Name="SYSTEM.S3BUCKET.ATHENA")["Parameter"]["Value"]

# To set notebook_id user are required to run the following in a new cell and retrieve the id(ResourceName) from there.
# Or can retrieve it from Amorphic's Datalab page.
# ```!cat /opt/ml/metadata/resource-metadata.json``` --> {"ResourceArn": "arn:aws:sagemaker:::notebook-instance/<notebook_id>", "ResourceName": "<notebook_id>"}
notebook_id = "<>"

# glue sessions disabled notebook has notebook_type -> sagemaker
# glue sessions enabled notebook has notebook_type -> sagemaker-glue-session
notebook_type = "<>"

# set the database name & view name to query the data from
database_name = "<>"
view_name = "<>"

# set the quey output loc -> s3://AthenaBucket/NotebookType/NotebookId/
query_output_loc = f"s3://{athena_bucket_name}/{notebook_type}/{notebook_id}/"

#initiate the query through athena client
queryStart = athena_client.start_query_execution(
QueryString = f'SELECT * FROM {database_name}.{view_name}',
QueryExecutionContext = {
'Database': database_name
},
ResultConfiguration = { 'OutputLocation': query_output_loc}
)

# after starting the execution of the query, user should wait for s3athena to process the equation
queryExecution = athena_client.get_query_execution(QueryExecutionId=queryStart['QueryExecutionId'])
results = athena_client.get_query_results(QueryExecutionId=queryStart['QueryExecutionId'])