Feature | Description |
Workflow | Notebook jobs can be scheduled in a workflow. |
Job history | Stdout and Stderr logs can be viewed on the Development Job Runs tab of the Job History page in the EMR console. Resources that are consumed to run a job can be viewed on the Development Job Runs tab of the Job History page in the EMR console, such as the memory size, CPU cores, and compute units (CUs).
|
Session management | Internal endpoints are supported by Spark Thrift Servers. Custom JAR packages are supported for Spark Thrift Servers whose engine version is esr-2.4 or later. LDAP authentication and Ranger authentication are supported for Spark Thrift Servers. Creation time and start time of a session can be viewed.
|
Gateway management | Internal endpoints are supported by Livy gateways. Spark sessions that are created by using the Livy interface can be viewed and the Spark UI can be accessed. The driver logs of Spark sessions that are created by using the Livy interface can be viewed by using a specified gateway. This applies to only Livy gateways whose version is esr-2.2.2 or later. Custom runtime environments are available for Livy gateways.
|
Data development | The maximum size of a notebook is increased. Code of a notebook can be run in another notebook. Custom runtime environments are available for PySpark jobs.
|
Others | Creation of folders is supported. Comments can be added in the Spark Configuration field. A Spark driver can be viewed after the spark_submit command is run.
|