This topic describes how to deploy Apache Spark on a single Elastic Compute Service (ECS) instance by creating a stack in the Resource Orchestration Service (ROS) console.

Background information

Apache Spark is a general-purpose computing engine designed for large-scale data processing. Apache Spark uses Scala as its application framework and uses resilient distributed datasets (RDDs) for in-memory computing. It provides interactive queries and can optimize workloads by means of iterative algorithms.

The Deploy Spark on a Single Instance in an Existing VPC sample template creates an ECS instance based on existing resources such as a virtual private cloud (VPC), a vSwitch, and a security group and associates an elastic IP address (EIP) with the instance. The following software versions are used in the sample template:

  • JDK: 1.8.0
  • Hadoop: 2.7.7
  • Scala: 2.12.1
  • Kafka: 2.1.0

After a stack is created by using the sample template, you can obtain the URL of the Spark web interface and log on to the Spark management console. If you want to access the URL of the Spark web interface over the Internet, add inbound rules to the security group to allow traffic on ports 8088 and 8080. For more information, see Add security group rules.

Step 1: Create a stack

  1. Log on to the ROS console.
  2. In the left-side navigation pane, choose Templates > Sample Templates.
  3. Find the Deploy Spark on a Single Instance in an Existing VPC template.
  4. Click Create Stack.
  5. In the Configure Template Parameters step, set Stack Name and the following parameters.
    Parameter Description Example
    Existing VPC ID The ID of the VPC.

    For more information about how to create and query a VPC, see Create a VPC.

    vpc-bp1m6fww66xbntjyc****
    VSwitch Zone ID The zone ID of the vSwitch in the VPC. Hangzhou Zone K
    VSwitch ID The ID of the vSwitch in the VPC.

    For more information about how to create and query a vSwitch, see Create a vSwitch.

    vsw-bp183p93qs667muql****
    Business Security Group ID The ID of the ECS security group.

    For more information about how to query security groups, see Query security groups.

    sg-bp15ed6xe1yxeycg7o****
    Instance Type The instance type of the ECS instance.

    Select a valid instance type. For more information, see Instance families.

    ecs.c5.large
    Image ID The image ID of the ECS instance. By default, centos_7 is used.

    For more information, see Image overview.

    centos_7
    Instance Password The password that is used to log on to the ECS instance. Test_12****
    Disk Type The disk category of the ECS instance. Valid values:
    • cloud_efficiency: ultra disk
    • cloud_ssd: standard SSD
    • cloud_essd: enhanced SSD (ESSD)
    • cloud: basic disk
    • ephemeral_ssd: local SSD

    For more information, see Cloud disks.

    cloud_efficiency
    System Disk Space The system disk size of the ECS instance.

    Valid values: 40 to 500.

    Unit: GB.

    40
  6. Click Create.
  7. View the stack status on the Stack Information tab of the stack management page. After the stack is created, click the Outputs tab to view the URL of the Spark web interface.
  8. Use the URL to log on to the Spark management console.

Step 2: View resources

  1. Log on to the ROS console.
  2. In the left-side navigation pane, click Stacks.
  3. On the Stacks page, click the stack that you created.
  4. On the stack management page, click the Resources tab to view the resource list.
    The following table describes the resources in this example.
    Resource type Quantity Description Specifications
    ALIYUN::ECS::Instance 1 Creates an ECS instance to deploy Apache Spark. A single instance of the following specifications is created:
    • InstanceType: ecs.c5.large
    • SystemDiskCategory: cloud_efficiency
    • SystemDiskSize: 40
    • AllocatePublicIP: true
    Note For more information about the resource charges, see the pricing schedule on the official website or the product pricing documentation.