This topic describes the types of E-MapReduce (EMR) events that ActionTrail records or CloudMonitor monitors and can be published to EventBridge.

Event types

The following table describes the types of EMR events that can be published to EventBridge.

Event type Value of the type parameter
Operation performed by Alibaba Cloud on a resource emr:ActionTrail:AliyunServiceEvent
API operation call emr:ActionTrail:ApiCall
Operation performed in a console emr:ActionTrail:ConsoleOperation
The heartbeat message of the ECmanaged agent expires emr:CloudMonitor:Agent[EcmAgentHeartbeatExpired]
The ECmanaged agent remains disconnected for a long time emr:CloudMonitor:Agent[Maintenance.EcmAgentTimeout]
A workflow is complete emr:CloudMonitor:EMR-110401002
A workflow is submitted emr:CloudMonitor:EMR-110401003
A job is submitted emr:CloudMonitor:EMR-110401004
A workflow node is started emr:CloudMonitor:EMR-110401005
The status of a workflow node is checked emr:CloudMonitor:EMR-110401006
A workflow node is complete emr:CloudMonitor:EMR-110401007
A workflow node is stopped emr:CloudMonitor:EMR-110401008
A workflow node is canceled emr:CloudMonitor:EMR-110401009
A workflow is canceled emr:CloudMonitor:EMR-110401010
A workflow is restarted emr:CloudMonitor:EMR-110401011
A workflow is resumed emr:CloudMonitor:EMR-110401012
A workflow is paused emr:CloudMonitor:EMR-110401013
A workflow is stopped emr:CloudMonitor:EMR-110401014
A workflow node fails emr:CloudMonitor:EMR-110401015
A job fails emr:CloudMonitor:EMR-110401016
A workflow fails emr:CloudMonitor:EMR-210401001
The startup of a workflow node times out emr:CloudMonitor:EMR-210401003
The startup of a job times out emr:CloudMonitor:EMR-210401004
The status of the Airflow scheduler fails to be checked emr:CloudMonitor:Maintenance[AIRFLOW.Scheduler.StatusCheck.Fail]
The status of the Airflow web server fails to be checked emr:CloudMonitor:Maintenance[AIRFLOW.WebServer.Check.Fail]
The service status of the Airflow web server fails to be checked emr:CloudMonitor:Maintenance[AIRFLOW.WebServer.StatusCheck.Fail]
The status of ApacheDS fails to be checked emr:CloudMonitor:Maintenance[APACHEDS.StatusCheck.Fail]
The status of the ClickHouse server fails to be checked emr:CloudMonitor:Maintenance[CLICKHOUSE.ServerStatusCheck.Fail]
Garbage collection (GC) for a Druid broker fails to be checked emr:CloudMonitor:Maintenance[DRUID.Broker.GcCheck.Fail]
The status of a Druid broker fails to be checked emr:CloudMonitor:Maintenance[DRUID.Broker.StatusCheck.Fail]
GC for a Druid coordinator fails to be checked emr:CloudMonitor:Maintenance[DRUID.Coordinator.GcCheck.Fail]
The status of a Druid coordinator fails to be checked emr:CloudMonitor:Maintenance[DRUID.Coordinator.StatusCheck.Fail]
GC for a Druid historical node fails to be checked emr:CloudMonitor:Maintenance[DRUID.Historical.GcCheck.Fail]
The status of a Druid historical node fails to be checked emr:CloudMonitor:Maintenance[DRUID.Historical.StatusCheck.Fail]
GC for a Druid middle manager fails to be checked emr:CloudMonitor:Maintenance[DRUID.MiddleManager.GcCheck.Fail]
The status of a Druid middle manager fails to be checked emr:CloudMonitor:Maintenance[DRUID.MiddleManager.StatusCheck.Fail]
GC for a Druid overlord fails to be checked emr:CloudMonitor:Maintenance[DRUID.Overlord.GcCheck.Fail]
The status of a Druid overlord fails to be checked emr:CloudMonitor:Maintenance[DRUID.Overlord.StatusCheck.Fail]
GC for a Druid router fails to be checked emr:CloudMonitor:Maintenance[DRUID.Router.GcCheck.Fail]
The status of a Druid router fails to be checked emr:CloudMonitor:Maintenance[DRUID.Router.StatusCheck.Fail]
GC for a Flink history server fails to be checked emr:CloudMonitor:Maintenance[FLINK.HistoryServer.GcCheckP0.Fail]
The status of a Flink history server fails to be checked emr:CloudMonitor:Maintenance[FLINK.HistoryServer.StatusCheck.Fail]
The status of a Flink Ververica Platform (VVP) server fails to be checked emr:CloudMonitor:Maintenance[FLINK.VVP.StatusCheck.Fail]
The status of the HAS administrator fails to be checked emr:CloudMonitor:Maintenance[HAS.Admin.StatusCheck.Fail]
The status of HAS fails to be checked emr:CloudMonitor:Maintenance[HAS.Server.StatusCheck.Fail]
The availability of an HBase cluster fails to be checked emr:CloudMonitor:Maintenance[HBASE.AvailabilityStatusCheck.Fail]
The inter-process communication (IPC) port of the HBase HMaster is unavailable emr:CloudMonitor:Maintenance[HBASE.HMaster.IpcPortUnAvailable]
The status of the HBase HMaster fails to be checked emr:CloudMonitor:Maintenance[HBASE.HMaster.StatusCheck.Fail]
The IPC port of an HBase HRegionServer is unavailable emr:CloudMonitor:Maintenance[HBASE.HRegionServer.IpcPortUnAvailable]
GC for an HBase RegionServer fails to be checked emr:CloudMonitor:Maintenance[HBASE.RegionServer.GcCheckP0.Fail]
The status of an HBase RegionServer fails to be checked emr:CloudMonitor:Maintenance[HBASE.RegionServer.StatusCheck.Fail]
GC for an HBase Thrift server fails to be checked emr:CloudMonitor:Maintenance[HBASE.ThriftServer.GcCheckP0.Fail]
The service port of an HBase Thrift server is unavailable emr:CloudMonitor:Maintenance[HBASE.ThriftServer.ServicePortUnAvailable]
The status of an HBase Thrift server fails to be checked emr:CloudMonitor:Maintenance[HBASE.ThriftServer.StatusCheck.Fail]
The availability of the Hadoop Distributed File System (HDFS) fails to be checked emr:CloudMonitor:Maintenance[HDFS.AvailabilityStatusCheck.Fail]
The data transmission port of a DataNode is unavailable emr:CloudMonitor:Maintenance[HDFS.DataNode.DataTransferPortUnAvailable]
A dead DataNode exists in the HDFS emr:CloudMonitor:Maintenance[HDFS.DataNode.DeadDataNodesExist]
An exception occurs in secureMain of a DataNode emr:CloudMonitor:Maintenance[HDFS.DataNode.ExceptionInSecureMain]
A DataNode unexpectedly exits emr:CloudMonitor:Maintenance[HDFS.DataNode.ExitUnexpected]
One or more damaged disks exist in a DataNode emr:CloudMonitor:Maintenance[HDFS.DataNode.FailueVolumes]
GC for a DataNode fails to be checked (P0) emr:CloudMonitor:Maintenance[HDFS.DataNode.GcCheckP0.Fail]
The IPC port of a DataNode is unavailable emr:CloudMonitor:Maintenance[HDFS.DataNode.IpcPortUnAvailable]
A DataNode cannot create a native thread due to an out-of-memory (OOM) error that occurs in the DataNode emr:CloudMonitor:Maintenance[HDFS.DataNode.OOM.UnableToCreateNewNativeThread]
Java heap space causes an OOM error emr:CloudMonitor:Maintenance[HDFS.DataNode.OomForJavaHeapSpace]
The status of a DataNode fails to be checked emr:CloudMonitor:Maintenance[HDFS.DataNode.StatusCheck.Fail]
Excessive dead DataNodes exists in the HDFS emr:CloudMonitor:Maintenance[HDFS.DataNode.TooManyDataNodeDead]
One or more damaged disks exist in the HDFS emr:CloudMonitor:Maintenance[HDFS.DataNode.VolumeFailuresExist]
The high availability (HA) status of the HDFS fails to be checked emr:CloudMonitor:Maintenance[HDFS.HaStateCheck.Fail]
GC for a JournalNode fails to be checked (P0) emr:CloudMonitor:Maintenance[HDFS.JournalNode.GcCheckP0.Fail]
The Remote Procedure Call (RPC) port of a JournalNode is unavailable emr:CloudMonitor:Maintenance[HDFS.JournalNode.RpcPortUnAvailable]
The status of a JournalNode fails to be checked emr:CloudMonitor:Maintenance[HDFS.JournalNode.StatusCheck.Fail]
A switchover occurs between active and standby NameNodes emr:CloudMonitor:Maintenance[HDFS.NameNode.ActiveStandbySwitch]
The block capacity of a NameNode is running out emr:CloudMonitor:Maintenance[HDFS.NameNode.BlockCapacityNearUsedUp]
Both NameNodes are active emr:CloudMonitor:Maintenance[HDFS.NameNode.BothActive]
Both NameNodes are standby emr:CloudMonitor:Maintenance[HDFS.NameNode.BothStandy]
One or more damaged blocks exist in the HDFS emr:CloudMonitor:Maintenance[HDFS.NameNode.CorruptBlocksOccured]
A directory is formatted in the HDFS emr:CloudMonitor:Maintenance[HDFS.NameNode.DirectoryFormatted]
A NameNode unexpectedly exits emr:CloudMonitor:Maintenance[HDFS.NameNode.ExitUnexpectely]
GC for a NameNode fails to be checked (P0) emr:CloudMonitor:Maintenance[HDFS.NameNode.GcCheckP0.Fail]
GC for a NameNode fails to be checked (P1) emr:CloudMonitor:Maintenance[HDFS.NameNode.GcCheckP1.Fail]
A NameNode remains in safe mode for a long time emr:CloudMonitor:Maintenance[HDFS.NameNode.InSafeMode]
The IPC port of a NameNode is unavailable emr:CloudMonitor:Maintenance[HDFS.NameNode.IpcPortUnAvailable]
An exception occurs when a NameNode loads FsImage emr:CloudMonitor:Maintenance[HDFS.NameNode.LoadFsImageException]
A NameNode is in safe mode due to insufficient disk space emr:CloudMonitor:Maintenance[HDFS.NameNode.LowAvailableDiskSpaceAndInSafeMode]
A data block is missing in the HDFS emr:CloudMonitor:Maintenance[HDFS.NameNode.MissingBlock]
An OOM error occurs in a NameNode emr:CloudMonitor:Maintenance[HDFS.NameNode.OOM]
A NameNode does not have sufficient resources emr:CloudMonitor:Maintenance[HDFS.NameNode.ResourceLow]
The RPC call queue of a NameNode is long emr:CloudMonitor:Maintenance[HDFS.NameNode.RpcPortCallQueueLengthTooLong]
The status of a NameNode fails to be checked emr:CloudMonitor:Maintenance[HDFS.NameNode.StatusCheck.Fail]
A NameNode fails to synchronize logs emr:CloudMonitor:Maintenance[HDFS.NameNode.SyncJournalFailed]
Excessive block space is used in the HDFS emr:CloudMonitor:Maintenance[HDFS.NameNode.TooMuchBlockCapacityUsed]
Excessive DataNode space is used emr:CloudMonitor:Maintenance[HDFS.NameNode.TooMuchDataNodeCapacityUsed]
Excessive storage space is used in the HDFS emr:CloudMonitor:Maintenance[HDFS.NameNode.TooMuchDfsCapacityUsed]
Excessive files and blocks consume a large amount of heap memory emr:CloudMonitor:Maintenance[HDFS.NameNode.TooMuchHeapUsedByTooManyFilesAndBlocks]
Data write by the HDFS to a JournalNode times out emr:CloudMonitor:Maintenance[HDFS.NameNode.WriteToJournalNodeTimeout]
The ZKFailoverController (ZKFC) triggers a switchover between active and standby NameNodes emr:CloudMonitor:Maintenance[HDFS.ZKFC.ActiveStandbySwitchOccured]
The port of the HDFS ZKFC is unavailable emr:CloudMonitor:Maintenance[HDFS.ZKFC.PortUnAvailable]
The status of the ZKFC fails to be checked emr:CloudMonitor:Maintenance[HDFS.ZKFC.StatusCheck.Fail]
A transport layer exception occurs when the ZKFC monitors the health status of a NameNode emr:CloudMonitor:Maintenance[HDFS.ZKFC.TransportLevelExceptionInMonitorHealth]
The ZKFC cannot connect to the ZooKeeper quorum emr:CloudMonitor:Maintenance[HDFS.ZKFC.UnableToConnectToQuorum]
The ZKFC cannot be started emr:CloudMonitor:Maintenance[HDFS.ZKFC.UnableToStartZKFC]
The availability of Apache Hive fails to be checked emr:CloudMonitor:Maintenance[HIVE.AvailabilityStatusCheck.Fail]
The communication link of a Hive Metastore (HMS) database fails emr:CloudMonitor:Maintenance[HIVE.HiveMetaStore.DataBaseCommunicationLinkFailure]
Fail to connect to an HMS database emr:CloudMonitor:Maintenance[HIVE.HiveMetaStore.DataBaseConnectionFailed]
An HMS database runs out of disk space emr:CloudMonitor:Maintenance[HIVE.HiveMetaStore.DataBaseDiskQuotaUsedup]
The port for communication between the HMS and HiveServer2 is unavailable emr:CloudMonitor:Maintenance[HIVE.HiveMetaStore.hiveServer2PortUnAvailable]
A Java Database Connectivity (JDBC) exception occurs in the HMS emr:CloudMonitor:Maintenance[HIVE.HiveMetaStore.JdbcCommunicationException]
The number of queries for the HMS exceeds the upper limit emr:CloudMonitor:Maintenance[HIVE.HiveMetaStore.MaxQuestionsExceeded]
The number of updates for the HMS exceeds the upper limit emr:CloudMonitor:Maintenance[HIVE.HiveMetaStore.MaxUpdatesExceeded]
The number of user connections for the HMS exceeds the upper limit emr:CloudMonitor:Maintenance[HIVE.HiveMetaStore.MaxUserConnectionExceeded]
An OOM error occurs in the HMS emr:CloudMonitor:Maintenance[HIVE.HiveMetaStore.OomOccured]
An error occurs when the HMS configuration file is parsed emr:CloudMonitor:Maintenance[HIVE.HiveMetaStore.ParseConfError]
The port of the HMS is unavailable emr:CloudMonitor:Maintenance[HIVE.HiveMetaStore.PortUnAvailable]
The required table of the HMS is missing emr:CloudMonitor:Maintenance[HIVE.HiveMetaStore.RequiredTableMissing]
GC for HiveServer fails to be checked (P0) emr:CloudMonitor:Maintenance[HIVE.HiveServer.GcCheckP0.Fail]
GC for HiveServer fails to be checked (P1) emr:CloudMonitor:Maintenance[HIVE.HiveServer.GcCheckP1.Fail]
The status of HiveServer fails to be checked emr:CloudMonitor:Maintenance[HIVE.HiveServer.StatusCheck.Fail]
Unable to connect to HiveServer2 by using the provided uniform resource identifiers (URIs) emr:CloudMonitor:Maintenance[HIVE.HiveServer2.CannotConnectByAnyURIsProvided]
The connection between HiveServer2 and ZooKeeper times out emr:CloudMonitor:Maintenance[HIVE.HiveServer2.ConnectToZkTimeout]
An error occurs when the HiveServer2 configuration is parsed emr:CloudMonitor:Maintenance[HIVE.HiveServer2.ErrorParseConf]
An error occurs when HiveServer2 is started emr:CloudMonitor:Maintenance[HIVE.HiveServer2.ErrorStartingHiveServer]
HiveServer2 fails to initialize a Metastore client emr:CloudMonitor:Maintenance[HIVE.HiveServer2.FailedInitMetaStoreClient]
HiveServer2 fails to connect to the Metastore server emr:CloudMonitor:Maintenance[HIVE.HiveServer2.FailedToConnectToMetaStoreServer]
An OOM error occurs in HiveServer2 emr:CloudMonitor:Maintenance[HIVE.HiveServer2.HiveServer2OOM]
The latency of the Metastore fails to be checked (P0) emr:CloudMonitor:Maintenance[HIVE.MetaStore.DelayCheckP0.Fail]
The latency of the Metastore fails to be checked (P1) emr:CloudMonitor:Maintenance[HIVE.MetaStore.DelayCheckP1.Fail]
GC for the Metastore fails to be checked (P0) emr:CloudMonitor:Maintenance[HIVE.MetaStore.GcCheckP0.Fail]
GC for the Metastore fails to be checked (P1) emr:CloudMonitor:Maintenance[HIVE.MetaStore.GcCheckP1.Fail]
The status of the MetaStore fails to be checked emr:CloudMonitor:Maintenance[HIVE.MetaStore.StatusCheck.Fail]
Stuttering occurs on a host due to high CPU utilization emr:CloudMonitor:Maintenance[HOST.CpuStuck]
The memory usage is high emr:CloudMonitor:Maintenance[HOST.HighMemoryUsage]
The size of the absolute memory available to a host is small emr:CloudMonitor:Maintenance[HOST.LowAbsoluteFreeMemory]
The size of the available space of the /mnt/disk1 directory is small emr:CloudMonitor:Maintenance[HOST.LowDiskForMntDisk1]
The size of the disk space available for the root file system is small emr:CloudMonitor:Maintenance[HOST.LowRootfsDisk]
An OOM error occurs in the /var/log/message directory on a host emr:CloudMonitor:Maintenance[HOST.OomFoundInVarLogMessage]
Excessive processes exist on a primary node emr:CloudMonitor:Maintenance[HOST.TooManyProcessesOnMasterHost]
A host is shut down emr:CloudMonitor:Maintenance[HOST.VmHostShutDown]
A host is started emr:CloudMonitor:Maintenance[HOST.VmHostStartUp]
The management port of Oozie is unavailable emr:CloudMonitor:Maintenance[HUE.OozieAdminPortUnAvailable]
The service port of HUE is unavailable emr:CloudMonitor:Maintenance[HUE.PortUnAvailable]
The status of the HUE RunCherryPyServer fails to be checked emr:CloudMonitor:Maintenance[HUE.RunCherryPyServer.StatusCheck.Fail]
The status of HUE fails to be checked emr:CloudMonitor:Maintenance[HUE.StatusCheck.Fail]
The availability of Apache Impala fails to be checked emr:CloudMonitor:Maintenance[IMPALA.AvailableCheck.Fail]
The availability of the Impala Catalog daemon (catalogd) fails to be checked emr:CloudMonitor:Maintenance[IMPALA.Catalogd.AvailableCheck.Fail]
The availability of the Impala daemon (impalad) fails to be checked emr:CloudMonitor:Maintenance[IMPALA.Impalad.AvailableCheck.Fail]
The availability of the Impala Statestore daemon (statestored) fails to be checked emr:CloudMonitor:Maintenance[IMPALA.StateStored.AvailableCheck.Fail]
The status of the JindoFS Manager service fails to be checked emr:CloudMonitor:Maintenance[JINDOFS.JindoFsManagerService.StatusCheck.Fail]
The status of the JindoFS Namespace service fails to be checked emr:CloudMonitor:Maintenance[JINDOFS.JindoFsNamespaceStatusCheck.Fail]
The status of the JindoFS Storage service fails to be checked emr:CloudMonitor:Maintenance[JINDOFS.JindoFsStorageServiceStatusCheck.Fail]
The status of JindoFS fails to be checked emr:CloudMonitor:Maintenance[JINDOFS.StatusCheck.Fail]
The availability of a Kafka broker fails to be checked emr:CloudMonitor:Maintenance[KAFKA.Broker.AvailableCheck.Fail]
GC for a Kafka broker fails to be checked (P0) emr:CloudMonitor:Maintenance[KAFKA.Broker.GcCheckP0.Fail]
GC for a Kafka broker fails to be checked (P1) emr:CloudMonitor:Maintenance[KAFKA.Broker.GcCheckP1.Fail]
The status of a Kafka broker fails to be checked emr:CloudMonitor:Maintenance[KAFKA.Broker.StateCheck.Fail]
Kafka Manager fails to be checked emr:CloudMonitor:Maintenance[KAFKA.KafkaManager.Check.Fail]
The Kafka metadata monitor fails to be checked emr:CloudMonitor:Maintenance[KAFKA.KafkaMetadataMonitor.Check.Fail]
The Kafka REST Proxy fails to be checked emr:CloudMonitor:Maintenance[KAFKA.RestProxy.Check.Fail]
The Kafka Schema Registry fails to be checked emr:CloudMonitor:Maintenance[KAFKA.SchemaRegistry.Check.Fail]
GC for Knox fails to be checked emr:CloudMonitor:Maintenance[KNOX.GcCheckP0.Fail]
The status of Knox fails to be checked emr:CloudMonitor:Maintenance[KNOX.StatusCheck.Fail]
The health status of Apache Kudu fails to be checked emr:CloudMonitor:Maintenance[KUDU.HealthyCheck.Fail]
The status of a Kudu master fails to be checked emr:CloudMonitor:Maintenance[KUDU.MasterStatusCheck.Fail]
The status of a Kudu tserver fails to be checked emr:CloudMonitor:Maintenance[KUDU.TServerStatusCheck.Fail]
GC for Apache Livy fails to be checked emr:CloudMonitor:Maintenance[LIVY.GcCheckP0.Fail]
The status of Apache Livy fails to be checked emr:CloudMonitor:Maintenance[LIVY.StatusCheck.Fail]
GC for Apache Oozie fails to be checked emr:CloudMonitor:Maintenance[OOZIE.GcCheckP0.Fail]
The status of Apache Oozie fails to be checked emr:CloudMonitor:Maintenance[OOZIE.StatusCheck.Fail]
The status of OpenLDAP fails to be checked emr:CloudMonitor:Maintenance[OPENLDAP.StatusCheck.Fail]
The availability of Presto fails to be checked emr:CloudMonitor:Maintenance[PRESTO.AvailabilityStatusCheck.Fail]
GC for a Presto coordinator fails to be checked emr:CloudMonitor:Maintenance[PRESTO.Coordinator.GcCheckP0.Fail]
The status of a Presto coordinator fails to be checked emr:CloudMonitor:Maintenance[PRESTO.Coordinator.StatusCheck.Fail]
GC for a Presto worker fails to be checked emr:CloudMonitor:Maintenance[PRESTO.Worker.GcCheckP0.Fail]
The status of a Presto worker fails to be checked emr:CloudMonitor:Maintenance[PRESTO.Worker.StatusCheck.Fail]
GC for a Ranger administrator fails to be checked emr:CloudMonitor:Maintenance[RANGER.ADMIN.GcCheck.Fail]
The status of a Ranger administrator fails to be checked emr:CloudMonitor:Maintenance[RANGER.ADMIN.StatusCheck.Fail]
The status of the Ranger Solr component fails to be checked emr:CloudMonitor:Maintenance[RANGER.Solr.StatusCheck.Fail]
The status of Ranger UserSync fails to be checked emr:CloudMonitor:Maintenance[RANGER.UserSync.StatusCheck.Fail]
GC for the Spark history server fails to be checked emr:CloudMonitor:Maintenance[SPARK.HistoryServer.GcCheckP0.Fail]
The status of the Spark history server fails to be checked emr:CloudMonitor:Maintenance[SPARK.HistoryServer.StatusCheck.Fail]
An OOM error occurs on the Spark history server emr:CloudMonitor:Maintenance[SPARK.SparkHistory.OomOccured]
The status of the Spark Thrift server fails to be checked emr:CloudMonitor:Maintenance[SPARK.ThriftServer.StatusCheck.Fail]
The Storm Nimbus Thrift port is unavailable emr:CloudMonitor:Maintenance[STORM.Nimbus.ThriftPortUnAvailable]
The status of Apache Superset fails to be checked emr:CloudMonitor:Maintenance[SUPERSET.StatusCheck.Fail]
GC for the TEZ Tomcat fails to be checked emr:CloudMonitor:Maintenance[TEZ.Tomcat.GcCheckP0.Fail]
The status of the TEZ Tomcat fails to be checked emr:CloudMonitor:Maintenance[TEZ.Tomcat.StatusCheck.Fail]
GC for AppTimeLine fails to be checked (P0) emr:CloudMonitor:Maintenance[YARN.AppTimeLine.GcCheckP0.Fail]
The status of AppTimeLine fails to be checked emr:CloudMonitor:Maintenance[YARN.AppTimeLine.StatusCheck.Fail]
The HA status of Yarn fails to be checked emr:CloudMonitor:Maintenance[YARN.HaStateCheck.Fail]
The JobHistory service unexpectedly exits emr:CloudMonitor:Maintenance[YARN.JobHistory.ExitUnExpectedly]
GC for JobHistory fails to be checked (P0) emr:CloudMonitor:Maintenance[YARN.JobHistory.GcCheckP0.Fail]
The service port of JobHistory is unavailable emr:CloudMonitor:Maintenance[YARN.JobHistory.PortUnAvailable]
An error occurs when the JobHistory service is started emr:CloudMonitor:Maintenance[YARN.JobHistory.StartingError]
The status of JobHistory fails to be checked emr:CloudMonitor:Maintenance[YARN.JobHistory.StatusCheck.Fail]
One or more dead NodeManagers are detected emr:CloudMonitor:Maintenance[YARN.NodeManager.DeadNodeDetected]
A NodeManager fails to start NodeStatusUpdater emr:CloudMonitor:Maintenance[YARN.NodeManager.ErrorRebootingNodeStatusUpdater]
GC for a NodeManager fails to be checked (P0) emr:CloudMonitor:Maintenance[YARN.NodeManager.GcCheckP0.Fail]
One or more NodeManagers are missing emr:CloudMonitor:Maintenance[YARN.NodeManager.LostNodesExist]
An OOM error occurs in a NodeManager emr:CloudMonitor:Maintenance[YARN.NodeManager.OOM]
An error occurs when a NodeManager is started emr:CloudMonitor:Maintenance[YARN.NodeManager.StartingError]
The status of a NodeManager fails to be checked emr:CloudMonitor:Maintenance[YARN.NodeManager.StatusCheck.Fail]
A NodeManager becomes unhealthy due to disk errors emr:CloudMonitor:Maintenance[YARN.NodeManager.UnHealthyForDiskFailed]
One or more unhealthy NodeManagers exist in Yarn emr:CloudMonitor:Maintenance[YARN.NodeManager.UnHealthyNodesExist]
A switchover occurs between active and standby ResourceManagers emr:CloudMonitor:Maintenance[YARN.ResourceManager.ActiveStandbySwitch]
Both ResourceManagers are active emr:CloudMonitor:Maintenance[YARN.ResourceManager.BothInActive]
Both ResourceManagers are standby emr:CloudMonitor:Maintenance[YARN.ResourceManager.BothInStandby]
A ResourceManager fails to be switched to the active state emr:CloudMonitor:Maintenance[YARN.ResourceManager.CouldNotTransitionToActive]
An error occurs when a ResourceManager is started emr:CloudMonitor:Maintenance[YARN.ResourceManager.ErrorInStarting]
An error occurs when a ResourceManager is switched to the active state emr:CloudMonitor:Maintenance[YARN.ResourceManager.ErrorInTransitionToActiveMode]
A ResourceManager unexpectedly exits emr:CloudMonitor:Maintenance[YARN.ResourceManager.ExitUnexpected]
GC for a ResourceManager fails to be checked (P0) emr:CloudMonitor:Maintenance[YARN.ResourceManager.GcCheckP0.Fail]
GC for a ResourceManager fails to be checked (P1) emr:CloudMonitor:Maintenance[YARN.ResourceManager.GcCheckP1.Fail]
RM_HA_ID cannot be found due to the invalid configuration of a ResourceManager emr:CloudMonitor:Maintenance[YARN.ResourceManager.InvalidConf.CannotFoundRMHAID]
An OOM error occurs in a ResourceManager emr:CloudMonitor:Maintenance[YARN.ResourceManager.OOM]
The service port of a ResourceManager in Yarn is unavailable emr:CloudMonitor:Maintenance[YARN.ResourceManager.PortUnAvailable]
The restart status of a ResourceManager fails to be checked emr:CloudMonitor:Maintenance[YARN.ResourceManager.RestartCheck.Fail]
The status of a ResourceManager fails to be checked emr:CloudMonitor:Maintenance[YARN.ResourceManager.StatusCheck.Fail]
An unknown host exception occurs in a ResourceManager emr:CloudMonitor:Maintenance[YARN.ResourceManager.UnkownHostException]
ZKRMStateStore cannot connect to ZooKeeper in Yarn emr:CloudMonitor:Maintenance[YARN.ResourceManager.ZKRMStateStoreCannotConnectZK]
The status of Yarn fails to be checked emr:CloudMonitor:Maintenance[YARN.StatusCheck.Fail]
An error occurs when the Timeline server is started emr:CloudMonitor:Maintenance[YARN.TimelineServer.ErrorInStarting]
The Timeline server unexpectedly exits emr:CloudMonitor:Maintenance[YARN.TimelineServer.ExistUnexpectedly]
The port of the Yarn Timeline server is unavailable emr:CloudMonitor:Maintenance[YARN.TimelineServer.PortUnAvailable]
The status of WebAppProxy fails to be checked emr:CloudMonitor:Maintenance[YARN.WebAppProxy.StatusCheck.Fail]
The service port of the Yarn WebAppProxy server is unavailable emr:CloudMonitor:Maintenance[YARN.WebAppProxyServer.PortUnAvailable]
The status of Zeppelin fails to be checked emr:CloudMonitor:Maintenance[ZEPPELIN.Server.StatusCheck.Fail]
The status of the Zeppelin component fails to be checked emr:CloudMonitor:Maintenance[ZEPPELIN.ServerCheck.Fail]
The client port of ZooKeeper is unavailable emr:CloudMonitor:Maintenance[ZOOKEEPER.ClientPortUnAvailable]
The status of a ZooKeeper cluster fails to be checked emr:CloudMonitor:Maintenance[ZOOKEEPER.ClusterStatusCheck.Fail]
GC for ZooKeeper fails to be checked emr:CloudMonitor:Maintenance[ZOOKEEPER.GcCheckP0.Fail]
An active/standby switchover occurs in ZooKeeper emr:CloudMonitor:Maintenance[ZOOKEEPER.LeaderFollowerSwitch]
The leader port of ZooKeeper is unavailable emr:CloudMonitor:Maintenance[ZOOKEEPER.LeaderPortUnAvailable]
The peer port of ZooKeeper is unavailable emr:CloudMonitor:Maintenance[ZOOKEEPER.PeerPortUnAvailable]
The status of a ZooKeeper process fails to be checked emr:CloudMonitor:Maintenance[ZOOKEEPER.StatusCheck.Fail]
ZooKeeper cannot run QuorumServer emr:CloudMonitor:Maintenance[ZOOKEEPER.UnableToRunQuorumServer]
A scaling activity fails emr:CloudMonitor:Scaling[ScalingActivity:Failed]
A scaling activity is rejected emr:CloudMonitor:Scaling[ScalingActivity:Rejected]
A scaling activity times out emr:CloudMonitor:Scaling[ScalingActivity:Timeout]
The status of a service component is checked emr:CloudMonitor:StatusCheck

For more information about the parameters defined in the CloudEvents specification, see Overview.