flintrock: Fail to launch jobs on Amazon EC2 with master-IP

I try to run several spark-bench benchmarks on EC2 launched by flintrock.

In the standalone mode of spark, I have to configure the IP address of EC2 machine as the “master” of spark.

The console and .conf is below:

[ec2-user@ip-172-31-15-16 new-spark-bench_2.1.1]$ ./bin/spark-bench.sh examples/kmeans.conf
 *** SPARK-SUBMIT: [/home/ec2-user/spark/bin/spark-submit, --class, com.ibm.sparktc.sparkbench.cli.CLIKickoff, --master, spark://34.215.233.221:7077, /home/ec2-user/new-spark-bench_2.1.1/lib/spark-bench-2.1.1_0.2.2-RELEASE.jar, {"spark-bench":{"spark-submit-config":[{"spark-args":{"master":"spark://34.215.233.221:7077"},"workload-suites":[{"benchmark-output":"console","descr":"datagen kmeans","workloads":[{"cols":4,"name":"data-generation-kmeans","output":"file:///tmp/kmeans-data.csv","parititions":32,"rows":10}]},{"benchmark-output":"console","descr":"run kmeans","workloads":[{"input":"file:///tmp/KMeen.csv","name":"kmeans"}]}]}]}}]
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
17/11/28 09:04:11 INFO SparkContext: Running Spark version 2.2.0
17/11/28 09:04:11 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
17/11/28 09:04:12 INFO SparkContext: Submitted application: com.ibm.sparktc.sparkbench.cli.CLIKickoff
17/11/28 09:04:12 INFO SecurityManager: Changing view acls to: ec2-user
17/11/28 09:04:12 INFO SecurityManager: Changing modify acls to: ec2-user
17/11/28 09:04:12 INFO SecurityManager: Changing view acls groups to:
17/11/28 09:04:12 INFO SecurityManager: Changing modify acls groups to:
17/11/28 09:04:12 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(ec2-user); groups with view permissions: Set(); users  with modify permissions: Set(ec2-user); groups with modify permissions: Set()
17/11/28 09:04:12 INFO Utils: Successfully started service 'sparkDriver' on port 37999.
17/11/28 09:04:12 INFO SparkEnv: Registering MapOutputTracker
17/11/28 09:04:12 INFO SparkEnv: Registering BlockManagerMaster
17/11/28 09:04:12 INFO BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
17/11/28 09:04:12 INFO BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
17/11/28 09:04:12 INFO DiskBlockManager: Created local directory at /tmp/blockmgr-f402d6cf-fed7-442c-abe5-a9ae2167051d
17/11/28 09:04:12 INFO MemoryStore: MemoryStore started with capacity 366.3 MB
17/11/28 09:04:12 INFO SparkEnv: Registering OutputCommitCoordinator
17/11/28 09:04:13 INFO Utils: Successfully started service 'SparkUI' on port 4040.
17/11/28 09:04:13 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started at http://172.31.15.16:4040
17/11/28 09:04:13 INFO SparkContext: Added JAR file:/home/ec2-user/new-spark-bench_2.1.1/lib/spark-bench-2.1.1_0.2.2-RELEASE.jar at spark://172.31.15.16:37999/jars/spark-bench-2.1.1_0.2.2-RELEASE.jar with timestamp 1511859853321
17/11/28 09:04:13 INFO StandaloneAppClient$ClientEndpoint: Connecting to master spark://34.215.233.221:7077...
17/11/28 09:04:33 INFO StandaloneAppClient$ClientEndpoint: Connecting to master spark://34.215.233.221:7077...
17/11/28 09:04:53 INFO StandaloneAppClient$ClientEndpoint: Connecting to master 

About this issue

  • Original URL
  • State: closed
  • Created 7 years ago
  • Comments: 22 (8 by maintainers)

Most upvoted comments

You’re right. I believe the README config example used to work fine, but #196 probably broke this. My apologies. I’ll fix this. (The config template does specify the HDFS version, though.)

In any case, are you able to launch a cluster without errors now?

Yes !! I have collected the spark job traces I need!

Thanks!!