在Google Compute Engine上部署了一個Hadoop(Yarn + Spark)羣集,其中包含一個主機&兩個從機。當我運行以下shell腳本時:SparkPi程序繼續在Yarn/Spark/Google計算引擎下運行
spark-submit --class org.apache.spark.examples.SparkPi --master yarn-cluster --num -executions 1 --driver-memory 1g --executor- 1G內存--executor-核心1 /home/hadoop/spark-install/lib/spark-examples-1.1.0-hadoop2.4.0.jar 10
工作只是保持運行&每一秒鐘我得到一個消息與此類似:
15/02/06 22:47:12 INFO yarn.Client: Application report from ResourceManager:
application identifier: application_1423247324488_0008<br>
appId: 8<br>
clientToAMToken: null<br>
appDiagnostics:<br>
appMasterHost: hadoop-w-zrem.c.myapp.internal<br>
appQueue: default<br>
appMasterRpcPort: 0<br>
appStartTime: 1423261517468<br>
yarnAppState: RUNNING<br>
distributedFinalState: UNDEFINED<br>
appTrackingUrl: http://hadoop-m-xxxx:8088/proxy/application_1423247324488_0008/<br>
appUser: achitre
我出現以下情況例外,當我使用的紗線客戶端: 異常線程「main」 org.apache.spark.SparkException: SparkContext已在org.apache.spark.SparkContext.runJob(SparkContext已經停運 。在org.apache.spark.SparkContext.runJob(SparkContext.scala:1158) at org.apache.spark.rdd.RDD.reduce(RDD.scala:861) at org.apache.spark。例如.SparkPi $ .main(SparkPi.scala:35) at org.apache.spark.examples.SparkPi.main(SparkPi.scala) – DilTeam 2015-02-07 17:01:12
您是否在完成作業 - 獲取結果或收到結果之前收到它? – baju 2015-02-07 22:19:16
當我運行'yarn logs -applicationId我注意到下面的異常; 線程「主」java.lang.NumberFormatException中的異常:對於輸入字符串:「5999m」 –
DilTeam
2015-02-08 03:29:10