報錯如下: WARN [2017-06-27 15:47:59,777] ({pool-2-thread-2} Logging.scala[logWarning]:66) - SPARK_CLASSPATH was detected (set to '/home/raini/spark/lib/mysql-connector-java-5.1.38-bin.jar:').This is deprecated in Spark 1.0+. Please instead use: - ./spark-submit with --driver-class-path to augment the driver classpath - spark.executor.extraClassPath to augment the executor classpath WARN [2017-06-27 15:47:59,778] ({pool-2-thread-2} Logging.scala[logWarning]:66) - Setting 'spark.executor.extraClassPath' to '/home/raini/spark/lib/mysql-connector-java-5.1.38-bin.jar:' as a work-around. ERROR [2017-06-27 15:47:59,780] ({pool-2-thread-2} Logging.scala[logError]:91) - Error initializing SparkContext. org.apache.spark.SparkException: Found both spark.driver.extraClassPath and SPARK_CLASSPATH. Use only the former. at org.apache.spark.SparkConf anonfun$apply$8.apply(SparkConf.scala:543) at org.apache.spark.SparkConf anonfun$apply$8.apply(SparkConf.scala:541) at scala.collection.immutable.List.foreach(List.scala:381) at org.apache.spark.SparkConf$$anonfun$validateSettings$7.apply(SparkConf.scala:541) at org.apache.spark.SparkConf$$anonfun$validateSettings$7.apply(SparkConf.scala:529) 解決: 1.提交的作業(yè)可以通過加入--driver-class-path參數(shù)來設(shè)置driver的classpath。 $ bin/spark-submit --master local[2] --driver-class-path lib/mysql-connector-java-5.1.35.jar --class spark.SparkToJDBC ./spark-test_2.10-1.0.jar export SPARK_CLASSPATH=$SPARK_CLASSPATH:/iteblog/com/mysql-connector-java-5.1.35.jar 所以,刪掉一個配置即可,這里刪掉了spark配置項:export SPARK_CLASSPATH=... |
|
來自: 飲茶仙人 > 《大數(shù)據(jù)》