There are multiple issues during running spark hive example HiveFromSpark:
- Copy hive-site.xml to ${SPARK_HOME}/conf/
- Correct command:
${SPARK_HOME}/bin/spark-submit --master yarn-cluster --driver-java-options "-XX:MaxPermSize=1G" --jars ../lib/datanucleus-api-jdo-3.2.1.jar,../lib/datanucleus-rdbms-3.2.1.jar,../lib/datanucleus-core-3.2.2.jar,../lib/mysql-connector-java-5.1.31-bin.jar --files ../conf/hive-site.xml --class org.apache.spark.examples.sql.hive.HiveFromSpark ../lib/spark-examples-1.3.0-cdh5.4.1-hadoop2.6.0-cdh5.4.1.jar - java.sql.SQLException: No suitable driver found for jdbc:mysql:
Add mysql-connector*.jar into hadoop classpath. One way is locate this jar in ${HADOOP_HOME}/share/hadoop/yarn/lib/ - java.lang.OutOfMemoryError thrown from the UncaughtExceptionHandler in thread "Driver":
To add driver memory size via increasing "--driver-java-options "-XX:MaxPermSize=1G"" option in spark-submit command.