vi /opt/spark-3.3.1/dev/make-distribution.sh export MAVEN_OPTS="${MAVEN_OPTS:--Xmx4g -XX:ReservedCodeCacheSize=2g -XX:MaxDirectMemorySize=256m}" MVN="/opt/softwear/apache-maven-3.6.3/bin/mvn" |
添加如下 # spark版本 VERSION=3.5.2 # scala版本 SCALA_VERSION=2.12 # hadoop版本 SPARK_HADOOP_VERSION=3.3.6 # 开启hive SPARK_HIVE=1 表明如下: #VERSION=$("$MVN" help:evaluate -Dexpression=project.version $@ \ # | grep -v "INFO"\ # | grep -v "WARNING"\ # | tail -n 1) #SCALA_VERSION=$("$MVN" help:evaluate -Dexpression=scala.binary.version $@ \ # | grep -v "INFO"\ # | grep -v "WARNING"\ # | tail -n 1) #SPARK_HADOOP_VERSION=$("$MVN" help:evaluate -Dexpression=hadoop.version $@ \ # | grep -v "INFO"\ # | grep -v "WARNING"\ # | tail -n 1) #SPARK_HIVE=$("$MVN" help:evaluate -Dexpression=project.activeProfiles -pl sql/hive $@ \ # | grep -v "INFO"\ # | grep -v "WARNING"\ # | fgrep --count "<id>hive</id>";\ # # Reset exit status to 0, otherwise the script stops here if the last grep finds nothing\ # # because we use "set -o pipefail" # echo -n) |
/opt/spark-3.5.2/dev/make-distribution.sh --name 3.5.2 --tgz -Pyarn -Phive -Phive-thriftserver -Dhadoop.version=3.3.6 -DskipTests |
sql("LOAD DATA LOCAL INPATH 'examples/src/main/resources/kv1.txt' INTO TABLE src") |
执行spark sql 报错: org.apache.spark.sql.AnalysisException: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.thrift.transport.TTransportException hive日志报错:metastore.RetryingHMSHandler: java.lang.IndexOutOfBoundsException: Index: 0, Size: 0 at java.util.ArrayList.rangeCheck(ArrayList.java:657) |
# 指定提交到 yarn 运行 spark.master yarn # 开启日志并存储到 HDFS 上 spark.eventLog.enabled true spark.eventLog.dir hdfs://nameservice1/spark-logDir # 指定每个执行器的内存 spark.executor.memory 1g # 指定每个调度器的内存 spark.driver.memory 1g spark.history.fs.logDirectory hdfs://nameservice1/spark-logDir spark.yarn.historyServer.address {spark-history-server IP}:18080 spark.history.ui.port 18080 # 需提前上传到 hdfs上 spark.yarn.jars hdfs://nameservice1/spark-jars/*.jar spark.sql.hive.metastore.version 3.1.3 spark.sql.hive.metastore.jars path # 需将jars上传到 yarn 每个nodemanager节点对应目录下 spark.sql.hive.metastore.jars.path file:///opt/hive/apache-hive-3.1.3-bin/lib/*.jar |
CREATE TABLE IF NOT EXISTS test.test_1202 (name string, age int) org.apache.spark.sql.AnalysisException: java.lang.ExceptionInInitializerError: null , Caused by: java.lang.IllegalArgumentException: URI scheme is not "file" at java.io.File.<init>(File.java:421) at org.apache.hadoop.hive.conf.HiveConf.findConfigFile(HiveConf.java:179) at org.apache.hadoop.hive.conf.HiveConf.<clinit>(HiveConf.java:146) |
欢迎光临 ToB企服应用市场:ToB评测及商务社交产业平台 (https://dis.qidao123.com/) | Powered by Discuz! X3.4 |