docker ubuntu下载jdk:
在opt里创建两个目录:software和modules
software用来存放安装包,modules用来存放解压后的文件
首先用命令在官网下载jdk:
wget 'https://download.oracle.com/java/17/latest/jdk-17_linux-x64_bin.tar.gz'
解压到modules目录下,可以给它重新改一个简单的名字,我叫它jdk17
mv jdk-17.0.1 jdk17
然后在etc/profile里追加:
export JAVA_HOME=/opt/modules/jdk17 export JRE_HOME=${JAVA_HOME}/jre export CLASSPATH=.:${JAVA_HOME}/lib:${JRE_HOME}/lib export PATH=${JAVA_HOME}/bin:$PATH
执行profile文件:
source etc/profile
下载scala:
wget'https://downloads.lightbend.com/scala/2.11.8/scala-2.11.8.tgz'
解压.
然后在etc/profile里追加::
export SCALA_HOME=/opt/modules/scala2 export PATH=${SCALA_HOME}/bin:$PATH
执行profile文件:
source etc/profile
测试是否安装成功:
scala -version
安装spark:
wget 'https://dlcdn.apache.org/spark/spark-3.2.0/spark-3.2.0-bin-hadoop3.2-scala2.13.tgz'
解压。
然后在etc/profile里追加::
export SPARK_HOME=/opt/modules/spark export PATH=${SPARK_HOME}/bin:$PATH
对了,我们可以将三个写在一起:
export JAVA_HOME=/opt/modules/jdk17 export JRE_HOME=${JAVA_HOME}/jre export CLASSPATH=.:${JAVA_HOME}/lib:${JRE_HOME}/lib export SCALA_HOME=/opt/modules/scala2 export SPARK_HOME=/opt/modules/spark export PATH=${JAVA_HOME}/bin:${SCALA_HOME}/bin:${SPARK_HOME}/bin:${SPARK_HOME}/sbin:$PATH
template文件cp:
root@b94e802c7b46:/opt/modules/spark/conf# cp log4j.properties.template log4j.properties root@b94e802c7b46:/opt/modules/spark/conf# cp spark-defaults.conf.template spark-defaults.conf root@b94e802c7b46:/opt/modules/spark/conf# cp spark-env.sh.template spark-env.sh root@b94e802c7b46:/opt/modules/spark/conf# cp workers.template workers
在root@b94e802c7b46:/opt/modules/spark/conf#
vim spark-env.sh
export JAVA_HOME=/opt/modules/jdk17 export SCALA_HOME=/opt/modules/scala2
验证成功: