1
1
FROM openjdk:8u212-b04-jdk-stretch
2
- LABEL author="Nathaniel Vala " email="nathanielvala@hotmail .com"
3
- LABEL version="0.2 "
2
+ LABEL author="Pradeep Sekar " email="pradipraniboy@gmail .com"
3
+ LABEL version="0.1 "
4
4
5
5
ENV DAEMON_RUN=true
6
- ENV SPARK_VERSION=2.4.3
7
- ENV HADOOP_VERSION=2.7
8
- ENV SCALA_VERSION=2.12.4
6
+ ENV SPARK_VERSION=3.0.1
7
+ ENV HADOOP_VERSION=3.2
8
+ ENV SCALA_VERSION=2.12.10
9
9
ENV SCALA_HOME=/usr/share/scala
10
10
ENV SPARK_HOME=/spark
11
11
12
-
12
+ # # Get all certificates.
13
13
RUN apt-get update && apt-get install -y curl vim wget software-properties-common ssh net-tools ca-certificates jq
14
14
15
- # apt update && apt -y upgrade \
16
- # apt install -y wget ca-certificates && \
17
- # apt install -y curl bash jq && \
15
+
16
+ # # Install all initial lirary
18
17
RUN cd "/tmp" && \
19
18
wget --no-verbose "https://downloads.typesafe.com/scala/${SCALA_VERSION}/scala-${SCALA_VERSION}.tgz" && \
20
19
tar xzf "scala-${SCALA_VERSION}.tgz" && \
@@ -24,34 +23,18 @@ RUN cd "/tmp" && \
24
23
ln -s "${SCALA_HOME}/bin/" * "/usr/bin/" && \
25
24
rm -rf "/tmp/" *
26
25
27
-
28
- # RUN apk add --no-cache --virtual=.build-dependencies wget ca-certificates && \
29
- # apk add --no-cache bash curl jq && \
30
- # cd "/tmp" && \
31
- # wget --no-verbose "https://downloads.typesafe.com/scala/${SCALA_VERSION}/scala-${SCALA_VERSION}.tgz" && \
32
- # tar xzf "scala-${SCALA_VERSION}.tgz" && \
33
- # mkdir "${SCALA_HOME}" && \
34
- # rm "/tmp/scala-${SCALA_VERSION}/bin/"*.bat && \
35
- # mv "/tmp/scala-${SCALA_VERSION}/bin" "/tmp/scala-${SCALA_VERSION}/lib" "${SCALA_HOME}" && \
36
- # ln -s "${SCALA_HOME}/bin/"* "/usr/bin/" && \
37
- # apk del .build-dependencies && \
38
- # rm -rf "/tmp/"*
39
-
40
-
41
26
# Add Dependencies for PySpark
42
27
RUN apt-get install -y python3 python3-pip python3-numpy python3-matplotlib python3-scipy python3-pandas python3-simpy
43
28
RUN update-alternatives --install "/usr/bin/python" "python" "$(which python3)" 1
44
29
45
30
46
- # Scala instalation
31
+ # Scala instalation ---> Check the spark and sbt versions
47
32
RUN export PATH="/usr/local/sbt/bin:$PATH" && apt update && apt install ca-certificates wget tar && mkdir -p "/usr/local/sbt" && wget -qO - --no-check-certificate "https://github.com/sbt/sbt/releases/download/v1.2.8/sbt-1.2.8.tgz" | tar xz -C /usr/local/sbt --strip-components=1 && sbt sbtVersion
48
33
49
34
RUN wget --no-verbose http://apache.mirror.iphh.net/spark/spark-${SPARK_VERSION}/spark-${SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}.tgz && tar -xvzf spark-${SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}.tgz \
50
35
&& mv spark-${SPARK_VERSION}-bin-hadoop${HADOOP_VERSION} spark \
51
36
&& rm spark-${SPARK_VERSION}-bin-hadoop${HADOOP_VERSION}.tgz
52
37
53
-
54
-
55
38
# Fix the value of PYTHONHASHSEED
56
39
# Note: this is needed when you use Python 3.3 or greater
57
40
ENV PYTHONHASHSEED 1
0 commit comments