Running Hadoop in docker and image making method


Repeat the wheel, here we use repackaging to generate a Hadoop image based on docker;

Hadoop cluster depends on the following software: JDK, SSH, etc., so as long as these two items are also related to Hadoop, they can be packaged into the image;

Configuration file preparation

1. Hadoop related configuration file: Core- site.xml 、hdfs- site.xml 、mapred- site.xml 、yarn- site.xml 、slaves、hadoop–
2. SSH configuration file: SSH_ config
3. Hadoop cluster startup file: start-

Making a mirror image

1. Installation dependency

RUN apt-get update && \
 apt-get install -y openssh-server openjdk-8-jdk wget

2. Download Hadoop package

RUN wget && \
tar -xzvf hadoop-2.10.0.tar.gz && \
mv hadoop-2.10.0 /usr/local/hadoop && \
rm hadoop-2.10.0.tar.gz && \
rm /usr/local/hadoop/share/doc -rf

3. Configure environment variables

ENV JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 
ENV HADOOP_HOME=/usr/local/hadoop 
ENV PATH=$PATH:/usr/local/hadoop/bin:/usr/local/hadoop/sbin

4. Generate SSH key for node secret free login

RUN ssh-keygen -t rsa -f ~/.ssh/id_rsa -P '' && \
cat ~/.ssh/ >> ~/.ssh/authorized_keys

5. Create Hadoop related directories, copy relevant configuration files, add execution permissions to relevant files, and finally format namenode. When each node starts, start SSH service;

RUN mkdir -p ~/hdfs/namenode && \ 
mkdir -p ~/hdfs/datanode && \
mkdir $HADOOP_HOME/logs
COPY config/* /tmp/
#Replication SSH, Hadoop configuration related
RUN mv /tmp/ssh_config ~/.ssh/config && \
mv /tmp/ /usr/local/hadoop/etc/hadoop/ && \
mv /tmp/hdfs-site.xml $HADOOP_HOME/etc/hadoop/hdfs-site.xml && \ 
mv /tmp/core-site.xml $HADOOP_HOME/etc/hadoop/core-site.xml && \
mv /tmp/mapred-site.xml $HADOOP_HOME/etc/hadoop/mapred-site.xml && \
mv /tmp/yarn-site.xml $HADOOP_HOME/etc/hadoop/yarn-site.xml && \
mv /tmp/slaves $HADOOP_HOME/etc/hadoop/slaves && \
mv /tmp/ ~/ && \
mv /tmp/ ~/
#Add execution permission
RUN chmod +x ~/ && \
chmod +x ~/ && \
chmod +x $HADOOP_HOME/sbin/ && \
chmod +x $HADOOP_HOME/sbin/ 
# format namenode
RUN /usr/local/hadoop/bin/hdfs namenode -format

Running Hadoop cluster in docker

After the image is generated through the dockerfile above, the Hadoop cluster can be built using the image generated above. Here, one master and two slave nodes are started;

Add bridging network:

docker network create --driver=bridge solinx-hadoop

Start the master node:

docker run -itd --net=solinx-hadoop -p 10070:50070 -p 8088:8088 --name solinx-hadoop-master --hostname solinx-hadoop-master solinx/hadoop:0.1

Start the slave1 node:

docker run -itd --net=solinx-hadoop --name solinx-hadoop-slave1 --hostname solinx-hadoop-slave1 solinx/hadoop:0.1

Start the slave 2 node:

docker run -itd --net=solinx-hadoop --name solinx-hadoop-slave2 --hostname solinx-hadoop-slave1 solinx/hadoop:0.1

Enter the master node and execute the Hadoop cluster startup script


The above is the introduction of Hadoop and image production in docker. I hope it will be helpful to you. If you have any questions, please leave me a message, and the editor will reply to you in time. Thank you very much for your support to the developeppaer website!
If you think this article is helpful to you, welcome to reprint, please indicate the source, thank you!