Ubuntu下搭建单机Hadoop和Spark集群环境
Posted Linux公社
tags:
篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了Ubuntu下搭建单机Hadoop和Spark集群环境相关的知识,希望对你有一定的参考价值。
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/ac552cfe8b2f4aae80840d91ac1c56d0.jpg)
1、Java JDK8的安装
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/1f492ae0517c4fac8722b24f83908497.jpg)
[sudo] linuxidc 的密码:
linuxidc@linuxidc:~/www.linuxidc.com$ cd /opt/java/
linuxidc@linuxidc:/opt/java$ ls
jdk-8u231-linux-x64.tar.gz
linuxidc@linuxidc:/opt/java$ ls
jdk1.8.0_231 jdk-8u231-linux-x64.tar.gz
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/d1f16deea3dd48a1b0527a739579c5f2.jpg)
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/8b792c96b9b64239b9f7c8f97824e378.jpg)
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/d546f9b571e04e38918484eaf6db9544.jpg)
export JRE_HOME=/opt/java/jdk1.8.0_231/jre
export PATH=${JAVA_HOME}/bin:$PATH
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/2105e62bd9504f10ba1ca7d632ac364f.jpg)
java version "1.8.0_231"
Java(TM) SE Runtime Environment (build 1.8.0_231-b11)
Java HotSpot(TM) 64-Bit Server VM (build 25.231-b11, mixed mode)
linuxidc@linuxidc:/opt/java$
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/45ba3e3e02a8467e97d48b29101c161b.jpg)
2、安装Hadoop
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/82cdcbffeb404fd59f02a1e1bcce82b6.jpg)
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/252be590080c482180af2b0cf7b4502a.jpg)
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/d546f9b571e04e38918484eaf6db9544.jpg)
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/8b792c96b9b64239b9f7c8f97824e378.jpg)
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/0be91fafb09e46639d38a96c8b8d72fd.jpg)
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/c0b856be21a843168a7c30fd0adc493f.jpg)
<property>
<name>fs.defaultFS</name>
<value>hdfs://localhost:9000</value>
</property>
</configuration>
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/b8c2c0facfc24a01aa9ebbe089c11e3c.jpg)
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
</configuration>
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/05b9d38af8c0427baa81bca66d027ac2.jpg)
Generating public/private rsa key pair.
Your identification has been saved in /home/linuxidc/.ssh/id_rsa.
Your public key has been saved in /home/linuxidc/.ssh/id_rsa.pub.
The key fingerprint is:
SHA256:zY+ELQc3sPXwTBRfKlTwntek6TWVsuQziHtu3N/6L5w linuxidc@linuxidc
The key's randomart image is:
+---[RSA 2048]----+
| . o.*+. .|
| + B o o.|
| o o =o+.o|
| B..+oo=o|
| S.*. ==.+|
| +.o .oo.|
| .o.o... |
| oo .E .|
| .. o==|
+----[SHA256]-----+
linuxidc@linuxidc:~/www.linuxidc.com$ cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
linuxidc@linuxidc:~/www.linuxidc.com$ chmod 0600 ~/.ssh/authorized_keys
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/f92d6f8e1218416c9490eaae5dbc3eef.jpg)
linuxidc@linuxidc:~/www.linuxidc.com$ ssh localhost
Welcome to Ubuntu 18.04.3 LTS (GNU/Linux 5.4.0-999-generic x86_64)
* Management: https://landscape.canonical.com
* Support: https://ubuntu.com/advantage
* Canonical Livepatch is available for installation.
- Reduce system reboots and improve kernel security. Activate at:
https://ubuntu.com/livepatch
0 个安全更新。
Last login: Sat Nov 30 23:25:35 2019 from 127.0.0.1
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/a9d8d94688d844c2945957f3a16a65dc.jpg)
19/11/30 23:29:06 INFO namenode.NameNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG: host = linuxidc/127.0.1.1
STARTUP_MSG: args = [-format]
STARTUP_MSG: version = 2.7.7
......
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/3a8881d105d94db583358e7aba4918de.jpg)
Starting namenodes on [localhost]
localhost: starting namenode, logging to /opt/hadoop/hadoop-2.7.7/logs/hadoop-linuxidc-namenode-linuxidc.out
localhost: starting datanode, logging to /opt/hadoop/hadoop-2.7.7/logs/hadoop-linuxidc-datanode-linuxidc.out
Starting secondary namenodes [0.0.0.0]
The authenticity of host '0.0.0.0 (0.0.0.0)' can't be established.
ECDSA key fingerprint is SHA256:OSXsQK3E9ReBQ8c5to2wvpcS6UGrP8tQki0IInUXcG0.
Are you sure you want to continue connecting (yes/no)? yes
0.0.0.0: Warning: Permanently added '0.0.0.0' (ECDSA) to the list of known hosts.
0.0.0.0: starting secondarynamenode, logging to /opt/hadoop/hadoop-2.7.7/logs/hadoop-linuxidc-secondarynamenode-linuxidc.out
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/1eeb7bfb0fdb4b548dce442c18b9f13d.jpg)
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/c8e7684d2c8549cb9d2458624478654d.jpg)
3、Scala安装:
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/12871f99434348a5a998b38c87ed023f.jpg)
[sudo] linuxidc 的密码:
linuxidc@linuxidc:~/下载$ cd /opt/scala
linuxidc@linuxidc:/opt/scala$ ls
scala-2.11.8
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/20feca6485194c11b4f72d5955171a9f.jpg)
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/265454b59b7243a98c94ceba737a8877.jpg)
4、安装spark
[sudo] linuxidc 的密码:
linuxidc@linuxidc:~/www.linuxidc.com$ cd /opt/spark/
linuxidc@linuxidc:/opt/spark$ ls
spark-2.4.4-bin-hadoop2.7.tgz
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/6a41fb12a2cc4ef28ac24f47a6d17c68.jpg)
[sudo] linuxidc 的密码:
linuxidc@linuxidc:/opt/spark$ ls
spark-2.4.4-bin-hadoop2.7 spark-2.4.4-bin-hadoop2.7.tgz
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/5d37c89b44c749f39368e42417a68f91.jpg)
export PATH=${JAVA_HOME}/bin:${HADOOP_HOME}/bin:${SPARK_HOME}/bin:$PATH
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/dedeb01a2cf346eaa488e5ac2c2b77e5.jpg)
export HADOOP_HOME=/opt/hadoop/hadoop-2.7.7
export HADOOP_CONF_DIR=/opt/hadoop/hadoop-2.7.7/etc/hadoop
export SPARK_HOME=/opt/spark/spark-2.4.4-bin-hadoop2.7
export SCALA_HOME=/opt/scala/scala-2.11.8
export SPARK_MASTER_IP=127.0.0.1
export SPARK_MASTER_PORT=7077
export SPARK_MASTER_WEBUI_PORT=8099
export SPARK_WORKER_CORES=3
export SPARK_WORKER_INSTANCES=1
export SPARK_WORKER_MEMORY=5G
export SPARK_WORKER_WEBUI_PORT=8081
export SPARK_EXECUTOR_CORES=1
export SPARK_EXECUTOR_MEMORY=1G
export LD_LIBRARY_PATH=${LD_LIBRARY_PATH}:$HADOOP_HOME/lib/native
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/576b367ab4594883afa9148a556471bb.jpg)
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/6422fed97f78441d920ba53467c80faa.jpg)
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/ece833f963d74a228f845eb02b979941.jpg)
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/3120f8aa4f154dbb918fa43ff991ca25.jpg)
Linux公社的RSS地址:https://www.linuxidc.com/rssFeed.aspx
![Ubuntu下搭建单机Hadoop和Spark集群环境](https://image.cha138.com/20210413/ce966a1494344271b74be771ed5436a9.jpg)
以上是关于Ubuntu下搭建单机Hadoop和Spark集群环境的主要内容,如果未能解决你的问题,请参考以下文章
大数据学习系列之七 ----- Hadoop+Spark+Zookeeper+HBase+Hive集群搭建 图文详解