Spark--安装与配置

Posted

tags:

篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了Spark--安装与配置相关的知识,希望对你有一定的参考价值。

一,环境

机器名

IP

集群状态

zookeeper

hadoop01

176.129.8.111

active

follower

hadoop02

176.129.8.112

standby

leader

hadoop03

176.129.8.113

observer

CentOS6.5

JDK1.8.0

Hadoop2.7.1

Zookeeper3.7.1

Scala-2.13.0

二,下载及解压

机器:hadoop01

下载地址​http://spark.apache.org/releases/spark-release-2-2-2.html​​

解压到/home/software下

三,配置

机器:hadoop01

进入conf目录

1,复制spark-env.sh

cp spark-env.sh.template spark-env.sh

 复制缓存的文件spark-env.sh.template改为spark识别的文件spark-env.sh,在文件尾部编辑如下内容:

export JAVA_HOME=/home/software/jdk1.8.0

export SCALA_HOME=/home/software/scala-2.13.0

export HADOOP_HOME=/home/hadoop-2.7.1

export HADOOP_CONF_DIR=/home/hadoop-2.7.1/etc/hadoop

export SPARK_MASTER_IP=hadoop01

export SPARK_WORKER_MEMORY=2g

export SPARK_WORKER_CORES=2

export SPARK_WORKER_INSTANCES=1

 

2,复制slaves

cp slaves.template slaves

编辑如下内容:

hadoop02

hadoop03

3,复制到hadoop02和hadoop03上

[root@hadoop01 software]# scp -r spark-2.2.2 hadoop02:/home/software/

[root@hadoop01 software]# scp -r spark-2.2.2 hadoop03:/home/software/

四,启动

1,进入spark/sbin下

sh start-all.sh

2,在web上访问

​​http://hadoop01:8080​​

Worker Id

Address

State

Cores

Memory

worker-20180919193657-176.129.8.113-46491

176.129.8.113:46491

ALIVE

2 (0 Used)

2.0 GB (0.0 B Used)

worker-20180919193658-176.129.8.112-38025

176.129.8.112:38025

ALIVE

2 (0 Used)

2.0 GB (0.0 B Used)

3,进入bin/spark-shell

a,启动并加载jar包准备进入hive库

[root@hadoop01 bin]# sh spark-shell --master local[3] --jars /home/software/hive-1.2.0/lib/mysql-connector-java-5.1.38-bin.jar

b,查看库

scala> spark.sql("show databases").show;

c,创建库

scala> spark.sql("create database test").show;

d,使用库

scala> spark.sql("use test").show;

e,创建表

scala> spark.sql("create table student(id int, name string,age int) row format delimited fields terminated by ,").show;

f,增加数据

spark.sql("insert into student values(1,Zenghaoxuan,10)").show;

g,查看数据

scala> spark.sql("select * from student").show;

+---+---------+---+

| id|     name|age|

+---+---------+---+

|  1|Zenghaoxuan| 10|

+---+---------+---+

以上是关于Spark--安装与配置的主要内容,如果未能解决你的问题,请参考以下文章

Spark的安装与配置

Spark 2.2.0 安装与配置

大数据高可用集群环境安装与配置(09)——安装Spark高可用集群

spark 2.0.0集群安装与hive on spark配置

Spark 安装和配置

Spark--安装与配置