kafka+zookeeper部署!
Posted handsomeboy-东
tags:
篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了kafka+zookeeper部署!相关的知识,希望对你有一定的参考价值。
kafka+zookeeper
基本环境部署
- 设备准备:
节点 ip地址 部署服务
node1 192.168.118.44 jdk,kafka,zookeeper
node2 192.168.118.55 jdk,kafka,zookeeper
node3 192.168.118.66 jdk,kafka,zookeeper
- 环境准备,三台同时部署
##关闭防火墙
systemctl stop firewalld
systemctl disable firewalld
setenforce 0
##设置时间同步
ntpdate ntp1.aliyun.com
##创建用户
useradd appUser
部署jdk环境
- 三个节点同部署
解压jdk包
tar xf jdk1.8.0_221.tar.gz -C /usr/local/
##设置环境变量
vim /etc/profile
export JAVA_HOME=/usr/local/jdk1.8.0_221
export JR_HOME=$JAVA_HOME/jre
export CLASSPATH=.:$JAVA_HOME/lib:$JRE_HOME/lib
export PATH=$JAVA_HOME/bin:$PATH
source /etc/profile
java -version ##查看版本
##创建对应目录
mkdir -p /export/packages/kafka,zookeeper,java
mkdir -p /export/data/kafka,zookeeper
mkdir -p /export/Logs/kafkalog,zkdatalog
mkdir -p /export/servers
部署zookeeper
- 三个节点同时部署
##解压zookeeper包
tar xf zookeeper-3.4.13.tar -C /export/packages/zookeeper/
##备份、修改配置文件
cd /export/packages/zookeeper/zookeeper-3.4.13/conf
cp zoo_sample.cfg zoo.cfg
vim zoo.cfg
tickTime=2000 #通信心跳时间,Zookeeper服务器与客户端心跳时间,单位毫秒
initLimit=10 #Leader和Follower初始连接时能容忍的最多心跳数(tickTime的数量),这里表示为10*2s
syncLimit=5 #Leader和Follower之间同步通信的超时时间,这里表示如果超过5*2s,Leader认为Follwer死掉,并从服务器列表中删除Follwer
dataDir=/export/packages/zookeeper/zookeeper-3.4.13/data ●修改,指定保存Zookeeper中的数据的目录,目录需要单独创建
dataLogDir=/export/packages/zookeeper/zookeeper-3.4.13/logs ●添加,指定存放日志的目录,目录需要单独创建
clientPort=2181 #客户端连接端口
#添加集群信息
server.1=192.168.226.128:3188:3288
server.2=192.168.226.130:3188:3288
server.3=192.168.226.131:3188:3288
#在"每个"节点的dataDir指定的目录下创建一个 myid 的文件
echo 1 > /export/packages/zookeeper/zookeeper-3.4.13/data/myid ##node1节点
echo 2 > /export/packages/zookeeper/zookeeper-3.4.13/data/myid ##node2节点
echo 3 > /export/packages/zookeeper/zookeeper-3.4.13/data/myid ##node3节点
##配置zookeeper启动脚本
vim /etc/init.d/zookeeper
#!/bin/bash
#chkconfig:2345 20 90
#description:Zookeeper Service Control Script
ZK_HOME='/export/packages/zookeeper/zookeeper-3.4.13/'
case $1 in
start)
echo "---------- zookeeper 启动 ------------"
$ZK_HOME/bin/zkServer.sh start
;;
stop)
echo "---------- zookeeper 停止 ------------"
$ZK_HOME/bin/zkServer.sh stop
;;
restart)
echo "---------- zookeeper 重启 ------------"
$ZK_HOME/bin/zkServer.sh restart
;;
status)
echo "---------- zookeeper 状态 ------------"
$ZK_HOME/bin/zkServer.sh status
;;
*)
echo "Usage: $0 start|stop|restart|status"
esac
##设置开机启动
chmod +x /etc/init.d/zookeeper
chkconfig --add zookeeper
service zookeeper start ##启动zookeeper
service zookeeper status ##查看zookeeper集群状态
---------- zookeeper 状态 ------------
ZooKeeper JMX enabled by default
Using config: /export/packages/zookeeper/zookeeper-3.4.13/bin/../conf/zoo.cfg
Mode: follower
[root@node1 export]# netstat -antp | grep 2181
tcp6 0 0 :::2181 :::* LISTEN 89681/java
tcp6 0 0 127.0.0.1:37636 127.0.0.1:2181 TIME_WAIT -
tcp6 0 0 192.168.118.44:48274 192.168.118.66:2181 ESTABLISHED 90562/java
kafka部署
- 三个节点同时部署
tar xf kafka_2.11-2.3.0.tar -C /export/packages/kafka/
cd /export/servers/kafka/config/
cp -r server.properties server.properties.back
vim /export/servers/kafka/config/server.properties ##修改配置文件
node1节点:
21 broker.id=0 #broker的全局唯一编号,每个节点不能重复
31 listeners=PLAINTEXT://192.168.118.44:9092 #指定监听的IP和端口,如果修改每个broker的IP需区分开
60 log.dirs=/export/Logs/kafkalog #kafka运行日志存放的路径,也是数据存放的路径
123 zookeeper.connect=192.168.118.44:2181,192.168.18.55:2181,192.168.118.66:2181 配置连接Zookeeper集群地址
node2节点:
broker.id=1
listeners=PLAINTEXT://192.168.118.55:9092
log.dirs=/export/Logs/kafkalog
zookeeper.connect=192.168.118.44:2181,192.168.18.55:2181,192.168.118.66:2181
node3节点:
broker.id=2
listeners=PLAINTEXT://192.168.118.66:9092
log.dirs=/export/Logs/kafkalog
zookeeper.connect=192.168.118.44:2181,192.168.18.55:2181,192.168.118.66:2181
- 修改文件属主属组,三节点同时部署
chown -R appUser.appUser /export/servers/kafka/
chown -R appUser.appUser /export/packages/kafka/
chown -R appUser.appUser /export/data/kafka/
chown -R appUser.appUser /export/Logs/kafkalog/
- 设置kafka启动脚本,三节点同时部署
vim /etc/init.d/kafka
#!/bin/bash
#chkconfig:2345 22 88
#description:Kafka Service Control Script
KAFKA_HOME='/export/servers/kafka/'
case $1 in
start)
echo "---------- Kafka 启动 ------------"
$KAFKA_HOME/bin/kafka-server-start.sh -daemon $KAFKA_HOME/config/server.properties
;;
stop)
echo "---------- Kafka 停止 ------------"
$KAFKA_HOME/bin/kafka-server-stop.sh
;;
restart)
$0 stop
$0 start
;;
status)
echo "---------- Kafka 状态 ------------"
count=$(ps -ef | grep kafka | egrep -cv "grep|$$")
if [ "$count" -eq 0 ];then
echo "kafka is not running"
else
echo "kafka is running"
fi
;;
*)
echo "Usage: $0 start|stop|restart|status"
esac
##设置自启
chmod +x /etc/init.d/kafka
chkconfig --add kafka
service kafka start ##启动kafka
[root@node1 export]# service kafka status
---------- Kafka 状态 ------------
kafka is running
[root@node1 export]# netstat -antp | grep 9092
tcp6 0 0 192.168.118.44:9092 :::* LISTEN 90562/java
tcp6 0 0 192.168.118.44:34968 192.168.118.66:9092 ESTABLISHED 90562/java
- 进入用户配置,这里再node1节点(三节点都可)
[appUser@node1 export]$ /export/servers/kafka/bin/kafka-topics.sh --create --zookeeper 192.168.118.44:2181 --replication-factor 2 --partitions 1 --topic whd
Created topic whd.
解释:
replication-factor 2:复制两份
partitions 1:创建一个分区
topic :主题
#创建一个broker分别在两台机器上,在一台服务器上创建一个发布者
##查看topic
[appUser@node1 export]$ /export/servers/kafka/bin/kafka-topics.sh --list --zookeeper 192.168.118.44:2181
test
whd
##查看topic状态
[appUser@node1 export]$ /export/servers/kafka/bin/kafka-topics.sh --describe --zookeeper 192.168.118.44:2181 --topic test
Topic:test PartitionCount:1 ReplicationFactor:2 Configs:
Topic: test Partition: 0 Leader: 2 Replicas: 2,0 Isr: 2,0
##调优实践,生产环境适当调大jvm内存,设置环境变量
export KAFKA_HEAP_OPTS="-Xms4g -Xmx4g"
以上是关于kafka+zookeeper部署!的主要内容,如果未能解决你的问题,请参考以下文章