Kafka+Zookeeper+Storm的docker化

Posted F11Team

tags:

篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了Kafka+Zookeeper+Storm的docker化相关的知识,希望对你有一定的参考价值。

please read in markdown format;
# Kafka+Zookeeper+Storm的docker化
> 因为项目需要,需要把kafka、zookeeper、storm等服务组件docker化,在借鉴相关的开源dockerfile文件后,终于完成,以此记录。

1. 首先是建立基础的Linux镜像,针对本项目做了定制。
```
FROM centos
MAINTAINER cc.zhang1024@gmail.com
COPY jq /usr/bin/
RUN yum  update -y && \
yum install wget -y && \
yum install openssh -y && \
yum install openssh-server -y && \
yum install vim -y && \
yum install zip unzip -y && \
yum install openssh-clients -y && \
yum groupinstall "Development Tools" -y && \
yum install python-setuptools -y && \
easy_install supervisor && \
echo 'root:root' | chpasswd && \
wget https://zcc2018.oss-cn-beijing.a ... 71-linux-x64.tar.gz &&\
tar -zxvf jdk-8u171-linux-x64.tar.gz -C /opt/ &&\
JAVA_HOME=/opt/jdk1.8.0_171 && \
CLASSPATH=$CLASSPATH:$JAVA_HOME/lib:$JAVA_HOME/jre/lib && \
PATH=$JAVA_HOME/bin:$JAVA_HOME/jre/bin:$PATH:HOME/bin && \
echo "export JAVA_HOME=$JAVA_HOME">>/etc/profile && \
echo "export CLASSPATH=$CLASSPATH">>/etc/profile && \
echo "export PATH=$PATH">>/etc/profile && \
ssh-keygen -A && \
chmod +x /usr/bin/jq && \
mkdir /var/run/sshd  && \
mkdir /var/log/supervisor -p && \
sed -i 's/PermitRootLogin without-password/PermitRootLogin yes/' /etc/ssh/sshd_config
ENV JAVA_HOME /opt/jdk1.8.0_171
ENV PATH $PATH:$JAVA_HOME/bin
ADD supervisord.conf /etc/supervisor/supervisord.conf
EXPOSE 22

```

· 其中jq需要从网上下载放在同一目录下。

2. 进行kafka的环境安装[参考github项目->https://github.com/wurstmeister/kafka-docker]
```
FROM centos
ARG kafka_version=1.0.0
ARG scala_version=2.12
ARG glibc_version=2.27-r0
ENV KAFKA_VERSION=$kafka_version \
    SCALA_VERSION=$scala_version \
    KAFKA_HOME=/opt/kafka \
    GLIBC_VERSION=$glibc_version
ENV PATH=${PATH}:${KAFKA_HOME}/bin
COPY download-kafka.sh start-kafka.sh broker-list.sh create-topics.sh /tmp/
RUN  chmod a+x /tmp/*.sh \
&& mv /tmp/start-kafka.sh /tmp/broker-list.sh /tmp/create-topics.sh /usr/bin \
&& sync && /tmp/download-kafka.sh \
&& tar xfz /tmp/kafka_${SCALA_VERSION}-${KAFKA_VERSION}.tgz -C /opt \
&& rm /tmp/kafka_${SCALA_VERSION}-${KAFKA_VERSION}.tgz \
&& ln -s /opt/kafka_${SCALA_VERSION}-${KAFKA_VERSION} /opt/kafka \
&& rm /tmp/* -rf
VOLUME ["/kafka"]
CMD ["start-kafka.sh"]

```

- 相关脚本请去github下载【本dockerfile也是参考该开源项目->https://github.com/wurstmeister/kafka-docker

3. zookeeper->[参考项目->https://github.com/wurstmeister/zookeeper-docker]
```
FROM centos
ENV ZOOKEEPER_VERSION 3.4.12
RUN wget -q http://mirrors.tuna.tsinghua.edu.cn/apache/zookeeper/zookeeper-${ZOOKEEPER_VERSION}/zookeeper-${ZOOKEEPER_VERSION}.tar.gz && \
wget -q https://www.apache.org/dist/zookeeper/KEYS && \
wget -q https://www.apache.org/dist/zookeeper/zookeeper-${ZOOKEEPER_VERSION}/zookeeper-${ZOOKEEPER_VERSION}.tar.gz.asc && \
wget -q https://www.apache.org/dist/zookeeper/zookeeper-${ZOOKEEPER_VERSION}/zookeeper-${ZOOKEEPER_VERSION}.tar.gz.md5
RUN md5sum -c zookeeper-${ZOOKEEPER_VERSION}.tar.gz.md5 && \
gpg --import KEYS && \
gpg --verify zookeeper-${ZOOKEEPER_VERSION}.tar.gz.asc
RUN tar -xzf zookeeper-${ZOOKEEPER_VERSION}.tar.gz -C /opt
RUN cp /opt/zookeeper-${ZOOKEEPER_VERSION}/conf/zoo_sample.cfg /opt/zookeeper-${ZOOKEEPER_VERSION}/conf/zoo.cfg
ENV JAVA_HOME $JAVA_HOME
ENV ZK_HOME /opt/zookeeper-${ZOOKEEPER_VERSION}
RUN sed  -i "s|/tmp/zookeeper|$ZK_HOME/data|g" $ZK_HOME/conf/zoo.cfg; mkdir $ZK_HOME/data
EXPOSE 2181 2888 3888
WORKDIR /opt/zookeeper-${ZOOKEEPER_VERSION}
VOLUME ["/opt/zookeeper-${ZOOKEEPER_VERSION}/conf", "/opt/zookeeper-${ZOOKEEPER_VERSION}/data"]
CMD /usr/sbin/sshd && sed -i -r 's|#(log4j.appender.ROLLINGFILE.MaxBackupIndex.*)|\1|g' $ZK_HOME/conf/log4j.properties && \
sed -i -r 's|#autopurge|autopurge|g' $ZK_HOME/conf/zoo.cfg && \
/opt/zookeeper-${ZOOKEEPER_VERSION}/bin/zkServer.sh start-foreground 
```

4. storm ->[参考项目-> https://github.com/wurstmeister/storm-docker]
```
FROM centos
RUN wget -q -O - https://zcc2018.oss-cn-beijing.a ... -storm-1.2.1.tar.gz | tar -xzf - -C /opt
ENV STORM_HOME /opt/apache-storm-1.2.1
RUN groupadd storm; useradd --gid storm --home-dir /home/storm --create-home --shell /bin/bash storm; chown -R storm:storm $STORM_HOME; mkdir /var/log/storm ; chown -R storm:storm /var/log/storm
RUN ln -s $STORM_HOME/bin/storm /usr/bin/storm
ADD storm.yaml $STORM_HOME/conf/storm.yaml
ADD cluster.xml $STORM_HOME/logback/cluster.xml
ADD config-supervisord.sh /usr/bin/config-supervisord.sh
ADD start-supervisor.sh /usr/bin/start-supervisor.sh 
RUN chmod a+x /usr/bin/config-supervisord.sh && \
chmod a+x /usr/bin/start-supervisor.sh 
RUN mkdir /etc/supervisor/conf.d/ -p
```

5. Docker-compose.yml【仅供参考】
```
version: '3'
services:
    #zookeeper
    zookeeper:
        image: zookeeper
        ports:
          - "2181:2181"
          - "10122:22"
        networks:
        - analyser_cluster
        #kafka
    kafka: 
        image: kafka
        ports:
            - "9092:9092"
        environment:
            KAFKA_ADVERTISED_HOST_NAME: "kafka"
            KAFKA_ADVERTISED_PORT: "9092"
            KAFKA_ZOOKEEPER_CONNECT: "zk:2181"
        volumes:
            - /var/run/docker.sock:/var/run/docker.sock
        links:
          - zookeeper:zk
        depends_on:
            - zookeeper
        networks:
        - analyser_cluster
    #storm ->nimbus
    nimbus:
        image: storm_nimbus
        ports:
            - "13773:3773"
            - "13772:3772"
            - "16627:6627"
            - "10322:22"
        links: 
            - zookeeper:zk
        depends_on:
            - zookeeper
        networks:
        - analyser_cluster
    #storm ->supervisor
    supervisor:
        image: storm_supervisor
        ports:
            - "8000:8000"
            - "10422:22"
        links:
            - nimbus:nimbus
            - zookeeper:zk
        depends_on:
            - zookeeper
        networks:
        - analyser_cluster
    #storm -> ui
    ui:
        image: storm_ui
        ports:
            - "18080:8080"
            - "10522:22"
        links:
          - nimbus:nimbus
          - zookeeper:zk 
        depends_on:
          - zookeeper
        networks:
        - analyser_cluster
networks:
  analyser_cluster:

```

> 主要是针对开源项目做了针对自己项目需求的定制,以及将一些软件源替换为国内源。还有针对docker
过程中一些无法运行的地方进行了修改,其实本项目还有elk集群,redis,mysql等与上述kafka、storm的链接,但是因为时间有限、所以没有发出来,等后续有时间再补上吧。

以上是关于Kafka+Zookeeper+Storm的docker化的主要内容,如果未能解决你的问题,请参考以下文章

无法在kafka-storm中将偏移数据写入zookeeper

zookeeper链接数导致kafka storm不能正常工作

Storm/Kafka - 无法获得 kafka 的偏移滞后

查看storm消费了多少kafka的数据

kafka和strom集群的环境安装

玩日志系统必备的17条操作命令|启动zookeeper kafka logstash storm 查看日志 我手上最珍贵的命令集