Elastic stack--ES/kibana/filebeat/metricbeat

Posted

tags:

篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了Elastic stack--ES/kibana/filebeat/metricbeat相关的知识,希望对你有一定的参考价值。

背景:

这篇文章主要是本人结合公司有关监控,详细介绍一下Elastic stack的几个组件:Elasticsearch/kibana/filebeat/metricbeat,通过安装配置部署,以及具体的case来介绍这四个组件,有关一些基本概念,直接从官方网站上copy,毕竟还是一首资料权威,本文英文水平有限,就不误人子弟了,

个人认为最核心的应该是ES;


过程:

用filebeat/metricbeat抓取数据,存在elasticsearch(后面简称ES)中,然后通过kibana展示并可视化,也就是做成漂亮的图形;


Note:

本文没有涉及到logstash,因为公司没有使用到它,所以先不讲,后面有时间我们在来研究;


ES简介:

Elasticsearch is a highly scalable open-source full-text search and analytics engine. It allows you to store, search, and analyze big volumes of data quickly and in near real time. It is generally used as the underlying engine/technology that powers applications that have complex search features and requirements.


Elasticsearch--Index--Type--Document

cluster:默认的cluster name - elasticsearch;

Node:一般是一个机器作为一个节点;

Index: 是具有一些相似特性的文档的集合,类似于数据库的Database,索引名必须小写(lowercase),单个集群中可以定义任意多个index;

Type: 属于index的一个逻辑种类或者分区,在6.0.0版本中不再使用;

Document: 可以被索引的最基本的一个单元;

Shards :当index数据量比较大的时候,select比较慢,你可以把index可以被分成多个shard,这样你可以分布和平行访问shard,increasing performance/throughput.

Replica shard (replicas):failover in case a shard/node fail.replica shard 不会再同一个node中;


创建index的时候可以指定shards和replicas数量,创建index之后,replicas的数量可以修改,但是shards不能修改。

By default, each index in Elasticsearch is allocated 5 primary shards and 1 replica which means that if you have at least two nodes in your cluster, your index will have 5 primary shards and another 5 replica shards (1 complete replica) for a total of 10 shards per index.


eg:0-4表示index:customer被分成5个shard,p代表primary shard, r代表replica shard,

GET _cat/shards/customer*?v

index    shard prirep state   docs store ip            node

customer 4     p      STARTED    0  162b 172.16.101.55 sht-sgmhadoopnn-01

customer 4     r      STARTED    0  162b 172.16.101.56 sht-sgmhadoopnn-02

customer 1     r      STARTED    0  162b 172.16.101.56 sht-sgmhadoopnn-02

customer 1     p      STARTED    0  162b 172.16.101.54 sht-sgmhadoopcm-01

customer 3     r      STARTED    1 3.4kb 172.16.101.55 sht-sgmhadoopnn-01

customer 3     p      STARTED    1 3.4kb 172.16.101.54 sht-sgmhadoopcm-01

customer 2     p      STARTED    0  162b 172.16.101.56 sht-sgmhadoopnn-02

customer 2     r      STARTED    0  162b 172.16.101.54 sht-sgmhadoopcm-01

customer 0     r      STARTED    0  162b 172.16.101.55 sht-sgmhadoopnn-01

customer 0     p      STARTED    0  162b 172.16.101.54 sht-sgmhadoopcm-01


1.installation ES  (three node cluster) and kiabana

Hostname/IP

Role

sht-sgmhadoopcm-01/ 172.16.101.54

ES/Kibana/Filebeat/Metricbeat/Cassandra/mysql

sht-sgmhadoopnn-01/ 172.16.101.55

ES

sht-sgmhadoopnn-02/ 172.16.101.56

ES


(1)Installation ES

Elasticsearch requires at least Java 8:

[[email protected] local]# java -version

java version "1.8.0_111"

Java(TM) SE Runtime Environment (build 1.8.0_111-b14)

Java HotSpot(TM) 64-Bit Server VM (build 25.111-b14, mixed mode)

[[email protected] ~]# useradd elasticsearch
[[email protected] ~]# wget https://artifacts.elastic.co/downloads/elasticsearch/elasticsearch-5.5.2.tar.gz
[[email protected] ~]# tar xf  elasticsearch-5.5.2.tar.gz
[[email protected] ~]# mkdir /usr/local/elasticsearch/{data,logs}
[[email protected] ~]# vim /usr/local/elasticsearch/config/elasticsearch.yml
cluster.name: myelasticsearch
node.name: ${HOSTNAME}
path.data: /usr/local/elasticsearch/data
path.logs: /usr/local/elasticsearch/logs
network.host: 172.16.101.54
#对外服务的端口
http.port: 9200
#集群节点之间交互的端口号
#transport.tcp.port:9300 
http.cors.enabled: true
http.cors.allow-origin: "*"
node.master: true
node.data: true
node.ingest: true
#设置默认索引副本数量:
#index.number_of_replicas:3 
#index.number_of_shards:5
discovery.zen.ping.unicast.hosts: ["172.16.101.54:9300","172.16.101.55:9300","172.16.101.56:9300"]
discovery.zen.minimum_master_nodes: 2  #防止脑裂(master_eligible_nodes / 2) + 1
[[email protected] ~]# chown -R elasticsearch.elasticsearch elasticsearch
[[email protected] ~]# vi /etc/security/limits.conf
*               soft    nproc           65536
*               hard    nproc           65536
*               soft    nofile          65536
*               hard    nofile          65536
[[email protected] ~]# vi /etc/sysctl.conf
vm.max_map_count= 262144
[[email protected] ~]# sysctl -p



注意:

同样需要在其他两个节点上172.16.101.55,172.16.101.56上安装配置,你也可以直接rsync到其他两个节点,然后修改对应的配置文件,然后启动elasticsearch,第一个启动的作为master,

其他两个会自动发现集群名为myelasticsearch的master,然后加入这个集群,作为data节点。节点发现是通过zen-disacovery技术。


[[email protected] local]# rsync -avz --progress /usr/local/elasticsearch sht-sgmhadoopnn-01:/usr/local/

[[email protected] local]# rsync -avz --progress /usr/local/elasticsearch sht-sgmhadoopnn-02:/usr/local/


依次启动:

[[email protected] ~]# su - elasticsearch -c "/usr/local/elasticsearch/bin/elasticsearch &"

[[email protected]  ~]# su - elasticsearch -c "/usr/local/elasticsearch/bin/elasticsearch &"

[[email protected]  ~]# su - elasticsearch -c "/usr/local/elasticsearch/bin/elasticsearch &"


[[email protected] elasticsearch]# ss -nltup|egrep "9200|9300"

tcp    LISTEN     0      128    ::ffff:172.16.101.54:9200                 :::*                   users:(("java",pid=2601,fd=158))

tcp    LISTEN     0      128    ::ffff:172.16.101.54:9300                 :::*                   users:(("java",pid=2601,fd=118))


[[email protected] elasticsearch]# curlhttp://172.16.101.54:9200/

{

  "name" : "sht-sgmhadoopcm-01",

  "cluster_name" : "myelasticsearch",

  "cluster_uuid" : "GOyhthoIQmebXKddPpZ4eQ",

  "version" : {

    "number" : "5.5.2",

    "build_hash" : "b2f0c09",

    "build_date" : "2017-08-14T12:33:14.154Z",

    "build_snapshot" : false,

    "lucene_version" : "6.6.0"

  },

  "tagline" : "You Know, for Search"

}


(2) Installation kibana

[[email protected] local]# tar xf kibana-5.5.2-linux-x86_64.tar.gz
[[email protected] local]# mv kibana-5.5.2-linux-x86_64 kibana
[[email protected] kibana]# vim config/kibana.yml
server.port: 5601
server.host: "172.16.101.54"
elasticsearch.url: "http://172.16.101.54:9200"
kibana.index: ".kibana"
[[email protected] kibana]# bin/kibana &
[[email protected] kibana]# ss -nltup|grep 5601
tcp    LISTEN     0      128    172.16.101.54:5601                  *:*                   users:(("node",pid=6503,fd=14))



Access kibana through browser web: http://172.16.101.54:5601/

Checking Kibana Status:http://172.16.101.54:5601/status

技术分享图片


(3)Loading Sample Data

? A set of fictitious accounts with randomly generated data. Download this data set by clicking here: accounts.zip

[[email protected] tmp]# unzip accounts.zip

[[email protected] tmp]# curl -H 'Content-Type: application/x-ndjson' -XPOST '172.16.101.54:9200/_bulk?pretty' --data-binary @logs.jsonl


Create index pattern: bank*

技术分享图片


技术分享图片


(4)Create a visualization

Now you can see what proportion of the 1000 accounts fall into each balance range.

技术分享图片


Now you can see the break down of the account holders' ages displayed in a ring around the balance ranges.

技术分享图片



2 Exploring Cluster 

有两种方式访问elasticsearch,终端上bulk API curl或者kinaba's console


(1)Bulk API


list health

[[email protected] local]# curl -X GET "172.16.101.54:9200/_cat/health?v"

epoch      timestamp cluster         status node.total node.data shards pri relo init unassign pending_tasks max_task_wait_time active_shards_percent

1527909806 11:23:26  myelasticsearch green           3         3     12   6    0    0        0             0                  -                100.0%


list all nodes

[[email protected] local]# curl -X GET "172.16.101.54:9200/_cat/nodes?v"

ip            heap.percent ram.percent cpu load_1m load_5m load_15m node.role master name

172.16.101.56            5          78   0    0.00    0.01     0.05 mdi       -      sht-sgmhadoopnn-02

172.16.101.54           10          88   1    0.00    0.05     0.05 mdi       *      sht-sgmhadoopcm-01

172.16.101.55            9          48   0    0.00    0.01     0.05 mdi       -      sht-sgmhadoopnn-01


list all indices

[[email protected] config]#curl -X GET "172.16.101.54:9200/_cat/indices?v"


create an index

[[email protected] config]#curl -X PUT "172.16.101.54:9200/customer?pretty"


(2) Kibana Dev Tools:

GET /_cat/health?v

GET /_cat/nodes?v

GET /_cat/indices?v

GET _cat/shards?v

GET _tasks


create index: 

PUT /customer?pretty


list all index:

GET /_cat/indices?v

health status index    uuid                   pri rep docs.count docs.deleted store.size pri.store.size

green  open   .kibana  RMboQeUhQ2-7sL9Ri1YvOg   1   1          1            0      6.4kb          3.2kb

green  open   customer ad1prAx1RM-NpYpt_8AJrA   5   1          1            0      8.1kb            4kb


Put something into customer index

<REST Verb> /<Index>/<Type>/<ID>

PUT /customer/external/1?pretty

{

  "name": "John Doe"

}

GET /customer/external/1?pretty

{

  "_index": "customer",

  "_type": "external",

  "_id": "1",

  "_version": 1,

  "found": true,

  "_source": {

    "name": "John Doe"

  }

}


delete an index:

DELETE /customer?pretty


Modify data

PUT /customer/external/1?pretty

{

  "name": "Jane Doe"

}


index a document without an explicit ID

POST /customer/external?pretty

{

  "name": "Jane Doe"

}


update document ID=1

POST /customer/external/1/_update?pretty

{

  "doc": { "name": "dehu", "age": 20 }

}


delete document id=AWO_DBrV5YtRPiTwnyJh

DELETE /customer/external/AWO_DBrV5YtRPiTwnyJh?pretty


batch processing

POST /customer/external/_bulk?pretty

{"index":{"_id":"1"}}

{"name": "John Doe" }

{"index":{"_id":"2"}}

{"name": "Jane Doe" }

POST /customer/external/_bulk?pretty

{"update":{"_id":"1"}}

{"doc": { "name": "Jane Doe" } }

{"delete":{"_id":"2"}}


3 Exploring data

#加载accounts.json文件内容到bank中

download data:https://raw.githubusercontent.com/elastic/elasticsearch/master/docs/src/test/resources/accounts.json#

and load it into our cluster as follows:

[[email protected] tmp]# curl -H "Content-Type: application/json" -XPOST '172.16.101.54:9200/bank/account/_bulk?pretty&refresh' --data-binary "@accounts.json"

GET /_cat/indices/bank?v

health status index uuid                   pri rep docs.count docs.deleted store.size pri.store.size

green  open   bank  czWxsadvTJmUA1VDZxUa7w   5   1       1000            0      1.3mb        680.2kb


The search API

GET /bank/_search

GET /bank/_search?q=*&sort=account_number:asc&pretty


ES的DSL语法

match_all表示整个document,

from表示从哪个开始,

size表示返回document个数,

sort表示排序

_source指定返回的key-values

GET /customer/_search

{

  "query": { "match_all": {} },

  "size": 10000

}


#在kibana的开发工具中执行,获得bank这个index的倒数前5个docuemnt

GET /bank/_search

{

  "query": { "match_all": {} },

  "from":1,

  "size": 5,

  "sort": { "account_number": { "order": "desc" } }

}


#返回指定的account_number和balance与其对应的value

GET /bank/_search

{

  "query": { "match_all": {} },

  "_source": ["account_number", "balance"]

}


#返回所有"account_number"=20

GET /bank/_search

{

  "query": { "match": { "account_number": 20 } }

}


#return contain “mill” or “lane”

GET /bank/_search

{

  "query": { "match": { "address": "mill" } }

}


#return contain “mill lane”

GET /bank/_search

{

  "query": { "match_phrase": { "address": "mill lane" } }

}

 

4 Case1: Display Cassandra system.log with error and warning log on Dashboard  through Filebeat fetch data

(1)Installing filebeat

安装filebeat之前,你需要安装和配置一下相关环境;

Elasticsearch

Kibana for the UI

Logstash(可选择的)


[[email protected] local]# wgethttps://artifacts.elastic.co/downloads/beats/filebeat/filebeat-5.5.2-linux-x86_64.tar.gz
[[email protected] local]# tar xf filebeat-5.5.2-linux-x86_64.tar.gz
[[email protected] local]# mv filebeat-5.5.2-linux-x86_64   filebeat



(2)configuring filebate

[[email protected] filebeat]# vim filebeat.yml
#define a single prospector with a single path.
filebeat.prospectors:
- type: log
  paths:
    - /var/log/*.log
  multiline.pattern: '^[[:space:]]'
  multiline.negate: false
  multiline.match: after
  tags: ["system"]
- input_type: log
  paths:
    - /usr/local/cassandra/logs/*.log
  include_lines: ["^ERR", "^WARN"]
  multiline.pattern: '^[[:space:]]'
  multiline.negate: false
  multiline.match: after
  tags: ["cassandra"]
#sending ouput to ES
output.elasticsearch:
  hosts: ["172.16.101.54:9200"]
#logging
logging.level: warning




test configruation file

[[email protected] filebeat]# ./filebeat -configtest -e

Config OK


(3)Loading the index template in ES

当filebeat启动,连接ES之后,默认自动加载index template(filebeat.template.json),如果已经加载index template,默认不会覆盖;

By default, Filebeat automatically loads the recommended template file(filebeat.template.json)

By default, if a template already exists in the index, it is not overwritten. To overwrite an existing template, set template.overwrite: true in the configuration file.

output.elasticsearch:

  hosts: ["172.16.101.54:9200"]

  template.name: "filebeat"

  template.path: "filebeat.template.json"

  template.overwrite: false


(4)Starting Filebeat

[[email protected] filebeat]# ./filebeat -c filebeat.yml &
[[email protected] filebeat]# curl -X GET "172.16.101.54:9200/_cat/indices?v"
health status index               uuid                   pri rep docs.count docs.deleted store.size pri.store.size
green  open   filebeat-2018.06.03 Z6NazqtBTT2CHJUM9vmBiA   5   1       3055            0      2.7mb          1.3mb
green  open   .kibana             41OJk7deTPivsetzt5ZDnw   1   1          2            0       40kb           20kb



(5)Loading the kibana index pattern

[[email protected] filebeat]# ./scripts/import_dashboards -only-index -eshttp://172.16.101.54:9200

然后在创建kibana/management/index pattern/filebeat-*

After you’ve created the index pattern, you can select the filebeat-* index pattern in Kibana to explore Filebeat data.

展示system.log日志中所有以ERR和WAR开头的日志:

技术分享图片

(6)也可以通过Dashboard过滤展示ERROR的日志,投影到显示屏,便于及时通知处理:

技术分享图片技术分享图片


(7)Display only ERROR log in the Dashboard.

技术分享图片





5.Case2: Display system and MySQL module through Metricbeat fetch data


Metricbeat is a lightweight shipper that you can install on your servers to periodically collect metrics from the operating system and from services running on the server

To get started with your own Metricbeat setup, install and configure these related products:

Elasticsearch for storage and indexing the data.

Kibana for the UI.


(1)Installing Metricbeat

[[email protected] local]# tar xf metricbeat-5.5.2-linux-x86_64.tar.gz

[[email protected] local]# mv metricbeat-5.5.2-linux-x86_64 metricbeat


(2)Configuring Metricbate

[[email protected] metricbeat]# vim metricbeat.yml
metricbeat.modules:
- module: system
  metricsets:
    - cpu
    - filesystem
    - memory
    - network
    - process
  enabled: true
  period: 10s
  processes: ['cassandra','mysql','postgres','java','redis','kafka','beat','mango','logstash','elastic','grafana']
- module: mysql
  metricsets: ["status"]
  enabled: true
  period: 30s
  hosts: ["root:[email protected](172.16.101.54:3306)/"]
output.elasticsearch:
  hosts: ["192.168.1.42:9200"]



(3)Loading the index template in ES

By default, Metricbeat automatically loads the recommended template file, metricbeat.template.json, if Elasticsearch output is enabled. You can configure metricbeat to load a different template by adjusting the template.name and template.path options in metricbeat.yml file:

output.elasticsearch:

  hosts: ["192.168.1.42:9200"]

  template.name: "metricbeat"

  template.path: "metricbeat.template.json"

  template.overwrite: false


(4)Starting Metricbeat

[[email protected] metricbeat]# ./metricbeat -c metricbeat.yml &


(5)Loading Sample Kibana Dashboards

[[email protected] metricbeat]# ./scripts/import_dashboards -eshttp://172.16.101.54:9200

Then create index pattern: metricbeat-*

After importing the dashboards, launch the Kibana web interface by pointing your browser to port 5601. 

技术分享图片

(6)Display System Module Data 

技术分享图片


(7)Display MySQL Module Data.

_exists_:mysql.status

技术分享图片技术分享图片


技术分享图片


6.Case3: Creating time series visualizations

(1)system.cpu.user.pct 

you’ll create will compare the real-time percentage of CPU time spent in user space to the results offset by one hour. In order to create this visualization, we’ll need to create two Timelion expressions. One with the real-time average of system.cpu.user.pct and another with the average offset by one hour.


Use the following expression to update your visualization:

.es(offset=-1h,index=metricbeat-*, timefield='@timestamp', metric='avg:system.cpu.user.pct').label('last hour').lines(fill=1,width=0.5).color(gray), .es(index=metricbeat-*, timefield='@timestamp', metric='avg:system.cpu.user.pct').label('current hour').title('CPU usage over time').color(#1E90FF).legend(columns=2, position=nw)



技术分享图片

(2)system.network.in.bytes and system.network.out.bytes


create a new Timelion visualization for inbound and outbound network traffic.


Use the following expression to update your visualization:

.es(index=metricbeat*, [email protected], metric=max:system.network.in.bytes).derivative().divide(1048576).lines(fill=2, width=1).color(green).label("Inbound traffic").title("Network traffic (MB/s)"), .es(index=metricbeat*, [email protected], metric=max:system.network.out.bytes).derivative().multiply(-1).divide(1048576).lines(fill=2, width=1).color(blue).label("Outbound traffic").legend(columns=2, position=nw)


技术分享图片


(3)system.memory.actual.used.bytes

monitors memory consumption

Use the following expression to update your visualization:

.es(index=metricbeat-*, timefield='@timestamp', metric='max:system.memory.actual.used.bytes').label('max memory').title('Memory consumption over time'), .es(index=metricbeat-*, timefield='@timestamp', metric='max:system.memory.actual.used.bytes').if(gt,6742000000,.es(index=metricbeat-*, timefield='@timestamp', metric='max:system.memory.actual.used.bytes'),null).label('warning').color('#FFCC11').lines(width=5), .es(index=metricbeat-*, timefield='@timestamp', metric='max:system.memory.actual.used.bytes').if(gt,6744000000,.es(index=metricbeat-*, timefield='@timestamp', metric='max:system.memory.actual.used.bytes'),null).label('severe').color('red').lines(width=5), .es(index=metricbeat-*, timefield='@timestamp', metric='max:system.memory.actual.used.bytes').mvavg(10).label('mvavg').lines(width=2).color(#5E5E5E).legend(columns=4, position=nw)


技术分享图片


今天就先介绍到这里,其实这些内容都是来自Elastic stack的官方:https://www.elastic.co/guide/index.html

本文只能算是入门,以后还要多多学习交流。




以上是关于Elastic stack--ES/kibana/filebeat/metricbeat的主要内容,如果未能解决你的问题,请参考以下文章

分布式任务elastic-job

hids Elastic Security 系列1-Elastic Security介绍

hids Elastic Security 系列1-Elastic Security介绍

Elastic:如何使用 Elastic 机器学习来侦测异常

Elastic:如何使用 Elastic 机器学习来侦测异常

elastic 线程池设置解决logstash-429