如何理解容器部署ELK7.10

发布时间:2021-10-20 16:04:58 作者:iii
来源:亿速云 阅读:147

这篇文章主要介绍“如何理解容器部署ELK7.10”,在日常操作中,相信很多人在如何理解容器部署ELK7.10问题上存在疑惑,小编查阅了各式资料,整理出简单好用的操作方法,希望对大家解答”如何理解容器部署ELK7.10”的疑惑有所帮助!接下来,请跟着小编一起来学习吧!

 一、elk架构简介

如何理解容器部署ELK7.10

二、环境

如何理解容器部署ELK7.10

三、系统参数优化

# 最大用户打开进程数 $ vim /etc/security/limits.d/20-nproc.conf  *           soft   nproc       65535 *           hard   nproc       65535  # 优化内核,用于 docker 支持 $ modprobe br_netfilter $ cat <<EOF >  /etc/sysctl.d/k8s.conf net.bridge.bridge-nf-call-ip6tables = 1 net.bridge.bridge-nf-call-iptables = 1 net.ipv4.ip_forward = 1 EOF $ sysctl -p /etc/sysctl.d/k8s.conf  # 优化内核,对 es 支持 $ echo 'vm.max_map_count=262144' >> /etc/sysctl.conf  # 生效配置 $ sysctl -p

四、部署 docker 和 docker-compose

部署 docker

# 安装必要的一些系统工具 $ yum install -y yum-utils device-mapper-persistent-data lvm2  # 添加软件源信息 $ yum-config-manager --add-repo http://mirrors.aliyun.com/docker-ce/linux/centos/docker-ce.repo  # 更新并安装 Docker-CE $ yum makecache fast $ yum -y install docker-ce  # 配置docker $ systemctl enable docker $ systemctl start docker $ vim /etc/docker/daemon.json {"data-root": "/var/lib/docker", "bip": "10.50.0.1/16", "default-address-pools": [{"base": "10.51.0.1/16", "size": 24}], "registry-mirrors": ["https://4xr1qpsp.mirror.aliyuncs.com"], "log-opts": {"max-size":"500m", "max-file":"3"}} $ sed  -i '/ExecStart=/i ExecStartPost=\/sbin\/iptables -P FORWARD ACCEPT' /usr/lib/systemd/system/docker.service $ systemctl enable docker.service $ systemctl daemon-reload $ systemctl restart docker

部署 docker-compose

# 安装 docker-compose $ sudo curl -L "https://github.com/docker/compose/releases/download/1.27.4/docker-compose-$(uname -s)-$(uname -m)" -o /usr/local/bin/docker-compose $ chmod +x /usr/local/bin/docker-compose

五、部署 ES

es-master1 操作

# 创建 es 目录 $ mkdir /data/ELKStack $ mkdir elasticsearch elasticsearch-data elasticsearch-plugins  # 容器es用户 uid 和 gid 都是 1000 $ chown 1000.1000 elasticsearch-data elasticsearch-plugins  # 临时启动一个es $ docker run --name es-test -it --rm docker.elastic.co/elasticsearch/elasticsearch:7.10.1 bash  # 生成证书,证书有效期10年,证书输入的密码这里为空 $ bin/elasticsearch-certutil ca --days 3660 $ bin/elasticsearch-certutil cert --ca elastic-stack-ca.p12 --days 3660  # 打开新的窗口,拷贝生成的证书 $ cd /data/ELKStack/elasticsearch $ mkdir es-p12 $ docker cp es-test:/usr/share/elasticsearch/elastic-certificates.p12 ./es-p12 $ docker cp es-test:/usr/share/elasticsearch/elastic-stack-ca.p12 ./es-p12 $ chown -R 1000.1000 ./es-p12  # 创建 docker-compose.yml $ vim docker-compose.yml  version: '2.2' services:   elasticsearch:     image: docker.elastic.co/elasticsearch/elasticsearch:7.10.1     container_name: es01     environment:       - cluster.name=es-docker-cluster       - cluster.initial_master_nodes=es01,es02,es03       - bootstrap.memory_lock=true       - "ES_JAVA_OPTS=-Xms10000m -Xmx10000m"     ulimits:       memlock:         soft: -1         hard: -1       nofile:         soft: 65536         hard: 65536     mem_limit: 13000m     cap_add:       - IPC_LOCK     restart: always     # 设置 docker host 网络模式     network_mode: "host"     volumes:        - /data/ELKStack/elasticsearch-data:/usr/share/elasticsearch/data        - /data/ELKStack/elasticsearch-plugins:/usr/share/elasticsearch/plugins        - /data/ELKStack/elasticsearch/elasticsearch.yml:/usr/share/elasticsearch/config/elasticsearch.yml        - /data/ELKStack/elasticsearch/es-p12:/usr/share/elasticsearch/config/es-p12  # 创建 elasticsearch.yml 配置文件 $ vim elasticsearch.yml  cluster.name: "es-docker-cluster" node.name: "es01" network.host: 0.0.0.0  node.master: true node.data: true  discovery.zen.minimum_master_nodes: 2 http.port: 9200 transport.tcp.port: 9300  # 如果是多节点es,通过ping来健康检查 discovery.zen.ping.unicast.hosts: ["172.20.166.25:9300", "172.20.166.24:9300", "172.20.166.22:9300", "172.20.166.23:9300", "172.20.166.26:9300"] discovery.zen.fd.ping_timeout: 120s discovery.zen.fd.ping_retries: 6 discovery.zen.fd.ping_interval: 10s  cluster.info.update.interval: 1m indices.fielddata.cache.size:  20% indices.breaker.fielddata.limit: 40% indices.breaker.request.limit: 40% indices.breaker.total.limit: 70% indices.memory.index_buffer_size: 20% script.painless.regex.enabled: true  # 磁盘分片分配 cluster.routing.allocation.disk.watermark.low: 100gb cluster.routing.allocation.disk.watermark.high: 50gb cluster.routing.allocation.disk.watermark.flood_stage: 30gb  # 本地数据分片恢复配置 gateway.recover_after_nodes: 3 gateway.recover_after_time: 5m gateway.expected_nodes: 3 cluster.routing.allocation.node_initial_primaries_recoveries: 8 cluster.routing.allocation.node_concurrent_recoveries: 2  # 允许跨域请求 http.cors.enabled: true http.cors.allow-origin: "*" http.cors.allow-headers: Authorization,X-Requested-With,Content-Length,Content-Type  # 开启xpack xpack.security.enabled: true xpack.monitoring.collection.enabled: true  # 开启集群中https传输 xpack.security.transport.ssl.enabled: true xpack.security.transport.ssl.verification_mode: certificate xpack.security.transport.ssl.keystore.path: es-p12/elastic-certificates.p12 xpack.security.transport.ssl.truststore.path: es-p12/elastic-certificates.p12  # 把 es 配置使用 rsync 同步到其它 es 节点 $ rsync -avp -e ssh /data/ELKStack 172.20.166.24:/data/ $ rsync -avp -e ssh /data/ELKStack 172.20.166.22:/data/ $ rsync -avp -e ssh /data/ELKStack 172.20.166.23:/data/ $ rsync -avp -e ssh /data/ELKStack 172.20.166.26:/data/  # 启动 es $ docker-compose up -d  # 查看 es $ docker-compose ps

es-master2 操作

$ cd /data/ELKStack/elasticsearch  # 修改 docker-compose.yml elasticsearch.yml 两个配置 $ sed -i 's/es01/es02/g' docker-compose.yml elasticsearch.yml  # 启动 es $ docker-compose up -d

es-master3 操作

$ cd /data/ELKStack/elasticsearch  # 修改 docker-compose.yml elasticsearch.yml 两个配置 $ sed -i 's/es01/es03/g' docker-compose.yml elasticsearch.yml  # 启动 es $ docker-compose up -d

es-data1 操作

$ cd /data/ELKStack/elasticsearch  # 修改 docker-compose.yml elasticsearch.yml 两个配置 $ sed -i 's/es01/es04/g' docker-compose.yml elasticsearch.yml  # 不做为 es master 节点,只做数据节点 $ sed -i 's/node.master: true/node.master: false/g' elasticsearch.yml  # 启动 es $ docker-compose up -d

es-data2 操作

$ cd /data/ELKStack/elasticsearch  # 修改 docker-compose.yml elasticsearch.yml 两个配置 $ sed -i 's/es01/es05/g' docker-compose.yml elasticsearch.yml  # 不做为 es master 节点,只做数据节点 $ sed -i 's/node.master: true/node.master: false/g' elasticsearch.yml  # 启动 es $ docker-compose up -d

设置 es 访问账号

# es-master1 操作 $ docker exec -it es01 bash  # 设置 elastic,apm_system,kibana,kibana_system,logstash_system,beats_system,remote_monitoring_user 等密码 # 密码都设置为 elastic123,这里只是举例,具体根据需求设置 $ ./bin/elasticsearch-setup-passwords interactive

六、部署 Kibana

logstash4 操作

$ mkdir -p /data/ELKStack/kibana $ cd /data/ELKStack/kibana  # 创建 kibana 相关目录,用于容器挂载 $ mkdir config data plugins $ chown 1000.1000 config data plugins  # 创建 docker-compose.yml $ vim docker-compose.yml  version: '2' services:   kibana:     image: docker.elastic.co/kibana/kibana:7.10.1     container_name: kibana     restart: always     network_mode: "bridge"     mem_limit: 2000m     environment:       SERVER_NAME: kibana.example.com     ports:       - "5601:5601"     volumes:        - /data/ELKStack/kibana/config:/usr/share/kibana/config        - /data/ELKStack/kibana/data:/usr/share/kibana/data        - /data/ELKStack/kibana/plugins:/usr/share/kibana/plugins  # 创建 kibana.yml $ vim config/kibana.yml  server.name: kibana server.host: "0" elasticsearch.hosts: ["http://172.20.166.25:9200","http://172.20.166.24:9200","http://172.20.166.22:9200"] elasticsearch.username: "kibana" elasticsearch.password: "elastic123" monitoring.ui.container.elasticsearch.enabled: true xpack.security.enabled: true xpack.encryptedSavedObjects.encryptionKey: encryptedSavedObjects12345678909876543210 xpack.security.encryptionKey: encryptionKeysecurity12345678909876543210 xpack.reporting.encryptionKey: encryptionKeyreporting12345678909876543210 i18n.locale: "zh-CN"  # 启动 kibana $ docker-compose up -d

七、部署 Zookeeper

logstash2 操作

# 创建 zookeeper 目录 $ mkdir /data/ELKStack/zookeeper $ cd /data/ELKStack/zookeeper $ mkdir data datalog $ chown 1000.1000 data datalog  # 创建 docker-compose.yml $ vim docker-compose.yml  version: '2' services:   zoo1:     image: zookeeper:3.6.2     restart: always     hostname: zoo1     container_name: zoo1     network_mode: "bridge"     mem_limit: 2000m     ports:       - 2181:2181       - 3888:3888       - 2888:2888     volumes:       - /data/ELKStack/zookeeper/data:/data       - /data/ELKStack/zookeeper/datalog:/datalog       - /data/ELKStack/zookeeper/zoo.cfg:/conf/zoo.cfg     environment:       ZOO_MY_ID: 1  # 表示 ZK服务的 id, 它是1-255 之间的整数, 必须在集群中唯一       ZOO_SERVERS: server.1=0.0.0.0:2888:3888;2181 server.2=172.20.166.28:2888:3888;2181 server.3=172.20.166.29:2888:3888;2181       # ZOOKEEPER_CLIENT_PORT: 2181  # 创建 zoo.cfg 配置 $ vim zoo.cfg  tickTime=2000 initLimit=10 syncLimit=5 dataDir=/data dataLogDir=/datalog autopurge.snapRetainCount=3 autopurge.purgeInterval=1 maxClientCnxns=60 server.1= 0.0.0.0:2888:3888;2181 server.2= 172.20.166.28:2888:3888;2181 server.3= 172.20.166.29:2888:3888;2181  # 拷贝配置到 logstash3 logstash4 机器上 $ rsync -avp -e ssh /data/ELKStack/zookeeper 172.20.166.28:/data/ELKStack/ $ rsync -avp -e ssh /data/ELKStack/zookeeper 172.20.166.29:/data/ELKStack/  # 启动 zookeeper $ docker-compose up -d

logstash3 操作

$ cd /data/ELKStack/zookeeper  # 修改 docker-compose.yml 文件 $ vim docker-compose.yml  version: '2' services:   zoo2:     image: zookeeper:3.6.2     restart: always     hostname: zoo2     container_name: zoo2     network_mode: "bridge"     mem_limit: 2000m     ports:       - 2181:2181       - 3888:3888       - 2888:2888     volumes:       - /data/ELKStack/zookeeper/data:/data       - /data/ELKStack/zookeeper/datalog:/datalog       - /data/ELKStack/zookeeper/zoo.cfg:/conf/zoo.cfg     environment:       ZOO_MY_ID: 2  # 表示 ZK服务的 id, 它是1-255 之间的整数, 必须在集群中唯一       ZOO_SERVERS: server.1=172.20.166.27:2888:3888;2181 server.2=0.0.0.0:2888:3888;2181 server.3=172.20.166.29:2888:3888;2181       # ZOOKEEPER_CLIENT_PORT: 2181  # 修改 zoo.cfg $ vim zoo.cfg  tickTime=2000 initLimit=10 syncLimit=5 dataDir=/data dataLogDir=/datalog autopurge.snapRetainCount=3 autopurge.purgeInterval=1 maxClientCnxns=60 server.1= 172.20.166.27:2888:3888;2181 server.2= 0.0.0.0:2888:3888;2181 server.3= 172.20.166.29:2888:3888;2181  # 启动 zookeeper $ docker-compose up -d

logstash4 操作

$ cd /data/ELKStack/zookeeper  # 修改 docker-compose.yml 文件 $ vim docker-compose.yml  version: '2' services:   zoo3:     image: zookeeper:3.6.2     restart: always     hostname: zoo3     container_name: zoo3     network_mode: "bridge"     mem_limit: 2000m     ports:       - 2181:2181       - 3888:3888       - 2888:2888     volumes:       - /data/ELKStack/zookeeper/data:/data       - /data/ELKStack/zookeeper/datalog:/datalog       - /data/ELKStack/zookeeper/zoo.cfg:/conf/zoo.cfg     environment:       ZOO_MY_ID: 3  # 表示 ZK服务的 id, 它是1-255 之间的整数, 必须在集群中唯一       ZOO_SERVERS: server.1=172.20.166.27:2888:3888;2181 server.2=172.20.166.28:2888:3888;2181 server.3=0.0.0.0:2888:3888;2181       # ZOOKEEPER_CLIENT_PORT: 2181  # 修改 zoo.cfg $ vim zoo.cfg  tickTime=2000 initLimit=10 syncLimit=5 dataDir=/data dataLogDir=/datalog autopurge.snapRetainCount=3 autopurge.purgeInterval=1 maxClientCnxns=60 server.1= 172.20.166.27:2888:3888;2181 server.2= 172.20.166.28:2888:3888;2181 server.3= 0.0.0.0:2888:3888;2181  # 启动 zookeeper $ docker-compose up -d  # 操作 zookeeper $ docker run -it zoo3 bash $ zkCli.sh -server 172.20.166.27:2181,172.20.166.28:2181,172.20.166.29:2181

八、部署 Kafka

logstash2 操作

# 创建 kafka 目录 $ mkdir -p /data/ELKStack/kafka $ cd /data/ELKStack/kafka  # 创建数据目录,用于存储kafka容器数据 $ mkdir data  # 把kafka配置拷贝到宿主机上 $ docker run --name kafka-test -it --rm wurstmeister/kafka:2.13-2.6.0 bash $ cd /opt/kafka $ tar zcvf /tmp/config.tar.gz config  # 打开一个新的窗口 $ docker cp kafka-test:/tmp/config.tar.gz ./  # 解压配置文件 $ tar xf config.tar.gz  # 创建 docker-compose.yml $ vim docker-compose.yml  version: '2'  services:   kafka1:     image: wurstmeister/kafka:2.13-2.6.0     restart: always     hostname: kafka1     container_name: kafka1     network_mode: "bridge"     mem_limit: 5120m     ports:     - 9092:9092     - 9966:9966     environment:       KAFKA_BROKER_ID: 1       KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://172.20.166.27:9092       # 宿主机的IP地址而非容器的IP,及暴露出来的端口       KAFKA_ADVERTISED_HOST_NAME: 172.20.166.27                        # 外网访问地址       KAFKA_ADVERTISED_PORT: 9092                                      # 端口       KAFKA_ZOOKEEPER_CONNECT: 172.20.166.27:2181,172.20.166.28:2181,172.20.166.29:2181           # 连接的zookeeper服务及端口       KAFKA_JMX_OPTS: "-Dcom.sun.management.jmxremote -Dcom.sun.management.jmxremote.authenticate=false -Dcom.sun.management.jmxremote.ssl=false -Djava.rmi.server.hostname=172.20.166.27 -Dcom.sun.management.jmxremote.rmi.port=9966"       JMX_PORT: 9966 # kafka需要监控broker和topic的数据的时候,是需要开启jmx_port的       KAFKA_HEAP_OPTS: "-Xmx4096M -Xms4096M"     volumes:     - /data/ELKStack/kafka/data:/kafka                    # kafka数据文件存储目录     - /data/ELKStack/kafka/config:/opt/kafka/config  # 优化 kafka server.properties 配置 $ vim config/server.properties  # 调大socket,防止报错 socket.send.buffer.bytes=1024000 socket.receive.buffer.bytes=1024000 socket.request.max.bytes=1048576000  # topic 数据保留多久,默认168小时(7day) log.retention.hours=72 log.cleanup.policy=delete  # 拷贝配置到 logstash3 logstash4 机器上 $ rsync -avp -e ssh /data/ELKStack/kafka 172.20.166.28:/data/ELKStack/ $ rsync -avp -e ssh /data/ELKStack/kafka 172.20.166.29:/data/ELKStack/  # 启动 kafka $ docker-compose up -d

logstash3 操作

$ cd /data/ELKStack/kafka  # 修改 docker-compose.yml 文件 $ vim docker-compose.yml  version: '2'  services:   kafka2:     image: wurstmeister/kafka:2.13-2.6.0     restart: always     hostname: kafka2     container_name: kafka2     network_mode: "bridge"     mem_limit: 5120m     ports:     - 9092:9092     - 9966:9966     environment:       KAFKA_BROKER_ID: 2       KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://172.20.166.28:9092       # 宿主机的IP地址而非容器的IP,及暴露出来的端口       KAFKA_ADVERTISED_HOST_NAME: 172.20.166.28                        # 外网访问地址       KAFKA_ADVERTISED_PORT: 9092                                      # 端口       KAFKA_ZOOKEEPER_CONNECT: 172.20.166.27:2181,172.20.166.28:2181,172.20.166.29:2181           # 连接的zookeeper服务及端口       KAFKA_JMX_OPTS: "-Dcom.sun.management.jmxremote -Dcom.sun.management.jmxremote.authenticate=false -Dcom.sun.management.jmxremote.ssl=false -Djava.rmi.server.hostname=172.20.166.28 -Dcom.sun.management.jmxremote.rmi.port=9966"       JMX_PORT: 9966  # kafka需要监控broker和topic的数据的时候,是需要开启jmx_port的       KAFKA_HEAP_OPTS: "-Xmx4096M -Xms4096M"     volumes:     - /data/ELKStack/kafka/data:/kafka                    # kafka数据文件存储目录     - /data/ELKStack/kafka/config:/opt/kafka/config  # 启动 kafka $ docker-compose up -d

logstash4 操作

$ cd /data/ELKStack/kafka  # 修改 docker-compose.yml 文件 $ vim docker-compose.yml  version: '2'  services:   kafka3:     image: wurstmeister/kafka:2.13-2.6.0     restart: always     hostname: kafka3     container_name: kafka3     network_mode: "bridge"     mem_limit: 5120m     ports:     - 9092:9092     - 9966:9966     environment:       KAFKA_BROKER_ID: 3       KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://172.20.166.29:9092       # 宿主机的IP地址而非容器的IP,及暴露出来的端口       KAFKA_ADVERTISED_HOST_NAME: 172.20.166.29                        # 外网访问地址       KAFKA_ADVERTISED_PORT: 9092                                      # 端口       KAFKA_ZOOKEEPER_CONNECT: 172.20.166.27:2181,172.20.166.28:2181,172.20.166.29:2181           # 连接的zookeeper服务及端口       KAFKA_JMX_OPTS: "-Dcom.sun.management.jmxremote -Dcom.sun.management.jmxremote.authenticate=false -Dcom.sun.management.jmxremote.ssl=false -Djava.rmi.server.hostname=172.20.166.29 -Dcom.sun.management.jmxremote.rmi.port=9966"       JMX_PORT: 9966  # kafka需要监控broker和topic的数据的时候,是需要开启jmx_port的       KAFKA_HEAP_OPTS: "-Xmx4096M -Xms4096M"     volumes:     - /data/ELKStack/kafka/data:/kafka                    # kafka数据文件存储目录     - /data/ELKStack/kafka/config:/opt/kafka/config  # 启动 kafka $ docker-compose up -d  # 部署 kafka-manager 管理 kafka 平台 $ mkdir /data/ELKStack/kafka-manager $ cd /data/ELKStack/kafka-manager $ vim docker-compose.yml  version: '3.6' services:   kafka_manager:     restart: always     container_name: kafa-manager     hostname: kafka-manager     network_mode: "bridge"     mem_limit: 1024m     image: hlebalbau/kafka-manager:3.0.0.5-7e7a22e     ports:       - "9000:9000"     environment:       ZK_HOSTS: "172.20.166.27:2181,172.20.166.28:2181,172.20.166.29:2181"       APPLICATION_SECRET: "random-secret"       KAFKA_MANAGER_AUTH_ENABLED: "true"       KAFKA_MANAGER_USERNAME: admin       KAFKA_MANAGER_PASSWORD: elastic123       JMX_PORT: 9966       TZ: "Asia/Shanghai"  # 启动 kafka-manager $ docker-compose up -d  # 访问 http://172.20.166.29:9000 ,把上面创建的三台 kafka 加入管理,这里不在阐述,网上很多配置教程

九、部署 logstash

logstash2 操作

$ mkdir /data/ELKStack/logstash $ cd /data/ELKStack/logstash $ mkdir config data $ chown 1000.1000 config data  # 创建 docker-compose.yml $ vim docker-compose.yml  version: '2' services:   logstash2:     image: docker.elastic.co/logstash/logstash:7.10.1     container_name: logstash2     hostname: logstash2     restart: always     network_mode: "bridge"     mem_limit: 4096m     environment:       TZ: "Asia/Shanghai"     ports:       - 5044:5044     volumes:       - /data/ELKStack/logstash/config:/config-dir       - /data/ELKStack/logstash/logstash.yml:/usr/share/logstash/config/logstash.yml       - /data/ELKStack/logstash/data:/usr/share/logstash/data       - /etc/localtime:/etc/localtime     user: logstash     command: bash -c "logstash -f /config-dir --config.reload.automatic"  # 创建 logstash.yml $ vim logstash.yml  http.host: "0.0.0.0" # 指发送到Elasticsearch的批量请求的大小,值越大,处理则通常更高效,但增加了内存开销 pipeline.batch.size: 3000 # 指调整Logstash管道的延迟,过了该时间则logstash开始执行过滤器和输出 pipeline.batch.delay: 200  # 创建 logstash 规则配置 $ vim config/01-input.conf  input {                                        # 输入组件     kafka {                                    # 从kafka消费数据         bootstrap_servers => ["172.20.166.27:9092,172.20.166.28:9092,172.20.166.29:9092"]         #topics => "%{[@metadata][topic]}"     # 使用kafka传过来的topic         topics_pattern => "elk-.*"             # 使用正则匹配topic         codec => "json"                        # 数据格式         consumer_threads => 3                  # 消费线程数量         decorate_events => true                # 可向事件添加Kafka元数据,比如主题、消息大小的选项,这将向logstash事件中添加一个名为kafka的字段         auto_offset_reset => "latest"          # 自动重置偏移量到最新的偏移量         group_id => "logstash-node"            # 消费组ID,多个有相同group_id的logstash实例为一个消费组         client_id => "logstash2"               # 客户端ID         fetch_max_wait_ms => "1000"            # 指当没有足够的数据立即满足fetch_min_bytes时,服务器在回答fetch请求之前将阻塞的最长时间   } }  $ vim config/02-output.conf  output {                                       # 输出组件     elasticsearch {         # Logstash输出到es         hosts => ["172.20.166.25:9200", "172.20.166.24:9200", "172.20.166.22:9200", "172.20.166.23:9200", "172.20.166.26:9200"]         index => "%{[fields][source]}-%{+YYYY-MM-dd}"      # 直接在日志中匹配,索引会去掉elk         # index => "%{[@metadata][topic]}-%{+YYYY-MM-dd}"  # 以日期建索引         user => "elastic"         password => "elastic123"     }     #stdout {     #    codec => rubydebug     #} }  $ vim config/03-filter.conf  filter {    # 当非业务字段时,无traceId则移除    if ([message] =~ "traceId=null") {          # 过滤组件,这里只是展示,无实际意义,根据自己的业务需求进行过滤       drop {}    } }  # 拷贝配置到 logstash3 logstash4 机器上 $ rsync -avp -e ssh /data/ELKStack/logstash 172.20.166.28:/data/ELKStack/ $ rsync -avp -e ssh /data/ELKStack/logstash 172.20.166.29:/data/ELKStack/  # 启动 logstash $ docker-compose up -d

logstash3 操作

$ cd /data/ELKStack/logstash $ sed -i 's/logstash2/logstash3/g' docker-compose.yml $ sed -i 's/logstash2/logstash3/g' config/01-input.conf  # 启动 logstash $ docker-compose up -d

logstash4 操作

$ cd /data/ELKStack/logstash $ sed -i 's/logstash2/logstash4/g' docker-compose.yml $ sed -i 's/logstash2/logstash4/g' config/01-input.conf  # 启动 logstash $ docker-compose up -d

十、部署 filebeat

# 配置 filebeat yum源,这里以 centos7 为例 $ rpm --import https://artifacts.elastic.co/GPG-KEY-elasticsearch  $ vim /etc/yum.repos.d/elastic.repo  [elastic-7.x] name=Elastic repository for 7.x packages baseurl=https://artifacts.elastic.co/packages/7.x/yum gpgcheck=1 gpgkey=https://artifacts.elastic.co/GPG-KEY-elasticsearch enabled=1 autorefresh=1 type=rpm-md  $ yum install -y filebeat-7.10.1 $ systemctl enable filebeat  # 配置 $ cd /etc/filebeat/ $ cp -a filebeat.yml filebeat.yml.old $ echo > filebeat.yml  # 以收集nginx访问日志为例 $ vim filebeat.yml  filebeat.inputs:                   # inputs为复数,表名type可以有多个 - type: log                        # 输入类型   access:   enabled: true                    # 启用这个type配置   json.keys_under_root: true       # 默认这个值是FALSE的,也就是我们的json日志解析后会被放在json键上。设为TRUE,所有的keys就会被放到根节点   json.overwrite_keys: true        # 是否要覆盖原有的key,这是关键配置,将keys_under_root设为TRUE后,再将overwrite_keys也设为TRUE,就能把filebeat默认的key值给覆盖   max_bytes: 20480                 # 单条日志的大小限制,建议限制(默认为10M,queue.mem.events * max_bytes 将是占有内存的一部分)   paths:     - /var/log/nginx/access.log    # 监控nginx 的access日志    fields:                          # 额外的字段     source: nginx-access-prod      # 自定义source字段,用于es建议索引(字段名小写,我记得大写好像不行)  # 自定义es的索引需要把ilm设置为false setup.ilm.enabled: false  output.kafka:            # 输出到kafka   enabled: true          # 该output配置是否启用   hosts: ["172.20.166.27:9092", "172.20.166.28:9092", "172.20.166.29:9092"]  # kafka节点列表   topic: "elk-%{[fields.source]}"   # kafka会创建该topic,然后logstash(可以过滤修改)会传给es作为索引名称   partition.hash:     reachable_only: true # 是否只发往可达分区   compression: gzip      # 压缩   max_message_bytes: 1000000  # Event最大字节数。默认1000000。应小于等于kafka broker message.max.bytes值   required_acks: 1  # kafka ack等级   worker: 1  # kafka output的最大并发数   bulk_max_size: 2048    # 单次发往kafka的最大事件数 logging.to_files: true   # 输出所有日志到file,默认true, 达到日志文件大小限制时,日志文件会自动限制替换,详细配置:https://www.cnblogs.com/qinwengang/p/10982424.html close_older: 30m         # 如果一个文件在某个时间段内没有发生过更新,则关闭监控的文件handle。默认1h force_close_files: false # 这个选项关闭一个文件,当文件名称的变化。只在window建议为true  # 没有新日志采集后多长时间关闭文件句柄,默认5分钟,设置成1分钟,加快文件句柄关闭 close_inactive: 1m  # 传输了3h后荏没有传输完成的话就强行关闭文件句柄,这个配置项是解决以上案例问题的key point close_timeout: 3h  # 这个配置项也应该配置上,默认值是0表示不清理,不清理的意思是采集过的文件描述在registry文件里永不清理,在运行一段时间后,registry会变大,可能会带来问题 clean_inactive: 72h  # 设置了clean_inactive后就需要设置ignore_older,且要保证ignore_older < clean_inactive ignore_older: 70h  # 限制 CPU和内存资源 max_procs: 1 # 限制一个CPU核心,避免过多抢占业务资源 queue.mem.events: 256 # 存储于内存队列的事件数,排队发送 (默认4096) queue.mem.flush.min_events: 128 # 小于 queue.mem.events ,增加此值可提高吞吐量 (默认值2048)  # 启动 filebeat $ systemctl start filebeat

十一、部署 curator,定时清理es索引

logstash4 机器操作

# 参考链接:https://www.elastic.co/guide/en/elasticsearch/client/curator/current/yum-repository.html  # 安装 curator 服务,以 centos7 为例 $ rpm --import https://packages.elastic.co/GPG-KEY-elasticsearch  $ vim /etc/yum.repos.d/elk-curator-5.repo  [curator-5] name=CentOS/RHEL 7 repository for Elasticsearch Curator 5.x packages baseurl=https://packages.elastic.co/curator/5/centos/7 gpgcheck=1 gpgkey=https://packages.elastic.co/GPG-KEY-elasticsearch enabled=1  $ yum install elasticsearch-curator -y  # 创建 curator 配置文件目录与输出日志目录 $ mkdir -p /data/ELKStack/curator/logs $ cd /data/ELKStack/curator  $ vim config.yml  --- # Remember, leave a key empty if there is no value.  None will be a string, # # not a Python "NoneType" client:   hosts: ["172.20.166.25", "172.20.166.24", "172.20.166.22", "172.20.166.23", "172.20.166.26"]   port: 9200   url_prefix:   use_ssl: False   certificate:   client_cert:   client_key:   ssl_no_validate: False   http_auth: elastic:elastic123   timeout: 150   master_only: False  logging:   loglevel: INFO   logfile: /data/ELKStack/curator/logs/curator.log   logformat: default   blacklist: ['elasticsearch', 'urllib3']  $ vim action.yml  --- # Remember, leave a key empty if there is no value.  None will be a string, # not a Python "NoneType" # # Also remember that all examples have 'disable_action' set to True.  If you # want to use this action as a template, be sure to set this to False after # copying it. actions:   1:     action: delete_indices     description: >-       Delete indices older than 30 days. Ignore the error if the filter does not result in an actionable list of indices (ignore_empty_list) and exit cleanly.     options:       ignore_empty_list: True       disable_action: False     filters:     - filtertype: pattern       kind: regex       value: '^((?!(kibana|json|monitoring|metadata|apm|async|transform|siem|security)).)*$'     - filtertype: age       source: creation_date       direction: older       #timestring: '%Yi-%m-%d'       unit: days       unit_count: 30   2:     action: delete_indices     description: >-       Delete indices older than 15 days. Ignore the error if the filter does not result in an actionable list of indices (ignore_empty_list) and exit cleanly.     options:       ignore_empty_list: True       disable_action: False     filters:     - filtertype: pattern       kind: regex       value: '^(nginx-).*$'     - filtertype: age       source: creation_date       direction: older       #timestring: '%Yi-%m-%d'       unit: days       unit_count: 15  # 设置定时任务清理es索引 $ crontab -e  0 0 * * * /usr/bin/curator --config /data/ELKStack/curator/config.yml /data/ELKStack/curator/action.yml

到此,关于“如何理解容器部署ELK7.10”的学习就结束了,希望能够解决大家的疑惑。理论与实践的搭配能更好的帮助大家学习,快去试试吧!若想继续学习更多相关知识,请继续关注亿速云网站,小编会继续努力为大家带来更多实用的文章!

推荐阅读:
  1. 如何理解docker容器数据卷
  2. Angular容器如何部署

免责声明:本站发布的内容(图片、视频和文字)以原创、转载和分享为主,文章观点不代表本网站立场,如果涉及侵权请联系站长邮箱:is@yisu.com进行举报,并提供相关证据,一经查实,将立刻删除涉嫌侵权内容。

elk

上一篇:javascript如何实现次方计算

下一篇:Spring Cloud Alibaba下额外支持的RPC方案Dubbo是什么

相关阅读

您好,登录后才能下订单哦!

密码登录
登录注册
其他方式登录
点击 登录注册 即表示同意《亿速云用户服务条款》