IT数码 购物 网址 头条 软件 日历 阅读 图书馆
TxT小说阅读器
↓语音阅读,小说下载,古典文学↓
图片批量下载器
↓批量下载图片,美女图库↓
图片自动播放器
↓图片自动播放器↓
一键清除垃圾
↓轻轻一点,清除系统垃圾↓
开发: C++知识库 Java知识库 JavaScript Python PHP知识库 人工智能 区块链 大数据 移动开发 嵌入式 开发工具 数据结构与算法 开发测试 游戏开发 网络协议 系统运维
教程: HTML教程 CSS教程 JavaScript教程 Go语言教程 JQuery教程 VUE教程 VUE3教程 Bootstrap教程 SQL数据库教程 C语言教程 C++教程 Java教程 Python教程 Python3教程 C#教程
数码: 电脑 笔记本 显卡 显示器 固态硬盘 硬盘 耳机 手机 iphone vivo oppo 小米 华为 单反 装机 图拉丁
 
   -> 大数据 -> elasticsearch 、logstash、filebeat、kibana日志执行流程 -> 正文阅读

[大数据]elasticsearch 、logstash、filebeat、kibana日志执行流程

elasticsearch

1.安装地址:wget?https://artifacts.elastic.co/downloads/elasticsearch/elasticsearch-7.5.1-linux-x86_64.tar.gz
2.创建普通账号 注意elasticsearch最后用非root账号运行
2.1创建过程
[root@master ~]# useradd yangyu 
[root@master ~]# passwd yangyu 
Changing password for user yangyu.
3.解压:
tar -zxvf elasticsearch-7.5.1-linux-x86_64.tar.gz
4.修改目录用户和用户组
[root@master ~]# chown -R yangyu:yangyu /home/elasticsearch

5.运行

?$cd /home/elasticsearch

启动。通过 -d 参数,表示后台运行。
$ bin/elasticsearch -d

6.使用普通账号启动elasticsearch可能出现的问题

[1]: max file descriptors [4096] for elasticsearch process is too low, increase to at least [65535]
[2]: max virtual memory areas vm.max_map_count [65530] is too low, increase to at least [262144]
[3]: the default discovery settings are unsuitable for production use; at least one of [discovery.seed_hosts, discovery.seed_providers, cluster.initial_master_nodes] must be configured
1解决方案之一
解决办法:
#切换到root用户修改
vim /etc/security/limits.conf
# 在最后面追加下面内容
*** hard nofile 65536 
*** soft nofile 65536
***  是启动ES的用户
2解决方案:
在/etc/sysctl.conf文件最后添加一行
vm.max_map_count=262144
执行/sbin/sysctl -p 立即生效
3.解决方案
修改elasticsearch.yml? 找到discovery那一块,做如下修改
?cluster.initial_master_nodes: ["node-1","node-2"]修改为:cluster.initial_master_nodes: ["node-1"]

7.

  1. 通过 su 普通用户启动

$ bin/elasticsearch -d

--------------------------------------------------------------------------------------------------------------------------------Logstash

Logstash

  1. 下载 wget https://artifacts.elastic.co/downloads/logstash/logstash-7.5.1.zip
  2. 解压

unzip logstash-7.5.1.zip

cd logstash-7.5.1

  1. 在root账户下启动

使用默认配置文件

执行?nohup bin/logstash -f config/logstash-sample.conf &?命令,后台启动 Logstash 服务。

?

?--------------------------------------------------------------------------------------------------------------------------------

?Beats

Beats 是一个全品类采集器的系列,包含多个:

Filebeat?:轻量型日志采集器。

Metricbeat?:轻量型指标采集器。

Packetbeat?:轻量型网络数据采集器。

Winlogbeat?:轻量型 Windows 事件日志采集器。

Auditbeat?:轻量型审计日志采集器。

Heartbeat?:面向运行状态监测的轻量型采集器。

Functionbeat?:面向云端数据的无服务器采集器。

演示日志搜集

  1. 下载
    curl -L -O https://artifacts.elastic.co/downloads/beats/filebeat/filebeat-7.5.1-linux-x86_64.tar.gz
  2. 解压

解压
$ tar -zxvf filebeat-7.5.1-linux-x86_64.tar.gz
$ cd filebeat-7.5.1-linux-x86_64

  1. 配置默认采集日志地址

Config filebeat.yml?定义inputs和输出地址output 选择的是5044

#=========================== Filebeat inputs
filebeat.inputs:
- type: log
??# Change to true to enable this input configuration.
??enabled: true
??# Paths that should be crawled and fetched. Glob based paths.
??paths:
????# - /var/log/*.log
????#- c:\programdata\elasticsearch\logs\*
????- /home/logs/spring.log # 配置我们要读取的 Spring Boot 应用的日志

#-------------------------- Elasticsearch output ----------------------------
#output.elasticsearch:
??# Array of hosts to connect to.
??# hosts: ["localhost:9200"]
#----------------------------- Logstash output --------------------------------
output.logstash:
??# The Logstash hosts
??hosts: ["localhost:5044"]

测试案例 spring.log

2021-08-04 11:08:21.388  INFO 11220 --- [           main] org.apache.catalina.core.StandardEngine  : Starting Servlet engine: [Apache Tomcat/9.0.46]
2021-08-04 11:08:21.663  INFO 11220 --- [           main] o.a.c.c.C.[Tomcat].[localhost].[/]       : Initializing Spring embedded WebApplicationContext
2021-08-04 11:08:21.663  INFO 11220 --- [           main] w.s.c.ServletWebServerApplicationContext : Root WebApplicationContext: initialization completed in 1656 ms
2021-08-04 11:08:22.034  INFO 11220 --- [           main] org.apache.spark.SparkContext            : Running Spark version 2.2.0
2021-08-04 11:08:22.353  INFO 11220 --- [           main] org.apache.spark.SparkContext            : Submitted application: sparkTest
2021-08-04 11:08:22.370  INFO 11220 --- [           main] org.apache.spark.SecurityManager         : Changing view acls to: GR-35
2021-08-04 11:08:22.370  INFO 11220 --- [           main] org.apache.spark.SecurityManager         : Changing modify acls to: GR-35
2021-08-04 11:08:22.371  INFO 11220 --- [           main] org.apache.spark.SecurityManager         : Changing view acls groups to: 
2021-08-04 11:08:22.371  INFO 11220 --- [           main] org.apache.spark.SecurityManager         : Changing modify acls groups to: 
2021-08-04 11:08:22.372  INFO 11220 --- [           main] org.apache.spark.SecurityManager         : SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(GR-35); groups with view permissions: Set(); users  with modify permissions: Set(GR-35); groups with modify permissions: Set()
2021-08-04 11:08:23.569  INFO 11220 --- [           main] org.apache.spark.util.Utils              : Successfully started service 'sparkDriver' on port 50792.
2021-08-04 11:08:23.584  INFO 11220 --- [           main] org.apache.spark.SparkEnv                : Registering MapOutputTracker
2021-08-04 11:08:23.599  INFO 11220 --- [           main] org.apache.spark.SparkEnv                : Registering BlockManagerMaster
2021-08-04 11:08:23.602  INFO 11220 --- [           main] o.a.s.s.BlockManagerMasterEndpoint       : Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
2021-08-04 11:08:23.602  INFO 11220 --- [           main] o.a.s.s.BlockManagerMasterEndpoint       : BlockManagerMasterEndpoint up
2021-08-04 11:08:23.608  INFO 11220 --- [           main] o.apache.spark.storage.DiskBlockManager  : Created local directory at C:\Users\GR-35\AppData\Local\Temp\blockmgr-f7ff5fd7-21ad-46b2-9a95-880d9c92c4dc
2021-08-04 11:08:23.634  INFO 11220 --- [           main] o.a.spark.storage.memory.MemoryStore     : MemoryStore started with capacity 891.0 MB
2021-08-04 11:08:23.671  INFO 11220 --- [           main] org.apache.spark.SparkEnv                : Registering OutputCommitCoordinator
2021-08-04 11:08:23.723  INFO 11220 --- [           main] org.spark_project.jetty.util.log         : Logging initialized @5971ms
2021-08-04 11:08:23.771  INFO 11220 --- [           main] org.spark_project.jetty.server.Server    : jetty-9.3.z-SNAPSHOT
2021-08-04 11:08:23.781  INFO 11220 --- [           main] org.spark_project.jetty.server.Server    : Started @6029ms
2021-08-04 11:08:23.798  INFO 11220 --- [           main] o.s.jetty.server.AbstractConnector       : Started ServerConnector@33634f04{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
2021-08-04 11:08:23.798  INFO 11220 --- [           main] org.apache.spark.util.Utils              : Successfully started service 'SparkUI' on port 4040.
2021-08-04 11:08:23.810  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@737d100a{/jobs,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.811  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@3c98781a{/jobs/json,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.811  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@4601203a{/jobs/job,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.812  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@80bfa9d{/jobs/job/json,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.812  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@4b039c6d{/stages,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.812  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@507d64aa{/stages/json,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.812  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@60b34931{/stages/stage,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.813  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@1e0a864d{/stages/stage/json,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.813  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@3e67f5f2{/stages/pool,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.814  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@4527f70a{/stages/pool/json,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.814  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@7132a9dc{/storage,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.814  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@2da66a44{/storage/json,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.815  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@61bfc9bf{/storage/rdd,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.816  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@329bad59{/storage/rdd/json,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.818  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@178f268a{/environment,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.819  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@5c723f2d{/environment/json,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.820  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@2d7a9786{/executors,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.820  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@366d8b97{/executors/json,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.820  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@7654f833{/executors/threadDump,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.821  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@d7109be{/executors/threadDump/json,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.827  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@f4a3a8d{/static,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.828  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@1f1ff879{/,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.829  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@24f870ee{/api,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.829  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@1f03fba0{/jobs/job/kill,null,AVAILABLE,@Spark}
2021-08-04 11:08:23.830  INFO 11220 --- [           main] o.s.jetty.server.handler.ContextHandler  : Started o.s.j.s.ServletContextHandler@3bbf6abe{/stages/stage/kill,null,AVAILABLE,@Spark}

--------------------------------------------------------------------------------------------------------------------------------?

?Kibana?

1.下载
curl -O https://artifacts.elastic.co/downloads/kibana/kibana-7.5.1-linux-x86_64.tar.gz
2.解压
tar -zxvf kibana-7.5.1-linux-x86_64.tar.gz
cd kibana-7.5.1-linux-x86_64
3.配置
server.port: 5601
server.host: "0.0.0.0"
elasticsearch.hosts: ["http://localhost:9200"]
kibana.index: ".kibana"
4使用跟elasticsearch 一样用户启动
改变文件属性
 chown -R yangyu:yangyu /home/Kibana
执行?nohup bin/kibana &?命令,后台启动 Kibana 服务。
可以通过?nohup.out?日志,查看启动是否成功。
5 .在使用kibana之前要启动 elasticsearch 、logstash、filebeat
且日志已经写入到配置地址 /home/logs/spring.log
filebeat会自动采集日志信息,通过logstash转换,存储到elasticsearch ,这样才能在kibana显示出来

?

第一步要搜索到filebeat-7.5.1-*,必须第五点执行完成

至此,elasticsearch 、logstash、filebeat安装和具体数据流程已经完毕,下章介绍通过sprinboot-LogBack组件来完成日志生成,使用filebeat进行日志采集

ELK 最简架构

  大数据 最新文章
实现Kafka至少消费一次
亚马逊云科技:还在苦于ETL?Zero ETL的时代
初探MapReduce
【SpringBoot框架篇】32.基于注解+redis实现
Elasticsearch:如何减少 Elasticsearch 集
Go redis操作
Redis面试题
专题五 Redis高并发场景
基于GBase8s和Calcite的多数据源查询
Redis——底层数据结构原理
上一篇文章      下一篇文章      查看所有文章
加:2021-08-05 17:25:10  更:2021-08-05 17:27:39 
 
开发: C++知识库 Java知识库 JavaScript Python PHP知识库 人工智能 区块链 大数据 移动开发 嵌入式 开发工具 数据结构与算法 开发测试 游戏开发 网络协议 系统运维
教程: HTML教程 CSS教程 JavaScript教程 Go语言教程 JQuery教程 VUE教程 VUE3教程 Bootstrap教程 SQL数据库教程 C语言教程 C++教程 Java教程 Python教程 Python3教程 C#教程
数码: 电脑 笔记本 显卡 显示器 固态硬盘 硬盘 耳机 手机 iphone vivo oppo 小米 华为 单反 装机 图拉丁

360图书馆 购物 三丰科技 阅读网 日历 万年历 2024年5日历 -2024/5/17 19:50:00-

图片自动播放器
↓图片自动播放器↓
TxT小说阅读器
↓语音阅读,小说下载,古典文学↓
一键清除垃圾
↓轻轻一点,清除系统垃圾↓
图片批量下载器
↓批量下载图片,美女图库↓
  网站联系: qq:121756557 email:121756557@qq.com  IT数码