diff --git a/Kibana.md b/Kibana.md
new file mode 100644
index 0000000..f27c5cd
--- /dev/null
+++ b/Kibana.md
@@ -0,0 +1,162 @@
+
Kibana
+
+------
+
+## 一:基本概念
+
+### 1. Kibana 简介
+
+ Kibana 是一个开源的数据分析和可视化平台,它是 Elastic Stack(包括 Elasticsearch、Logstash、Kibana 和 Beats)的一部分,主要用于对 Elasticsearch 中的数据进行搜索、查看、交互操作。
+
+### 2. Kibana 功能
+
+- 数据可视化:Kibana 提供了丰富的数据可视化选项,如柱状图、线图等,帮助用户以图形化的方式理解数据
+- 数据探索:Kibana 提供了强大的数据探索功能,用户可以通过 Kibana 的界面进行数据筛选和排序
+- 仪表盘:用户可以将多个可视化组件组合在一起,创建交互式的仪表盘,用于实时监控数据
+- 机器学习:Kibana 还集成了 Elasticsearch 的机器学习功能,可以用于异常检测、预测等任务
+- 定制和扩展:Kibana 提供了丰富的 API 和插件系统,用户可以根据自己的需求定制和扩展 Kibana
+
+## 二:安装部署
+
+### 1. 获取安装包
+
+ 对应Elasticsearch版本获取Kibana的安装包
+
+ 官网获取:[Past Releases of Elastic Stack Software | Elastic](https://www.elastic.co/downloads/past-releases#kibana)
+
+```bash
+[root@kibana ~]# wget https://artifacts.elastic.co/downloads/kibana/kibana-8.13.4-linux-x86_64.tar.gz
+```
+
+### 2. 安装部署
+
+```bash
+[root@kibana ~]# tar xf kibana-8.13.4-linux-x86_64.tar.gz -C /usr/local/
+[root@kibana ~]# mv /usr/local/kibana-8.13.4/ /usr/local/kibana
+[root@kibana ~]# mkdir /usr/local/kibana/config/certs
+```
+
+### 3. 修改配置
+
+```bash
+[root@kibana ~]# vim /usr/local/kibana/config/kibana.yml
+server.port: 5601
+server.host: "192.168.159.134"
+server.name: "kibana"
+elasticsearch.hosts: ["https://192.168.159.131:9200"]
+elasticsearch.username: "kibana"
+elasticsearch.password: "123456"
+elasticsearch.ssl.certificateAuthorities: [ "/usr/local/kibana/config/certs/elasticsearch-ca.pem" ]
+i18n.locale: "zh-CN"
+```
+
+### 4. 获取Elasticsearch 的CA证书
+
+```bash
+[root@kibana ~]# scp 192.168.159.131:/usr/local/elasticsearch/config/certs/elasticsearch-ca.pem /usr/local/kibana/config/certs/
+```
+
+### 5. 创建运行用户
+
+```bash
+[root@kibana ~]# useradd kibana
+[root@kibana ~]# echo kibana | passwd --stdin kibana
+[root@kibana ~]# chown -R kibana:kibana /usr/local/kibana/
+```
+
+### 6. 启动Kibana
+
+```bash
+[root@kibana ~]# su - kibana
+[kibana@kibana ~]$ /usr/local/kibana/bin/kibana
+```
+
+### 7. 常见故障
+
+```bash
+Error: [config validation of [elasticsearch].username]: value of "elastic" is forbidden. This is a superuser account that cannot write to system indices that Kibana needs to function. Use a service account token instead. Learn more: https://www.elastic.co/guide/en/elasticsearch/reference/8.0/service-accounts.html
+ at ensureValidConfiguration (/usr/local/kibana/node_modules/@kbn/core-config-server-internal/src/ensure_valid_configuration.js:23:11)
+```
+
+解决方案:
+
+ 该版本的kibana在连接es集群中,不支持使用elasticsearch中elastic(超级管理员)账户连接
+
+ 在es集群中自带的kibana账户设定密码
+
+```bash
+[elasticsearch@elasticsearch1 ~]$ /usr/local/elasticsearch/bin/elasticsearch-reset-password -u kibana -i
+warning: ignoring JAVA_HOME=/usr/local/elasticsearch/jdk; using bundled JDK
+This tool will reset the password of the [kibana] user.
+You will be prompted to enter the password.
+Please confirm that you would like to continue [y/N]y
+
+
+Enter password for [kibana]:
+Re-enter password for [kibana]:
+Password for the [kibana] user successfully reset.
+```
+
+扩展:
+
+Elasticsearch中创建用户
+
+```bash
+[elasticsearch@elasticsearch1 ~]$ /usr/local/elasticsearch/bin/elasticsearch-users useradd test_account
+```
+
+为用户分配角色
+
+```bash
+[elasticsearch@elasticsearch1 ~]$ /usr/local/elasticsearch/bin/elasticsearch-users roles -a superuser test_account
+```
+
+### 8. 浏览器访问
+
+
+
+注意:
+
+ 用户名和密码是es集群登录的用户和密码
+
+
+
+
+
+## 三:配置使用
+
+注意:此部分的使用在做完《Logstash》中第一节和第二节后在使用
+
+### 1. Logstash采集到数据进行展示
+
+进入管理界面
+
+
+
+
+
+进入索引管理
+
+
+
+进入要管理的索引
+
+
+
+展示索引数据
+
+
+
+创建数据视图
+
+
+
+
+
+
+
+注意:
+
+- 这里我们就可以看到Logstash采集的数据,并以图形的方式展示
+- 在图形界面我们可以设置时间范围
+- 在图形界面我们可以设置时间自动刷新时间
diff --git a/Logstash.md b/Logstash.md
new file mode 100644
index 0000000..202e720
--- /dev/null
+++ b/Logstash.md
@@ -0,0 +1,301 @@
+Logstash
+
+------
+
+## 一:基本概念
+
+### 1. Logstash 简介
+
+ Logstash 是 Elastic Stack 的中央数据流引擎,用于收集、丰富和统一所有数据,而不管格式或模式。
+
+ 相当与Elasticsearch,Kibana,及 Beats 共同使用的时候便会拥有特别强大的实时处理能力。
+
+### 2. Logstash 原理
+
+ Logstash 管道中的每个输入阶段都在其自己的线程中运行。输入将事件写入位于内存(默认)或磁盘上的中央队列。每个管道工作线程从此队列中取出一批事件,通过配置的过滤器运行这批事件,然后通过任何输出运行过滤后的事件。批次的大小和管道工作线程的数量是可配置的。
+
+ 默认情况下,Logstash 使用管道阶段(输入 → 过滤器和过滤器 → 输出)之间的内存有界队列来缓冲事件。如果 Logstash 不安全地终止,则存储在内存中的所有事件都将丢失。为了防止数据丢失,您可以启用 Logstash 将传输中的事件保存到磁盘。
+
+## 二:安装部署
+
+### 1. 下载安装包
+
+官网:[Past Releases of Elastic Stack Software | Elastic](https://www.elastic.co/downloads/past-releases#logstash)
+
+```bash
+[root@logstash ~]# wget https://artifacts.elastic.co/downloads/logstash/logstash-8.13.4-linux-x86_64.tar.gz
+```
+
+### 2. 解压安装
+
+```bash
+[root@logstash ~]# tar xf logstash-8.13.4-linux-x86_64.tar.gz -C /usr/local/
+[root@logstash ~]# mv /usr/local/logstash-8.13.4/ /usr/local/logstash
+```
+
+### 3. 配置环境
+
+```bash
+[root@logstash ~]# vim /etc/profile
+JAVA_HOME=/usr/local/logstash/jdk
+PATH=$PATH:$JAVA_HOME/bin
+export JAVA_HOME PATH
+[root@logstash ~]# source /etc/profile
+[root@logstash ~]# java -version
+openjdk version "17.0.11" 2024-04-16
+OpenJDK Runtime Environment Temurin-17.0.11+9 (build 17.0.11+9)
+OpenJDK 64-Bit Server VM Temurin-17.0.11+9 (build 17.0.11+9, mixed mode, sharing)
+```
+
+### 4. 数据流向
+
+
+
+### 5. 模拟运行
+
+**输入和输出都来自于终端**
+
+```bash
+[root@logstash ~]# /usr/local/logstash/bin/logstash -e 'input { stdin { } } output { stdout {} }'
+nihao (INPUT)
+{
+ "@version" => "1",
+ "message" => "nihao (INPUT)",
+ "@timestamp" => 2025-05-04T08:54:48.278302130Z,
+ "event" => {
+ "original" => "nihao (INPUT)"
+ },
+ "host" => {
+ "hostname" => "logstash"
+ }
+}
+```
+
+**输入来自终端,输出到ES集群**
+
+Elasticsearch 从8.0开始, 默认开启了SSL安全验证
+
+因此我们需要为 Logstash 配置身份验证凭据才能建立与 Elasticsearch 集群的通信
+
+Logstash创建目录存储Elasticsearch的CA证书
+
+```bash
+[root@logstash ~]# mkdir /usr/local/logstash/config/certs
+```
+
+Elasticsearch中将证书拷贝到Logstash
+
+```bash
+[root@logstash ~]# scp 192.168.159.131:/usr/local/elasticsearch/config/certs/elasticsearch-ca.pem /usr/local/logstash/config/certs/
+```
+
+创建Logstash采集数据配置文件
+
+```bash
+[root@logstash ~]# vim /opt/stdin.conf
+input {
+ stdin{}
+}
+output {
+ elasticsearch {
+ index => "stdin_logstash"
+ hosts => [ "https://192.168.159.131:9200" ]
+ cacert => "/usr/local/logstash/config/certs/elasticsearch-ca.pem"
+ user => "elastic"
+ password => "123456"
+ }
+}
+```
+
+执行Logstash
+
+```bash
+[root@logstash ~]# /usr/local/logstash/bin/logstash -f /opt/stdin.conf
+hello world
+hello wxin
+nice to meet you
+```
+
+ES插件查看索引信息
+
+
+
+**输入来自日志文件,输出到ES集群**
+
+此处只展示Logstash配置文件
+
+```bash
+[root@logstash ~]# vim /opt/nginx_access_logstash.conf
+input {
+ file {
+ path => "/var/log/nginx/access.log"
+ start_position => "beginning"
+ }
+}
+output {
+ elasticsearch {
+ index => "nginx_access_logstash"
+ hosts => [ "https://192.168.159.131:9200" ]
+ cacert => "/usr/local/logstash/config/certs/elasticsearch-ca.pem"
+ user => "elastic"
+ password => "123456"
+ }
+}
+```
+
+ES插件展示索引
+
+
+
+Kibana展示数据
+
+
+
+**输入来自日志多文件,输出到ES集群**
+
+Logstash采集配置文件
+
+```bash
+[root@logstash ~]# vim /opt/files.conf
+input {
+ file {
+ path => "/var/log/message"
+ type => "system"
+ start_position => "beginning"
+ }
+}
+input {
+ file {
+ path => "/var/log/yum.log"
+ type => "safeware"
+ start_position => "beginning"
+ }
+}
+output {
+ if [type] == 'system' {
+ elasticsearch {
+ index => "system-%{+YYYY.MM.dd}"
+ hosts => [ "https://192.168.159.131:9200" ]
+ cacert => "/usr/local/logstash/config/certs/elasticsearch-ca.pem"
+ user => "elastic"
+ password => "123456"
+ }
+ }
+ if [type] == 'safeware' {
+ elasticsearch {
+ index => "safeware-%{+YYYY.MM.dd}"
+ hosts => [ "https://192.168.159.131:9200" ]
+ cacert => "/usr/local/logstash/config/certs/elasticsearch-ca.pem"
+ user => "elastic"
+ password => "123456"
+ }
+ }
+}
+```
+
+ES插件展示索引
+
+
+
+
+
+## 三:数据过滤
+
+### 1. 插件 grok 简介
+
+- grok插件有非常强大的功能,他能匹配一切数据,但是他的性能和对资源的损耗同样让人诟病
+- filter的grok是目前logstash中解析非结构化日志数据最好的方式
+- grok位于正则表达式之上,所以任何正则表达式在grok中都是有效的
+
+**官网地址:**
+
+正则:https://github.com/logstash-plugins/logstash-patterns-core/blob/main/patterns/ecs-v1/grok-patterns
+
+### 2. 语法格式
+
+- Grok 模式的语法是:%{SYNTAX:SEMANTIC}
+- 是`SYNTAX`将与您的文本匹配的模式的名称
+- 是`SEMANTIC`您为匹配的文本片段指定的标识符
+
+### 3. 应用案例
+
+**针对nginx的访问日志,获取对应的IP地址**
+
+Logstash采集数据配置文件:
+
+```bash
+[root@logstash ~]# vim /opt/grok_nginx_access_logstash.conf
+input {
+ file {
+ path => "/var/log/nginx/access.log"
+ start_position => "beginning"
+ }
+}
+filter {
+ grok {
+ match => {
+ "message" => "%{IP:remote_addr}"
+ }
+ }
+}
+output {
+ elasticsearch {
+ index => "grok_nginx_access_logstash"
+ hosts => [ "https://192.168.159.131:9200" ]
+ cacert => "/usr/local/logstash/config/certs/elasticsearch-ca.pem"
+ user => "elastic"
+ password => "123456"
+ }
+}
+```
+
+执行:
+
+```bash
+[root@logstash ~]# /usr/local/logstash/bin/logstash -f /opt/grok_nginx_access_logstash.conf
+```
+
+ES插件确认索引:
+
+
+
+Kibana查看新字段:
+
+
+
+**针对nginx的访问日志,对应生成新的字段**
+
+Logstash采集数据配置文件:
+
+```bash
+[root@logstash ~]# vim /opt/grok_nginx_access_logstash.conf
+input {
+ file {
+ path => "/var/log/nginx/access.log"
+ start_position => "beginning"
+ }
+}
+filter {
+ grok {
+ match => {
+ "message" => "%{IPORHOST:client_ip} - %{DATA:user} \[%{HTTPDATE:timestamp}\] \"%{WORD:method} %{DATA:request} HTTP/%{NUMBER:httpversion}\" %{NUMBER:response_code} %{NUMBER:bytes_sent} \"%{DATA:referrer}\" \"%{DATA:user_agent}\""
+ }
+ }
+ date {
+ match => [ "timestamp", "dd/MM/YYYY:HH:mm:ss Z" ]
+ }
+}
+output {
+ elasticsearch {
+ index => "grok_nginx_access_logstash"
+ hosts => [ "https://192.168.159.131:9200" ]
+ cacert => "/usr/local/logstash/config/certs/elasticsearch-ca.pem"
+ user => "elastic"
+ password => "123456"
+ }
+}
+```
+
+Kibana查看新字段:
+
+
\ No newline at end of file