1.搭建Elasticsearch環境並測試:
(1)刪除es的容器
(2)刪除es的鏡像
(3)宿主機調內存: 執行命令:sudo sysctl -w vm.max_map_count=655360
(4)通過ftp軟件修改docker-compose.yml中的 mem_limit: 2048M
(5)找到虛擬機, 執行命令:cd /home/px2/envdm/springcloudV2.0/
執行命令:docker-compose up -d elasticsearch
(6)測試es:http://192.168.115.158:9200 出現es的版本號說明配置環境成功。
2.搭建logstash
(1)打開物料中logstash
步驟:進入logstash的容器
vi /usr/local/logstash-6.3.0/config/logstash.yml(修改連接es的ip)
修改成:
http.host: "0.0.0.0"
xpack.monitoring.elasticsearch.url: http://192.168.115.158:9200
xpack.monitoring.elasticsearch.username: elastic
xpack.monitoring.elasticsearch.password: changeme
xpack.monitoring.enabled: false
(2)vi /usr/local/logstash-6.3.0/bin/logstash.conf
修改ip以及加入日志參數
input{
kafka{
bootstrap_servers => ["192.168.115.158:9092"]
group_id => "test-consumer-group"
auto_offset_reset => "latest"
consumer_threads => 5
decorate_events => true
topics => ["dm"]
type => "bhy"
}
}
output {
elasticsearch{
hosts=> ["192.168.115.158:9200"]
index=> "dmservice-%{+YYYY.MM.dd}"
}
stdout{
codec => json_lines
}
}
3.修改kibana配置文件
(1)找到elasticsearch.url:並修改成 "http://192.168.115.158:9200"
(2)訪問:http://192.168.115.158:5601沒有出現無es數據即說明成功。
4.kafka
(1)找到server.properties
listeners=PLAINTEXT://0.0.0.0:9092
advertised.listeners=PLAINTEXT://192.168.115.158:9092
(2)啟動kafka命令
./kafka-console-consumer.sh --bootstrap-server 127.0.0.1:9092 --topic dm --from-beginning