The ELK stack consists of Elasticsearch, Logstash, and Kibana.
They provide a powerful, flexible, and scalable solution for managing and making sense of large amounts of data.
Logstash: a data processing pipeline which gathers, processes and forwards data (logs, metrics) from various sources to Elasticsearch.
Elasticsearch: a distributed search and analytics engine that stores this processed data and enables powerful search and analytics capabilities.
Kibana: a data visualization and exploration tool. It provides a graphical interface to visualize and explore the data stored in Elasticsearch.
Benefits of ELK Stack
Real-time Insights: They enable real-time data processing and visualization, which is crucial for monitoring and quick decision-making.
Scalability: They can handle large-scale data operations, making them suitable for big data applications.
Flexibility: They support a wide range of data types and sources, providing flexibility in how data is ingested, stored, and analyzed.
We'll implement the workflow below:
-
Provision Servers with Terraform
-
User Configuration on Linux Servers
-
Setup ELK Stack in Central Server
-
Enable ELK Clustering
-
Add Remote Hosts
-
Develop Kibana Visualization
-
Create Kibana Dashboard
Machine image used in terraform configuration is CentOS 8.
Install AWS CLI in local machine
sudo apt install curl unzip
curl "https://awscli.amazonaws.com/awscli-exe-linux-x86_64.zip" -o "awscliv2.zip"
unzip awscliv2.zip
sudo ./aws/install -i /usr/local/aws-cli -b /usr/local/binConfirm the AWS CLI installation
aws --versionClone this repository in the local machine
cd /
git clone git@github.com:odennav/elk-centralized-syslog-system.gitExecute these Terraform commands sequentially in the local machine to create the AWS VPC(Virtual Private Cloud) and EC2 instances.
Initializes terraform working directory
cd elk-centralized-syslog-system/terraform
terraform initValidate the syntax of the terraform configuration files
terraform validateCreate an execution plan that describes the changes terraform will make to the infrastructure
terraform planApply the changes described in execution plan
terraform apply -auto-approveCheck AWS console for instances created and running
SSH access
Use .pem key from AWS to SSH into the public EC2 instance. IPv4 address of public EC2 instance will be shown in terraform outputs.
ssh -i private-key/terraform-key.pem ec2-user@<ipaddress>We can use public EC2 instance as a jumpbox to securely SSH into private EC2 instances within the VPC.
Note, the ansible inventory is built dynamically by terraform with the private ip addresses of the EC2 machines.
Add New User
We'll use the central-server-1 virtual machine as our build machine. Integrations to pipeline is implemented on this server
Change password for root user
sudo passwdSwitch to root user. Add new user 'odennav' to sudo group.
sudo useradd odennav
sudo usermod -aG wheel odennavNotice the prompt to enter your user password. To disable password prompt for every sudo command, implement the following:
Add sudoers file for odennav-admin
echo "odennav ALL=(ALL) NOPASSWD: ALL" | sudo tee /etc/sudoers.d/odennavEnsure correct permissions for sudoers file
sudo chmod 0440 /etc/sudoers.d/odennav
sudo chown root:root /etc/sudoers.d/odennavTest sudo privileges by switching to new user
su - odennav
sudo ls -la /rootTo change the PermitRootLogin setting, modify the SSH server configuration file /etc/ssh/sshd_config as shown below:
PermitRootLogin noRestart the SSH service for the changes to take effect
sudo systemctl restart sshdVerify the the configuration has been applied
sudo grep PermitRootLogin /etc/ssh/sshd_configInstall Elasticsearch with RPM
Download from elastic website and install the RPM manually
wget https://artifacts.elastic.co/downloads/elasticsearch/elasticsearch-8.13.4-x86_64.rpmDownload the SHA512 checksum file
wget https://artifacts.elastic.co/downloads/elasticsearch/elasticsearch-8.13.4-x86_64.rpm.sha512Verify the checksum
shasum -a 512 -c elasticsearch-8.13.4-x86_64.rpm.sha512Comparing the SHA of the downloaded RPM and the published checksum, should output
elasticsearch-8.13.4-x86_64.rpm: OK.
When the checksums match, this confirms that the file is intact and hasn't been tampered with.
Install the RPM
sudo rpm --install elasticsearch-8.13.4-x86_64.rpmWhen installing Elasticsearch, security features are enabled and configured by default.
The password and certificate and keys are output to your terminal.
Store the elastic password as an environment variable in your shell.
export ELASTIC_PASSWORD="my_password"Name Elasticsearch Cluster
sudo vi /etc/elasticsearch/elasticsearch.ymlFor a production environment, it's beneficial to have shards distributed. We'll configure elasticsearch to communicate with outside network and look for an additional node, central-server-2.
Add this to end of elasticsearch.yml and save the configuration.
cluster.name: syslog
node.name: central-server-1
network.host: [10.33.10.1, _local_]
discovery.zen.ping.unicast.hosts: ["10.33.10.1", "10.33.10.6"]
action.auto_create_index: .monitoring*,.watches,.triggered_watches,.watcher-history*,.ml*Start and enable elasticsearch service
sudo systemctl daemon-reload
sudo systemctl start elasticsearch.service
sudo systemctl enable elasticsearch.service
Confirm connection to elasticsearch
curl --cacert /etc/elasticsearch/certs/http_ca.crt -u elastic:$ELASTIC_PASSWORD https://localhost:9200 Install Logstash with RPM
Ensure java is available for Logstash
sudo yum install -y java-1.8.0-openjdkDownload and install the public signing key
sudo rpm --import https://artifacts.elastic.co/GPG-KEY-elasticsearchCreate logstash.repo in /etc/yum.repos.d/
cd /etc/yum.repos.d/
sudo touch logstash.repoAdd this to logstash.repo file
[logstash-8.x]
name=Elastic repository for 8.x packages
baseurl=https://artifacts.elastic.co/packages/8.x/yum
gpgcheck=1
gpgkey=https://artifacts.elastic.co/GPG-KEY-elasticsearch
enabled=1
autorefresh=1
type=rpm-md
Install logstash
sudo yum install logstashConfigure Logstash
Create the syslog configuration file for logstash
sudo touch /etc/logstash/conf.d/syslog.confOur logstash configuration will have three main blocks:
-
Input: cause logstash to listen for syslog messages on port 5141 -
Filter: process messages it receives that match the given patterns.It extracts the authentication method, the username, the source IP address, and source
port for ssh connection attempts. Also tags the messages withssh_successful_loginorssh_failed_login. -
Output: store the messages into the elasticsearch instance we just created.
Add this to syslog.conf and save the configuration
input {
syslog {
type => syslog
port => 5141
}
}
filter {
if [type] == "syslog" {
grok {
match => { "message" => "Accepted %{WORD:auth_method} for %{USER:username} from %{IP:src_ip} port %{INT:src_port} ssh2" }
add_tag => "ssh_successful_login"
}
grok {
match => { "message" => "Failed %{WORD:auth_method} for %{USER:username} from %{IP:src_ip} port %{INT:src_port} ssh2" }
add_tag => "ssh_failed_login"
}
grok {
match => { "message" => "Invalid user %{USER:username} from %{IP:src_ip}" }
add_tag => "ssh_failed_login"
}
}
geoip {
source => "src_ip"
}
}
output {
elasticsearch { }
}
Start and enable logstash service
sudo systemctl start logstash.service
sudo systemctl enable logstash.serviceForward Syslogs to Logstash
Next, we configure central-server-1 node tp forward its syslog messages to logstash.
Create logstash configuration file
sudo touch /etc/rsyslog.d/logstash.confAdd this to logstash.conf and save
*.* @10.33.10.1:5141Restart rsyslog service
sudo systemctl restart rsyslogConfirm logstash is now receiving syslog messages from central-server-1 node and and storing them in Elasticsearch.
curl --cacert /etc/elasticsearch/certs/http_ca.crt -u elastic:$ELASTIC_PASSWORD https://localhost:9200/_cat/indices?vInstall Kibana
Download from elastic website and install the RPM manually
wget https://artifacts.elastic.co/downloads/kibana/kibana-8.13.4-x86_64.rpmDownload the SHA512 checksum file
shasum -a 512 -c kibana-8.13.4-x86_64.rpm.sha512Verify the checksum
shasum -a 512 -c kibana-8.13.4-x86_64.rpm.sha512Comparing the SHA of the downloaded RPM and the published checksum, should output
kibana-8.13.4-x86_64.rpm: OK
Install the RPM
sudo rpm --install kibana-8.13.4-x86_64.rpmTo enable connection to Kibana from outside the localhost
sudo vi /etc/kibana/kibana.ymlAdd this to the configuration file, kibana.yml
server.host: "10.33.10.1"Securely connect kibana with elasticsearch
The elasticsearch-create-enrollment-token command creates enrollment tokens for Elasticsearch nodes and Kibana instances.
Generate an enrollment token for kibana
/usr/share/elasticsearch/bin/elasticsearch-create-enrollment-token -s kibanaStart and enable kibana service
sudo systemctl daemon-reload
sudo systemctl start kibana.service
sudo systemctl enable kibana.serviceTo receive feedback whether Kibana was started successfully or not
sudo journalctl -u kibana.serviceBrowse 10.33.10.1:5601 to view Kibana UI and click on Explore on my own link to get started with Elastic.
Configure Node to Join Cluster
When Elasticsearch was installed in first node central-server-1, the installation process configured a single-node cluster by default.
To enable a node to join an existing cluster instead, implement the following:
- Generate an enrollment token on an existing node,
central-server-1before you start the new nodecs2for the first time.
On central-server-1 in our existing cluster, generate a node enrollment token
/usr/share/elasticsearch/bin/elasticsearch-create-enrollment-token -s node-
Copy the enrollment token, which is output to your terminal.
We'll use
central-server-2node as our additional elasticsearch cluster member. -
Implement the same steps done for
central-server-1node incentral-server-2node: Install Elasticsearch, Logstash, Kibana -
Ensure the elasticsearch cluster in second node is named as
syslog
sudo vi /etc/elasticsearch/elasticsearch.ymlAdd this to end of elasticsearch.yml and save the configuration.
cluster.name: syslog
node.name: central-server-2
network.host: [10.33.10.6, _local_]
discovery.zen.ping.unicast.hosts: ["10.33.10.1", "10.3.10.6"]
action.auto_create_index: .monitoring*,.watches,.triggered_watches,.watcher-history*,.ml*This new second server will automatically discover and join the cluster as long as it has the same cluster.name as the first node.
- On your new Elasticsearch node,
central-server-2, pass the enrollment token generated in step1 as a parameter to the elasticsearch-reconfigure-node tool
/usr/share/elasticsearch/bin/elasticsearch-reconfigure-node --enrollment-token <enrollment-token>Start and enable elasticsearch service
sudo systemctl daemon-reload
sudo systemctl start elasticsearch.service
sudo systemctl enable elasticsearch.service
Start other remote machines, run:
vagrant up
vagrant ssh Install Ansible
To install ansibe without upgrading current python version, we'll make use of the yum packae manager.
sudo yum update
sudo yum upgradeInstall EPEL repository
sudo yum install epel-releaseVerify installation of EPEL repository
sudo yum repolistInstall Ansible
sudo yum install ansibleConfirm installation
ansible --versionConfigure Ansible Vault
Ansible communicates with target remote servers using SSH and usually we generate RSA key pair and copy the public key to each remote server, instead we'll use username and password credentials of odennav user.
This credentials are added to inventory host file but encrypted with ansible-vault
Ensure all IPv4 addresses and user variables of remote servers are in the inventory file as shown
View ansible-vault/values.yml which has the secret password
cat /elk-centralized-logging-system/ansible/ansible-vault/values.ymlGenerate vault password file
openssl rand -base64 2048 > /elk-centralized-logging-system/ansible/ansible-vault/secret-vault.passCreate ansible vault with vault password file
ansible-vault create /elk-centralized-logging-system/ansible/ansible-vault/values.yml --vault-password-file=/elk-centralized-logging-system/ansible/ansible-vault/secret-vault.passView content of ansible vault
ansible-vault view /elk-centralized-logging-system/ansible/ansible-vault/values.yml --vault-password-file=/elk-centralized-logging-system/ansible/ansible-vault/secret-vault.passRead ansible vault password from environment variable
export ANSIBLE_VAULT_PASSWORD_FILE=/elk-centralized-logging-system/ansible/ansible-vault/secret-vault.passConfirm environment variable has been exported
export ANSIBLE_VAULT_PASSWORD_FILETest Ansible by pinging all remote servers in inventory list
ansible all -m pingConfigure Remote Hosts
We'll use ansible playbook to configure the remote systems to forward their syslog messages to the centralized syslog server.
ansible-playbook -i hosts.inventory /elk-centralized-logging-system/ansible/add_hosts/add_hosts.ymlCreate Index Pattern
Impplement the following steps below:
-
Return to the Kibana UI at
10.33.10.1:5601. -
Click on the hamburger menu icon, then click on the
Stack Managementlink under theManagementsection of the menu.
- Scroll down. Under the
Kibanasection, click theIndex Patternslink.
-
There will be a pop-up display labeled
About Index Patternsin the right-hand side of your screen. Click thexto close it. -
Now click on the
Create index patternbutton.
- In the
Index pattern namefield, enterlogstash*and then clickNext Step >button.
This tells Kibana to use any indices in Elasticsearch that start with logstash.
- In the
Time Fielddropdown menu, select@timestamp, then click theCreate index patternbutton.
A screen will appear that shows information about the index pattern that we've just created.
Confirm Log Sources from Remote Hosts
Now we can start searching for log messages
- Click on the hamburger menu icon, then click on the
Discoverlink under theKibanasection of the menu.
- At the left-hand
fieldmenu, click onlogsource.
You should now see the other remote hosts appear in addition to the cs1 host.
Now you can search forlog records across multiple hosts in one single place.
Generate Syslog
We'll use two scripts in syslog-gen directory to simulate syslog generation.
The logstash config will process and filter this messages, while also tagging them.
We can then search for this syslog records stored in elasticsearch and visually analyze with Kibana.
Create Kibana Visualization
Next, we use Kibana to create a visualization object that analyzes the unsuccessful ssh login data.
Implement the following steps:
-
Click on the hamburger menu icon, then click on the
Visualizelink under theKibanasection of the menu. -
Click on the
Create new visualizationbutton. -
Scroll down and click
Vertical Bar.
- Next, click on
logstash*index pattern.
-
In the search bar, type in
tags:ssh_failed_loginand click on theRefreshbutton. Set date asLast 1 hourto capture events of multiple failed ssh login attempts. -
Under "Buckets", click on the
+ Addlink. The click onX-Axis. -
Under aggregation, select
Date Histogram. To apply the changes, click theUpdatebutton at the bottom-right of your screen. -
Now we'll see that one big bar break into smaller bars.
When you hover over the bar, it tells us the number of times the search occurs during that time period.
-
Under the
MetricsSection, click onY-Axis. Now supply a custom label ofFailed Logins.
To apply this change, click the Update button in the bottom-right of your screen.
- Finally, click the
Saveicon in the top-left and give this graph a Title ofFailed SSH Logins. ClickSave.
Use Kibana Dashboard
Next, we use Kibana to create a dashboard for the the unsuccessful ssh login visualization object we just created.
Implement the following steps:
-
Click on the hamburger menu icon, then click on the
Dashboardlink under theKibanasection of the menu. -
Click on the
Create new dashboardbutton.
-
Click on the
Add an existing objectlink, then click onFailed SSH Logins.Next, click on the
xto close the pop-up window.Adjust the graph to whatever size you desire.
-
Click on the
savelink at the top of your screen and name the dashboard `SSH Login Analysis" then save it.
The visualization object enables us to represent the syslog data in a meaningful way and also adding them to a dashboard for real-time view.
Enjoy!










