Visualizing Logs from a Dockerized Node Application Using the Elastic Stack

栏目: IT技术 · 发布时间: 4年前

内容简介:Application logsare events that are typically outputted by an application. It contains sometimes detailed information about the faults and warnings that occur within the said application.Logs play an important part in running and maintaining an application
Visualizing Logs from a Dockerized Node Application Using the Elastic Stack

Visualizing Logs from a Dockerized Node Application Using the Elastic Stack

Application logsare events that are typically outputted by an application. It contains sometimes detailed information about the faults and warnings that occur within the said application.

Logs play an important part in running and maintaining an application; logs help during debugging, in the case of web servers, it can be used to track incoming requests and potential malicious attacks.

There are five different levels of logs categorized by their severity:

  • Debug
  • Info
  • Warning
  • Error
  • Critical

In this article, we’ll see how we can stream the logs from a Node application to a log visualization application like Kibana . Kibana is part of the Elastic Stack, it gets fed data from Elasticsearch and with that data, we can track, query and review how requests flow through our application.

Prerequisites

Before you begin this guide you’ll need the following:

  • A working PC with Node and NPM installed.
  • A working installation of Docker.
  • A basic understanding of Javascript.

Building the Node application

To get started, we’ll need a create a simple Node application that outputs logs detailing incoming requests. The library of choice for this application is Express; with Express , we can easily bootstrap a lightweight application that serves our purpose.

Create a new folder called logs - demo ; next, we’ll initialize NPM within the directory by running the following command:

npm init

Running this command should create a package . json file in the directory. Install the following dependencies next:

npm install express morgan

morgan is an HTTP request logger middleware for node.js. It is a great tool that can be used to log requests, errors, and more to the console. With the help of morgan, we’ll have details of requests like the status, response time and date logged.

Create a file index . js within the logs - demo directory and copy the following content into it:

const app = require("express")();
const morgan = require("morgan");
 
app.use(morgan("common"));
 
app.get("/", (req, res) => {
  console.log("A new request was here");
  res.send("Hello there");
});
 
app.listen(4000, () => {
  console.log("Listening on port 4000");
});

The application is a pretty basic one. We set up the server to listen on port 4000 , the base route responds with a string but not before logging to the console. The app also uses the morgan middleware with the common log format. Start the server by running the following command:

node index

Then test the server using curl by running the following command in a different terminal tab or window:

curl http://localhost:4000

You should see Hello there as the resulting response from the server. Meanwhile, in the terminal tab/window that has the server running, you should see an output similar to the snippet below on the console:

Listening on port 4000
A new request here
127.0.0.1 - - [12/Jan/2020:06:48:15 +0000] "GET / HTTP/1.1" 200 11

Now, we have a working server that logs content to the console. The next step is to run the application in a docker file. In the next section, we’ll create a Dockerfile to configure the container.

Dockerize the Node application

Before writing the Dockerfile , we have to make sure that we have Docker installed and working on our system. If you haven’t installed Docker already, follow the instructions for your OS here to install setup Docker on your PC. After installing, run the following command to test the installation:

docker version --format '{{.Server.Version}}'

This command should output your current Docker version. If it doesn’t, then you might need to troubleshoot your installation.

Now that we have Docker installed, let’s create a Dockerfile for our application. Create a file within the logs - demo directory, name the file Dockerfile .

The first thing we do in the Dockerfile is to select the image we want to build from. In this case, we select node : 10 - alpine which is based on the Alpine Linux project . Alpine Linux is lightweight which leads to smaller images.

FROM node:10-alpine

Create a directory / usr / src / app to hold the files for our application; this newly created directory is also set as the working directory using the WORKDIR command.

# Create working directory 
RUN mkdir -p /usr/src/app
 
# Select the newly created directory as the work directory
WORKDIR /usr/src/app

Next, we use the wildcard to copy the package . json and package - lock . json files alone; this is done to take advantage of cached Docker layers.

# Copy package.json file
COPY package*.json ./

Next, we install the dependencies by running npm install .

# Install app dependencies
RUN npm install

Then we bundle our application’s resources using the COPY instruction.

# Get all the code needed to run the app
COPY . /usr/src/app

Before running the application, we EXPOSE the port which our app binds to ( 4000 ).

# expose port 4000
EXPOSE 4000

Finally, we define the command that starts the server using CMD .

# Run the app
CMD ["node", "index"]

Altogether, the Dockerfile should look like this:

#Dockerfile
FROM node:10-alpine
 
# Create working directory 
RUN mkdir -p /usr/src/app
 
# Select the newly created directory as the work directory
WORKDIR /usr/src/app
 
# Copy package.json file
COPY package*.json ./
 
# Install app dependencies
RUN npm install
 
# Get all the code needed to run the app
COPY . /usr/src/app
 
# expose port 4000
EXPOSE 4000
 
# Run the app
CMD ["node", "index"]

When copying files from your directory, we wouldn’t want Docker to necessarily copy everything within the directory. To help with that, we can define a . dockerignore file to specify the folders or files we wish to ignore. Create a . dockerignore file within the logs - demo directory and copy the following content into it:

//.dockerignore
node_modules

This will prevent docker from copying the local node_modules folder which could potentially lead to a clash with the modules installed on the image.

Building and running the image

In the logs - demo directory, run the command below to build the Docker image.

docker build -t <your username>/logs-demo .

The - t flag is used to tag the image for easy identification. Next, we are going to run the application in the background, we’ll also bind the private port exposed within the image to a public one. We’ll run the image using the - d flag to run the image in a detached mode and with the - p flag to specify the redirect port.

docker run -p 8082:4000 -d <your username>/logs-demo

To check the details of your running container, run the following command:

$ docker ps
 
# Example 
outputID            IMAGE                                COMMAND    ...   PORTS
ecce33b30ebf  <your username>/logs-demo:latest  npm start  ...   8082->4000

You can try making requests to the application using cURL. We bound port 8082 to the exposed port in the container so:

$ curl http://localhost:8082
 
Hello there

We now have a running containerized version of our application. In the next section, we’ll look at how to set up Elasticsearch and Kibana.

Setting up Elasticsearch

To get up and running with Elasticsearch, follow any of the following installation commands that apply to your system.

For Debian:

curl -L -O https://artifacts.elastic.co/downloads/elasticsearch/elasticsearch-7.5.1-amd64.deb
sudo dpkg -i elasticsearch-7.5.1-amd64.deb
sudo /etc/init.d/elasticsearch start

For Mac using brew:

brew tap elastic/tap
brew install elastic/tap/elasticsearch-full
elasticsearch

For Windows:

  1. Download the Elasticsearch 7.5.1 Windows zip file from the Elasticsearch download page .
  2. Extract the contents of the zip file to a directory on your computer.
  3. Open a command prompt as an Administrator and navigate to the directory that contains the extracted files, for example:
cd C:\Program Files\elasticsearch-7.5.1

If you can’t installation instructions for your page here, you can check the official documentation to see if it’s listed.

Start Elasticsearch:

bin\elasticsearch.bat

To test that the Elasticsearch daemon is up and running, try sending an HTTP GET request on port 9200.

curl http://127.0.0.1:9200

On Windows, if you don’t have cURL installed, point your browser to the URL. You should see a response similar to this:

{
  "name": "ip-171-31-8-14",
  "cluster_name": "elasticsearch",
  "cluster_uuid": "bDaHmaaURmicHPjCIju3qA",
  "version": {
    "number": "7.5.1",
    "build_flavor": "default",
    "build_type": "deb",
    "build_hash": "3ae9ac9a93c95bd0cdc154951cf95d88e1e18d96",
    "build_date": "2019-12-16T22:57:37.835892Z",
    "build_snapshot": false,
    "lucene_version": "8.3.0",
    "minimum_wire_compatibility_version": "6.8.0",
    "minimum_index_compatibility_version": "6.0.0-beta1"
  },
  "tagline": "You Know, for Search"
}

Now that we have Elasticsearch up and running, we’ll get to set up Kibana. Setting up should involve steps similar to the setup for Elasticsearch.

Setting up Kibana

Kibana is an open-source analytics and visualization platform designed to work with Elasticsearch. With Kibana, you can search, view, and interact with data stored in Elasticsearch indices. Run the following command to install Kibana on your system:

For deb , rpm , or linux :

curl -L -O https://artifacts.elastic.co/downloads/kibana/kibana-7.5.1-linux-x86_64.tar.gz
tar xzvf kibana-7.5.1-linux-x86_64.tar.gz
cd kibana-7.5.1-linux-x86_64/
./bin/kibana

For brew :

brew tap elastic/tap
brew install elastic/tap/kibana-full
kibana

For Windows :

  1. Download the Kibana 7.5.1 Windows zip file from the Kibana download page.
  2. Extract the contents of the zip file to a directory on your computer, for example, C : \ Program Files .
  3. Open a command prompt as an Administrator and navigate to the directory that contains the extracted files, for example:
cd C:\Program Files\kibana-7.5.1-windows

Start Kibana :

bin\kibana.bat

If you can’t find install instructions for your system here, you can check the Kibana official downloads page.

After installing and starting Kibana, you can visit the Kibana web interface on port 5601 . http://127.0.0.1:5601 .

Now that we have Elasticsearch and Kibana up and running, you’re probably asking yourself “How do we get the data into Elasticsearch for Kibana to display?”. The answer to that lies in the next section. Okay let me give a hint; we’ll be using Beats; Filebeats to be precise. We’ll get to installing and configuring Filebeats to send data to Elasticsearch.

Installing and configuring Filebeats

Filebeats belongs to a list of data transporters collectively known as Beats. Beats are data transporters are installed on servers to send all kinds of data to Elasticsearch. Asides from Filebeats, some other available Beats are:

  • Metricbeats : for collecting metrics from systems
  • Heartbeat : for monitoring service availability.

You can check the rest of the list here.

To install Filebeats you can run any of the following commands listed below:

For debian :

curl -L -O https://artifacts.elastic.co/downloads/beats/filebeat/filebeat-7.5.1-amd64.deb
sudo dpkg -i filebeat-7.5.1-amd64.deb

For brew :

brew tap elastic/tap
brew install elastic/tap/filebeat-full

For Windows

  1. Download the Filebeat Windows zip file from the downloads page.
  2. Extract the contents of the zip file into C : \ Program Files .
  3. Rename the filebeat - < version > - windows directory to Filebeat .
  4. Open a PowerShell prompt as an Administrator (right-click the PowerShell icon and select Run As Administrator).
  5. From the PowerShell prompt, run the following commands to install Filebeat as a Windows service:
PS > cd 'C:\Program Files\Filebeat'
PS C:\Program Files\Filebeat> .\install-service-filebeat.ps1

Configuring Filebeats

To configure Filebeats, we need to edit the filebeat . yml file. The filebeat.yml file contains the default configuration.  The location of the file varies by platform, to locate the file, see Directory layout . Once you locate the file, open it and make the following changes to it:

FileBeat has an input type called docker that is specifically designed to import logs from docker. You will need to specify which of the docker containers’ logs you want to target. In our case, we want to get logs for all the containers. So we’ll use a wildcard value "*" . Update the filebeats . inputs block:

filebeat.inputs:
 - type: docker
     containers.ids: "*"

Next, we update the processors block; within it, we add_docker_metadata . The goal is to add more detail to the docker logs, such as the name of the image or the name of the container. Without this processor, only the ids of the containers will show up in the logs.

processors:
- add_docker_metadata:    
  host: "unix:///var/run/docker.sock"

The other processor is decode_json_fields , it is useful for parsing JSON encoded logs. The logs in FileBeat, ElasticSearch, and Kibana consist of multiple fields and the message field is what the application (running inside a docker container) writes to the standard output. This message is only a string, but it may contain useful information such as the log level.

- decode_json_fields:    
  fields: ["message"]    
  target: "json"    
  overwrite_keys: true

The elasticsearch setting that is quite straightforward. It lets you configure the ElasticSearch address and the indices where the logs are imported. The index pattern setup filebeat - % { [ agent . version ] } - % { + yyyy . MM . dd } includes the date which shows that the logs are imported into the index according to the date of occurrence.

output.elasticsearch:  
  hosts: ["localhost:9200"]  
  index: "filebeat-%{[agent.version]}-%{+yyyy.MM.dd}"

Setup Kibana dashboards

Filebeat comes packaged with preset Kibana dashboards, visualizations, and searches for visualizing Filebeat data in Kibana. We need to create an index pattern and load the dashboards into Kibana before we can make use of the dashboards. To do this, you can run the command for your system specified below:

deband rpm :

filebeat setup --dashboards

mac:

./filebeat setup --dashboards

brew:

filebeat setup --dashboards

win:

Open a PowerShell prompt as an Administrator (right-click the PowerShell icon and select Run As Administrator).

From the PowerShell prompt, change to the directory where you installed Filebeat, and run:

PS > .\filebeat.exe setup --dashboards

After running this command, you should see an output similar to the one below if the dashboards were loaded successfully:

Loading dashboards (Kibana must be running and reachable)
Loaded dashboards

We have Elasticsearch installed and running, same with Kibana and we have also loaded preset dashboards on Kibana using Filebeat. Now, the moment we’ve been waiting for (*drum roll*), let’s start Filebeat. Starting Filebeat will stream out container’s logs to Elasticsearch and then will be displayed by Kibana. To start Filebeat, run the command that applies to your system below:

deband rpm :

sudo service filebeat start

macand linux :

sudo chown root filebeat.yml
sudo ./filebeat -e

You’ll be running Filebeat as root, so you need to change ownership of the configuration file, or run Filebeat with -- strict . perms = false specified. See Config File Ownership and Permissions in the Beats Platform Reference.

brew:

To have launchd start elastic/tap/filebeat and then restart it at login, run:

brew services start elastic/tap/filebeat-full

To run Filebeat in the foreground instead of running it as a background service, run:

filebeat -e

win:

PS C:\Program Files\Filebeat> Start-Service filebeat

At this moment, Filebeat should be sending logs to Elasticsearch already. Let’s check out Kibana to see if we can find our logs. Visit this link on your browser http://localhost:5601/app/infra#/logs/stream . You should see a view similar to the one below:

Visualizing Logs from a Dockerized Node Application Using the Elastic Stack

Kibana Sidebar

In the screenshot above, we can see the logs from our application shown. To view more details, we can expand each log by clicking on the expand icon annotated above. Clicking the button above will show a screen similar to the one below. Showing more details about the container where the logs emanated from:

Visualizing Logs from a Dockerized Node Application Using the Elastic Stack

Kibana

You can see information like the container name in container.image.name and the container id in container.id . More details can be found in the opened sidebar.

Conclusion

In this article, we’ve seen how to set up and dockerize a minimal Node application. We didn’t stop there, we also set up Elasticsearch, Kibana, and Filebeat to stream and visualize logs from the container. You can take this setup a step further by adding Logstash to parse and transform the logs from the application.

Stay good and happy coding.


以上就是本文的全部内容,希望对大家的学习有所帮助,也希望大家多多支持 码农网

查看所有标签

猜你喜欢:

本站部分资源来源于网络,本站转载出于传递更多信息之目的,版权归原作者或者来源机构所有,如转载稿涉及版权问题,请联系我们

算法技术手册(原书第2版)

算法技术手册(原书第2版)

George T. Heineman、Gary Pollice、Stanley Selkow / 杨晨、曹如进 / 机械工业出版社 / 2017-8-1 / 89.00元

本书使用实际代码而非伪代码来描述算法,并以经验主导支撑数学分析,侧重于应用且规范严谨。本书提供了用多种程序设计语言实现的文档化的实际代码解决方案,还介绍了近40种核心算法,其中包括用于计算点集的Voronoi图的Fortune算法、归并排序、多线程快速排序、AVL平衡二叉树实现以及空间算法。一起来看看 《算法技术手册(原书第2版)》 这本书的介绍吧!

JSON 在线解析
JSON 在线解析

在线 JSON 格式化工具

html转js在线工具
html转js在线工具

html转js在线工具

RGB HSV 转换
RGB HSV 转换

RGB HSV 互转工具