Logprep allows to collect, process and forward log messages from various data sources. Log messages are being read and written by so-called connectors. Currently, connectors for Kafka, Opensearch, S3, HTTP and JSON(L) files exist.
The log messages are processed in serial by a pipeline of processors, where each processor modifies an event that is being passed through. The main idea is that each processor performs a simple task that is easy to carry out. Once the log message is passed through all processors in the pipeline the resulting message is sent to a configured output connector.
Logprep is primarily designed to process log messages. Generally, Logprep can handle JSON messages, allowing further applications besides log handling.
This readme provides basic information about the following topics:
- About Logprep
- Getting Started
- Deployment Examples
- Event Generation
- Documentation
- Contributing
- License
- Changelog
More detailed information can be found in the Documentation.
Logprep processes incoming log messages with a configured pipeline that can be spawned
multiple times via multiprocessing.
The following chart shows a basic setup that represents this behaviour.
The pipeline consists of three processors: the Dissector
, Geo-IP Enricher
and the
Dropper
.
Each pipeline runs concurrently and takes one event from it's Input Connector
.
Once the log messages is fully processed the result will be forwarded to the Output Connector
,
after which the pipeline will take the next message, repeating the processing cycle.
flowchart LR
A1[Input\nConnector] --> B
A2[Input\nConnector] --> C
A3[Input\nConnector] --> D
subgraph Pipeline 1
B[Dissector] --> E[Geo-IP Enricher]
E --> F[Dropper]
end
subgraph Pipeline 2
C[Dissector] --> G[Geo-IP Enricher]
G --> H[Dropper]
end
subgraph Pipeline n
D[Dissector] --> I[Geo-IP Enricher]
I --> J[Dropper]
end
F --> K1[Output\nConnector]
H --> K2[Output\nConnector]
J --> K3[Output\nConnector]
Every processor has one simple task to fulfill.
For example, the Dissector
can split up long message fields into multiple subfields
to facilitate structural normalization.
The Geo-IP Enricher
, for example, takes an ip-address and adds the geolocation of it to the
log message, based on a configured geo-ip database.
Or the Dropper
deletes fields from the log message.
As detailed overview of all processors can be found in the processor documentation.
To influence the behaviour of those processors, each can be configured with a set of rules. These rules define two things. Firstly, they specify when the processor should process a log message and secondly they specify how to process the message. For example which fields should be deleted or to which IP-address the geolocation should be retrieved.
For performance reasons on startup all rules per processor are aggregated to a generic and a specific rule tree, respectively. Instead of evaluating all rules independently for each log message the message is checked against the rule tree. Each node in the rule tree represents a condition that has to be meet, while the leafs represent changes that the processor should apply. If no condition is met, the processor will just pass the log event to the next processor.
The following chart gives an example of such a rule tree:
flowchart TD
A[root]
A-->B[Condition 1]
A-->C[Condition 2]
A-->D[Condition 3]
B-->E[Condition 4]
B-->H(Rule 1)
C-->I(Rule 2)
D-->J(rule 3)
E-->G(Rule 4)
To further improve the performance, it is possible to prioritize specific nodes of the rule tree,
such that broader conditions are higher up in the tree.
And specific conditions can be moved further down.
Following json gives an example of such a rule tree configuration.
This configuration will lead to the prioritization of tags
and message
in the rule tree.
{
"priority_dict": {
"category": "01",
"message": "02"
},
"tag_map": {
"check_field_name": "check-tag"
}
}
Instead of writing very specific rules that apply to single log messages, it is also possible to define generic rules that apply to multiple messages. It is possible to define a set of generic and specific rules for each processor, resulting in two rule trees.
Connectors are responsible for reading the input and writing the result to a desired output. The main connectors that are currently used and implemented are a kafka-input-connector and a kafka-output-connector allowing to receive messages from a kafka-topic and write messages into a kafka-topic. Addionally, you can use the Opensearch or Opensearch output connectors to ship the messages directly to Opensearch or Opensearch after processing.
The details regarding the connectors can be found in the input connector documentation and output connector documentation.
To run Logprep, certain configurations have to be provided. Because Logprep is designed to run in a containerized environment like Kubernetes, these configurations can be provided via the filesystem or http. By providing the configuration via http, it is possible to control the configuration change via a flexible http api. This enables Logprep to quickly adapt to changes in your environment.
First, a general configuration is given that describes the pipeline and the connectors, and lastly, the processors need rules in order to process messages correctly.
The following yaml configuration shows an example configuration for the pipeline shown in the graph above:
process_count: 3
timeout: 0.1
pipeline:
- dissector:
type: dissector
specific_rules:
- https://your-api/dissector/
generic_rules:
- rules/01_dissector/generic/
- geoip_enricher:
type: geoip_enricher
specific_rules:
- https://your-api/geoip/
generic_rules:
- rules/02_geoip_enricher/generic/
tree_config: artifacts/tree_config.json
db_path: artifacts/GeoDB.mmdb
- dropper:
type: dropper
specific_rules:
- rules/03_dropper/specific/
generic_rules:
- rules/03_dropper/generic/
input:
mykafka:
type: confluentkafka_input
bootstrapservers: [127.0.0.1:9092]
topic: consumer
group: cgroup
auto_commit: true
session_timeout: 6000
offset_reset_policy: smallest
output:
opensearch:
type: opensearch_output
hosts:
- 127.0.0.1:9200
default_index: default_index
error_index: error_index
message_backlog_size: 10000
timeout: 10000
max_retries:
user: the username
secret: the passord
cert: /path/to/cert.crt
The following yaml represents a dropper rule which according to the previous configuration
should be in the rules/03_dropper/generic/
directory.
filter: "message"
drop:
- message
description: "Drops the message field"
The condition of this rule would check if the field message
exists in the log.
If it does exist then the dropper would delete this field from the log message.
Details about the rule language and how to write rules for the processors can be found in the rule configuration documentation.
For installation instructions see: https://logprep.readthedocs.io/en/latest/installation.html For execution instructions see: https://logprep.readthedocs.io/en/latest/user_manual/execution.html
A config_refresh_interval
can be set to periodically and automatically refresh the given configuration.
This can be useful in case of containerized environments (such as Kubernetes), when pod volumes often change
on the fly.
If the configuration does not pass a consistency check, then an error message is logged and Logprep keeps running with the previous configuration. The configuration should be then checked and corrected on the basis of the error message.
The documentation for Logprep is online at https://logprep.readthedocs.io/en/latest/ or it can be built locally via:
sudo apt install pandoc
pip install -e .[doc]
cd ./doc/
make html
A HTML documentation can be then found in doc/_build/html/index.html
.