Best Practice: Choosing a Log Shipper

LogScale can ingest data from a variety of sources including:

This KB article covers recommendations specifically related to choosing log shippers to use and how to use them in a highly available way.

Log Shipper Recommendations

The choice of log shippers to use in conjunction with LogScale should be made based on the following factors:

  • Functionality

    Can the log shipper handle all of the log sources that an organization needs to send to their logging platform? Does the log shipper feature the ability to transform events and/or fields if required by the organization?

  • Ease of deployment

    How easy is it to deploy the selected log shipper, configure it to send data to LogScale, and deploy updates as required?

  • Reliability

    How reliable is the chosen log shipper? Does your organization have the ability to troubleshoot issues should they arise?

  • Speed

    Will the log shipper be able to keep up with the volume of data you need to send to LogScale?

LogScale offers it's own Falcon LogScale Collector and additionally supports a wide variety of log shippers (see Third-Party Log Shippers for more information), however the three most commonly used are Filebeat, Vector, and Fluentd.

Comparing LogScale Log Collector, Filebeat, Vector, and Fluentd

Each of these log shippers is extremely capable and have their pros and cons delineated below (however this is not an exhaustive list):

Feature LogScale Log Collector Filebeat Vector Fluentd
Platform Support Installers for Ubuntu Linux, Redhat Linux, Windows, Mac OS Installers for Debian, RPM, Mac OS, Brew, Linux (tar.gz), Windows Amazon Linux, CentOS, Debian, MacOS, NixOS, Raspbian, RHEL, Ubuntu, Windows Installers for RPM, Debian, MacOS, Windows, platforms running Ruby
License Commercial Commercial and OSS; Note: Only the OSS licensed (Apache 2.0) package will work with LogScale OSS - Mozilla Public License 2.0 OSS - Apache 2.0
Centralized Management No, Under development No No No
Deployment Models Agent Agent Agent (daemon, sidecar), aggregator Agent, Aggregator
Supported Inputs/Sources File, Syslog, Windows Event Logs, JournalD, Exec AWS CloudWatch, AWS S3, Azure Event Hub, Cloud Foundry, Container, Docker, GCP Pub/Sub, HTTP Endpoint, HTTP JSON, Kafka, Log, MQTT, NetFlow, Office 365 Management Activity API, Redis, Stdin, Syslog, TCP, UDP; Important: Bolded inputs require X-Pack which is not a part of the open source package and requires a paid license. Apache metrics, AWS Kinesis Firehose, AWS ECS Metrics, AWS S3, Datadog Logs, Docker, File, Generator, Heroku Logplex, Host Metrics, Internal Metrics, Journald, Kafka, Kubernetes Logs, MongoDB Metrics, Nginx Metrics, PostgresSQL Metrics, Prometheus Remote Write, Prometheus Scrape, Socket, Splunk HTTP Event Collector (HEC), StatsD, STDIN, Syslog, Vector Inputs supported via Input Plugins: in_tail, in_forward, in_udp, in_tcp, In_unix, in_http, in_syslog, in_exec, in_sample, in_windows_eventlog. Additional input formats (793 input/output plugins) can be found in the full list here: https://docs.fluentd.org/input
Supported Outputs/Sinks LogScale Elasticsearch Service, Elasticsearch, Logstash, Kafka, Redis, File, Console. Important: Events are typically sent to LogScale from Filebeat using the Elasticsearch output method. AWS Cloudwatch Logs, AWS Cloudwatch Metrics, AWS Kinesis Firehose, AWS Kinesis Data Streams, AWS S3, Amazon Simple Queue Service (SQS), Azure Monitor Logs, Blackhole, Clickhouse, Console, Datadog Logs, Datadog Metrics, Elasticsearch, File, GCP Cloud Storage (GCS), GCS PubSub, GCP Operations (Stackdriver) Logs, GCP Cloud Monitoring (Stackdrive) Metrics, Honeycomb, HTTP, LogScale Logs, LogScale Metrics, InfluxDB Logs, InfluxDB Metrics, Kafka, LogDNA, Loki, NATS, New Relic Logs, Papertrail, Prometheus Exporter, Prometheus Remote Write, Apache Pulsar, Sematext Logs, Sematex Metrics, Socket, Splunk HEC, StatsD, Vector Output is supported vial plugins: out_copy, out_null, out_roundrobin, out_stdout, out_exec_filter, out_forward, out_mongo / out_mongo_replset, out_exec, out_file, out_s3, out_webhdfs. Additional output formats (793 input/output plugins) can be found in the full list here: https://www.fluentd.org/plugins
ETL Functionality Yes, limited support is under development Yes, limited, however organizations typically pair Filebeat with LogStash for more complicated transformations Yes - Vector has its own language designed for transforming log events/metrics Yes - Fluentd has the ability to filter and transform events using its Filter functionality: https://docs.fluentd.org/filter
LogScale End Point Utilized LogScale Unstructured Elastic HEC Elastic / HEC (preferred)

Filebeat, Vector and Fluentd do not cover all the possible log sources and there are use cases that will require organizations to implement additional log shippers. Windows Event Logs for example are a common source that neither Filebeat or Vector currently handle |---| LogScale Log Collector and Winlogbeat are great choices.

Windows Event Logs

LogScale Log collector can be used for this specific use case. Vector and Filebeat do not have access to the Windows APIs that are required to read the Windows Event Log.

If LogScale Log collector is not suitable for your use case you can use Winlogbeat. Organizations can choose to put Winlogbeat or LogScale Log Collector on each server individually however at scale this becomes difficult to manage in terms of deploying log shippers to every Windows machine. The solution is to use the Windows Event Forwarding to forward events from each Windows machine to a central aggregation server that will then forward the events on to LogScale. While Winlogbeat is often used in this role, there are other log shippers like LogScale Log Collector and Fluentd that forward the events on to LogScale.

Kafka

For organizations using Kafka as part of their data pipeline there are multiple solutions for moving data from Kafka to LogScale. Vector and Fluentd for example can both read from Kafka and then ship that data to LogScale. Alternatively LogScale has a Kafka Connect Log Format that is also capable of sending data from Kafka topics to LogScale.

One of the primary advantages to using Kafka in front of LogScale as part of your log shipping pipeline is that Kafka can be used as a queue to help buffer events under various failure conditions (network outages, log shipper failures, etc.)

Highly Available Configurations

Most modern log shippers are designed to be able to handle common failure scenarios including:

  • Log shipper failure

    Log shippers maintain a record of the last event successfully transmitted to the target platform. When a log shipper recovers from its failure state it will refer to this record to begin sending data again.

  • Network failure/target unreachable

    Under network failure conditions, when the target source is unavailable, or when the target source is unable to keep up with the volume of logs being sent most log shippers are able to buffer outgoing logs until the condition resolves.

While it isn't possible to configure log shippers to stop network or target system failures it is possible to deploy log shippers in highly available configurations using a load balancer or queuing system that can provide the following benefits:

  • Increased uptime

    Minimize the downtime associated with log shipper failures and application upgrades.

  • Increased throughput

    Increase the number of events per second that can be shipped by combining multiple log shippers behind a firewall or queue which can be particularly useful in scenarios like when using the Windows Event Forwarding Framework.

Recommendations for how to deploy the log shippers and which load balancer or queuing system to use are use case specific. For many organizations the decision of which approach to use will be obvious but for others the answer might be a combination of both load balancing and queue based approaches. The following sections briefly describe load balancer and queue based high availability models and their pros and cons.

High Availability Using A Load Balancer

One way to achieve high availability with log shipping is to place a pool of log shipping agents behind a load balancer. In this model source systems sending logs to LogScale would send their logs to a single address (URL or IP) and the load balancer would manage the process of balancing the traffic across a series of log shippers. This model of load balancing log shippers is commonly used under for the following reasons:

  • The organization wants to minimize the number of devices sending log events through their firewall to external targets (LogScale);

  • The organization uses a framework like the Windows Event Forwarding framework that gathers event logs from Windows servers and desktops in a centralized manner before forwarding on to LogScale;

  • The organization wants to take advantage of the transformation capabilities offered by a log shipper like Fluentd or Vector without having to deploy the log shipper on every source machine (when sending syslog from firewalls or network switches for example).

For most organizations the primary reason to deploy log shippers in a highly available configuration is when those log shippers are acting as a centralized point for forwarding logs on to LogScale. Most common load balancers (HAProxy, Nginx, etc.) would support this use case. The following diagram presents a high level architectural overview of a load balanced log shipping solution:

graph TD src1[Source 1] --> lb1[Load Balancer] src2[Source 2] --> lb1 src3[Source 3] --> lb1 src4[Source 4] --> lb1 src5[Source 5] --> lb1 srcN[Source N] --> lb1 lb1 --> logS1 lb1 --> logS2 lb1 --> logS3 logS1[Log Shipper] --> lb2[Load Balancer] logS2[Log Shipper] --> lb2 logS3[Log Shipper] --> lb2 lb2 --> humio1 lb2 --> humio2 lb2 --> humioN humio1[LogScale 1] humio2[LogScale 2] humioN[LogScale N] style lb1 fill:#2ac76d; style lb2 fill:#32a852;
Example Nginx Configuration

The following sample Nginx configuration demonstrates how to load balance Syslog connections over TCP and UDP to two log shipping nodes (10.100.0.5:514, 10.100.0.6:514) from the NGINX proxy (10.100.0.26:514):

json
stream
{
   upstream syslog_udp {
      Server 10.100.0.5:514;
      Server 10.100.0.6:514;
   }

   upstream syslog_tcp {
      Server 10.100.0.5:514;
      Server 10.100.0.6:514;
   }

   server {
      Listen 10.100.0.26:514 udp;
      Proxy_pass syslog_udp;
      Proxy_responses 0;
   }

   server {
      Listen 10.100.0.26:514;
      Proxy_pass syslog_tcp;
   }
}

worker_rlimit_nofile 1000000;

events
{
   worker_connections 20000;
}
Load Balancing using a Queue

An alternative solution is to use a queue. In this high availability model all of your log producers write events to the target queue. Multiple agents (log shippers) are deployed to read from the queue. There are several possible queue platforms that you can use including Kafka, Amazon SQS, and Amazon S3. The decision of which platform to use for your queue will depend on your organization's existing infrastructure. It should be noted however that one of the important aspects of a queuing system is that there is a mechanism in place to prevent events from being read and forwarded by multiple log shippers.

Generally speaking organizations should not attempt to build ad hoc queuing systems when it is possible to use a platform like Amazon SQS or Kafka. These platforms have existing integrations with LogScale that minimize the level of effort required to implement a queuing solution.