Amazon Bucket Storage

LogScale supports writing a copy of the ingested logs to Amazon S3 and Google Cloud Storage using the native file format of LogScale, allowing LogScale to fetch those files and search them efficiently if the local copies are lost or deleted. This page will explain how to set up bucket storage with Amazon S3. For more details on this topic in general, see the Bucket Storage page.

If you use S3 for bucket storage, server-side encryption must be turned off. Otherwise, LogScale cannot verify the checksum and upload will fail. The files are encrypted by LogScale, as described under Security section from the overview page.

Self-Hosting with S3

LogScale supports multiple ways of configuring access to the bucket, allowing you to use any of the 5 listed in Using the Default Credential Provider Chain on top of the following, which takes precedence if set.

If you run your LogScale installation outside AWS, you need an IAM user with write access to the buckets used for storage. That user must have programmatic access to S3, so when adding a new user through the AWS console make sure programmatic access is set.

Self-Hosting with S3

Figure 280. Self-Hosting with S3


Later in the process, you can retrieve the access key and secret key:

Self-Hosting with S3

Figure 281. Self-Hosting with S3


Which in LogScale is needed in the following configuration:

ini
# These two take precedence over all other AWS access methods.
S3_STORAGE_ACCESSKEY=$ACCESS_KEY
S3_STORAGE_SECRETKEY=$SECRET_KEY

# Also supported:
AWS_ACCESS_KEY_ID=$ACCESS_KEY
AWS_SECRET_ACCESS_KEY=$SECRET_KEY

The keys are used for authenticating the user against the S3 service. For more guidance on how to retrieve S3 access keys, see AWS access keys. More details on creating a new user in IAM.

Configuring the user to have write access to a bucket can be done by attaching the right access policy to the user.

IAM User Example Policy

The following JSON is an example policy configuration:

json
{
 "Version": "2012-10-17",
 "Statement": [
     {
         "Effect": "Allow",
         "Action": [
             "s3:ListBucket"
         ],
         "Resource": [
             "arn:aws:s3:::BUCKET_NAME"
         ]
     },
     {
         "Effect": "Allow",
         "Action": [
             "s3:PutObject",
             "s3:GetObject",
     "s3:DeleteObject"
         ],
         "Resource": [
             "arn:aws:s3:::BUCKET_NAME/*"
         ]
     }
 ]
}

The policy can be used as an inline policy attached directly to the user through the AWS console:

IAM User Example Policy

Figure 282. IAM User Example Policy


You must also tell which bucket to use. These settings must be identical on all nodes across the entire cluster.

ini
S3_STORAGE_BUCKET=$BUCKET_NAME
S3_STORAGE_REGION=$BUCKET_REGION
S3_STORAGE_ENCRYPTION_KEY=$ENCRYPTION_SECRET
S3_STORAGE_OBJECT_KEY_PREFIX=/basefolder
USING_EPHEMERAL_DISKS=true

The first option here is to set the name of the bucket to use. The encryption key given with S3_STORAGE_ENCRYPTION_KEY can be any UTF-8 string. The suggested value is 64 or more random ASCII characters. The S3_STORAGE_OBJECT_KEY_PREFIX is used to set the optional prefix for all object keys. This option allows multiple LogScale clusters to share a bucket, but requires them each to write to a unique prefix. It's empty by default. Note, there is a performance penalty when using a non-empty prefix. We recommend an unset prefix. If there are any ephemeral disks in the cluster, you must set the last option here to true.

You can change the settings using the S3_STORAGE_BUCKET to point to a fresh bucket at any point in time. From that point, LogScale will write new files to that bucket while still reading from any previously-configured buckets. Existing files already written to any previous bucket will not get written to the new bucket. LogScale will continue to delete files from the old buckets that match the file names that LogScale would put there.

Switching to a Fresh Bucket

You can change the settings using the S3_STORAGE_BUCKET and S3_STORAGE_REGION to point to a fresh bucket at any point in time. From that point, LogScale will write new files to that bucket while still reading from any previously-configured buckets. Existing files already written to any previous bucket will not get written to the new bucket. LogScale will continue to delete files from the old buckets that match the file names that LogScale would put there.

Other Options

Configuring for Use With Non-Default Endpoints

You can point to your own hosting endpoint for the S3 to use for bucket storage if you host an S3-compatible service.

There are two styles of S3 base URL LogScale can use, depending on which URLs your service supports.

Virtual host style (default)

LogScale will construct virtual host-style URLs like:

https://my-bucket.my-own-s3:8080/path/inside/bucket/file.txt

For this style of access, you need to set your base URL, so it contains a placeholder for the bucket name.

ini
S3_STORAGE_ENDPOINT_BASE=http://{bucket}.my-own-s3:8080

LogScale will replace the placeholder, {bucket} with the relevant bucket name at runtime.

Path-style

Some services do not support virtual host style access, and require path-style access.

Such URLs have the format:

uri
https://my-own-s3:8080/my-bucket/path/inside/bucket/file.txt

If you are using such a service, your endpoint base URL should not contain a bucket placeholder.

ini
S3_STORAGE_ENDPOINT_BASE=http://my-own-s3:8080

Additionally, you must set S3_STORAGE_PATH_STYLE_ACCESS to true

Hitachi Content Platform compatibility

Bucket storage can be used with the Hitachi Content Platform by setting S3_STORAGE_HCP_COMPAT to true.

HTTP Proxy

If LogScale is set up to use an HTTP proxy, it will per default be used for communicating with S3. It can be disabled using the following:

ini
# Use the globally configured HTTP proxy for communicating with S3.
# Default is true.
S3_STORAGE_USE_HTTP_PROXY=false
Amazon Bucket Parameters

The following allows tuning for performance. Note that there may be a costs associated with increasing these as S3 is billed also based on the number of operations executed.

ini
# How many parallel chunks to split each file into when uploading and downloading.
# Defaults to 4, the maximum is 16.
S3_STORAGE_CHUNK_COUNT=4

# Maximum number of files that LogScale will run concurrent downloads for at a time.
# Default is the number of hyperthreads / 2
S3_STORAGE_DOWNLOAD_CONCURRENCY=8

# Maximum number of files that LogScale will run concurrent uploads for at a time.
# Default is the number of hyperthreads / 2
# S3_STORAGE_UPLOAD_CONCURRENCY=8

# Chunk size for uploads and download ranges. Max 8 MB, which is the default.
# Minimum value is 5 MB.
S3_STORAGE_CHUNK_SIZE=8388608

# Prefer to fetch data files from the bucket when possible even if another
# node in the LogScale cluster also has a copy of the file.
# In some environments, it may be less expensive to do the transfers this way.
# The transfer from the bucket may be billed at a lower cost than a transfer from
# a node in another region or in another data center.  This preference does not
# guarantee that the bucket copy will be used, as the cluster can
# still make internal replicas directly in case the file is not yet in
# a bucket.
# Default is false.
S3_STORAGE_PREFERRED_COPY_SOURCE=false

Export to Bucket with Amazon

By default LogScale allows downloading the results of a query to a file. This file is generated as a HTTP stream directly from LogScale, and can be long-lasting with long periods of no data being transmitted when LogScale is searching for rare hits in large data sets. This can cause issues for some networks and load balancers.

As an alternative, LogScale allows exporting to Amazon S3. The result of the query will be uploaded to the bucket storage provider and the user will be given a URL to download the file once the upload is complete.

As LogScale uses signed URLs for downloads the user does not need read access to the bucket, LogScale needs write access to the target bucket.

Following is the most basic example configuration for exporting to Amazon S3:

ini
# The credentials used to authenticate with
S3_EXPORT_ACCESSKEY=$ACCESS_KEY
S3_EXPORT_SECRETKEY=$SECRET_KEY

# The name of the region and bucket to export to
S3_EXPORT_REGION=$BUCKET_REGION
S3_EXPORT_BUCKET=$BUCKET_NAME

LogScale supports many other kinds of authentication on AWS, as described earlier on this page.

HTTP Proxy

If LogScale is set up to use an HTTP proxy, it will per default be used for communicating with S3. It can be disabled using the following:

ini
# Use the globally configured HTTP proxy for communicating with S3.
# Default is true.
S3_EXPORT_USE_HTTP_PROXY=false