S3 input plugin
editS3 input plugin
edit- A component of the aws integration plugin
- Integration version: v7.1.7
- Released on: 2024-07-02
- Changelog
For other versions, see the Versioned plugin docs.
Getting Help
editFor questions about the plugin, open a topic in the Discuss forums. For bugs or feature requests, open an issue in Github. For the list of Elastic supported plugins, please consult the Elastic Support Matrix.
Description
editStream events from files from a S3 bucket.
The S3 input plugin only supports AWS S3. Other S3 compatible storage solutions are not supported.
Each line from each file generates an event.
Files ending in .gz
are handled as gzip’ed files.
Files that are archived to AWS Glacier will be skipped.
Event Metadata and the Elastic Common Schema (ECS)
editThis plugin adds cloudfront metadata to event.
When ECS compatibility is disabled, the value is stored in the root level.
When ECS is enabled, the value is stored in the @metadata
where it can be used by other plugins in your pipeline.
Here’s how ECS compatibility mode affects output.
ECS disabled | ECS v1 | Availability | Description |
---|---|---|---|
cloudfront_fields |
[@metadata][s3][cloudfront][fields] |
available when the file is a CloudFront log |
column names of log |
cloudfront_version |
[@metadata][s3][cloudfront][version] |
available when the file is a CloudFront log |
version of log |
S3 Input Configuration Options
editThis plugin supports the following configuration options plus the Common Options described later.
Setting | Input type | Required |
---|---|---|
No |
||
No |
||
No |
||
No |
||
No |
||
No |
||
Yes |
||
No |
||
No |
||
No |
||
No |
||
No |
||
No |
||
No |
||
No |
||
No |
||
No |
||
No |
||
No |
||
No |
||
No |
||
No |
||
No |
||
No |
||
No |
Also see Common Options for a list of options supported by all input plugins.
access_key_id
edit- Value type is string
- There is no default value for this setting.
This plugin uses the AWS SDK and supports several ways to get credentials, which will be tried in this order:
-
Static configuration, using
access_key_id
andsecret_access_key
params in logstash plugin config -
External credentials file specified by
aws_credentials_file
-
Environment variables
AWS_ACCESS_KEY_ID
andAWS_SECRET_ACCESS_KEY
-
Environment variables
AMAZON_ACCESS_KEY_ID
andAMAZON_SECRET_ACCESS_KEY
- IAM Instance Profile (available when running inside EC2)
additional_settings
edit- Value type is hash
-
Default value is
{}
Key-value pairs of settings and corresponding values used to parametrize the connection to s3. See full list in the AWS SDK documentation. Example:
input { s3 { access_key_id => "1234" secret_access_key => "secret" bucket => "logstash-test" additional_settings => { force_path_style => true follow_redirects => false } } }
aws_credentials_file
edit- Value type is string
- There is no default value for this setting.
Path to YAML file containing a hash of AWS credentials.
This file will only be loaded if access_key_id
and
secret_access_key
aren’t set. The contents of the
file should look like this:
:access_key_id: "12345" :secret_access_key: "54321"
backup_add_prefix
edit- Value type is string
-
Default value is
nil
Append a prefix to the key (full path including file name in s3) after processing. If backing up to another (or the same) bucket, this effectively lets you choose a new folder to place the files in
backup_to_bucket
edit- Value type is string
-
Default value is
nil
Name of a S3 bucket to backup processed files to.
backup_to_dir
edit- Value type is string
-
Default value is
nil
Path of a local directory to backup processed files to.
bucket
edit- This is a required setting.
- Value type is string
- There is no default value for this setting.
The name of the S3 bucket.
delete
edit- Value type is boolean
-
Default value is
false
Whether to delete processed files from the original bucket.
ecs_compatibility
edit- Value type is string
-
Supported values are:
-
disabled
: does not use ECS-compatible field names -
v1
,v8
: uses metadata fields that are compatible with Elastic Common Schema
-
Controls this plugin’s compatibility with the Elastic Common Schema (ECS). See Event Metadata and the Elastic Common Schema (ECS) for detailed information.
endpoint
edit- Value type is string
- There is no default value for this setting.
The endpoint to connect to. By default it is constructed using the value of region
.
This is useful when connecting to S3 compatible services, but beware that these aren’t
guaranteed to work correctly with the AWS SDK.
exclude_pattern
edit- Value type is string
-
Default value is
nil
Ruby style regexp of keys to exclude from the bucket.
Note that files matching the pattern are skipped after they have been listed.
Consider using prefix
instead where possible.
Example:
"exclude_pattern" => "\/2020\/04\/"
This pattern excludes all logs containing "/2020/04/" in the path.
gzip_pattern
edit- Value type is string
-
Default value is
"\.gz(ip)?$"
Regular expression used to determine whether an input file is in gzip format.
include_object_properties
edit- Value type is boolean
-
Default value is
false
Whether or not to include the S3 object’s properties (last_modified, content_type, metadata) into each Event at
[@metadata][s3]
. Regardless of this setting, [@metadata][s3][key]
will always be present.
interval
edit- Value type is number
-
Default value is
60
Interval to wait between to check the file list again after a run is finished. Value is in seconds.
prefix
edit- Value type is string
-
Default value is
nil
If specified, the prefix of filenames in the bucket must match (not a regexp)
proxy_uri
edit- Value type is string
- There is no default value for this setting.
URI to proxy server if required
role_arn
edit- Value type is string
- There is no default value for this setting.
The AWS IAM Role to assume, if any. This is used to generate temporary credentials, typically for cross-account access. See the AssumeRole API documentation for more information.
role_session_name
edit- Value type is string
-
Default value is
"logstash"
Session name to use when assuming an IAM role.
secret_access_key
edit- Value type is string
- There is no default value for this setting.
The AWS Secret Access Key
session_token
edit- Value type is string
- There is no default value for this setting.
The AWS Session token for temporary credential
sincedb_path
edit- Value type is string
-
Default value is
nil
Where to write the since database (keeps track of the date the last handled file was added to S3). The default will write sincedb files to in the directory {path.data}/plugins/inputs/s3/
If specified, this setting must be a filename path and not just a directory.
temporary_directory
edit- Value type is string
-
Default value is
"/tmp/logstash"
Set the directory where logstash will store the tmp files before processing them.
Common Options
editThe following configuration options are supported by all input plugins:
Details
edit
enable_metric
edit- Value type is boolean
-
Default value is
true
Disable or enable metric logging for this specific plugin instance by default we record all the metrics we can, but you can disable metrics collection for a specific plugin.
id
edit- Value type is string
- There is no default value for this setting.
Add a unique ID
to the plugin configuration. If no ID is specified, Logstash will generate one.
It is strongly recommended to set this ID in your configuration. This is particularly useful
when you have two or more plugins of the same type, for example, if you have 2 s3 inputs.
Adding a named ID in this case will help in monitoring Logstash when using the monitoring APIs.
input { s3 { id => "my_plugin_id" } }
Variable substitution in the id
field only supports environment variables
and does not support the use of values from the secret store.
tags
edit- Value type is array
- There is no default value for this setting.
Add any number of arbitrary tags to your event.
This can help with processing later.
type
edit- Value type is string
- There is no default value for this setting.
Add a type
field to all events handled by this input.
Types are used mainly for filter activation.
The type is stored as part of the event itself, so you can also use the type to search for it in Kibana.
If you try to set a type on an event that already has one (for example when you send an event from a shipper to an indexer) then a new input will not override the existing type. A type set at the shipper stays with that event for its life even when sent to another Logstash server.