http_poller
edithttp_poller
editThis Logstash input plugin allows you to call an HTTP API, decode the output of it into event(s), and send them on their merry way. The idea behind this plugins came from a need to read springboot metrics endpoint, instead of configuring jmx to monitor my java application memory/gc/ etc.
Example
editReads from a list of urls and decodes the body of the response with a codec. The config should look like this:
input { http_poller { urls => { test1 => "http://localhost:9200" test2 => { # Supports all options supported by ruby's Manticore HTTP client method => get url => "http://localhost:9200/_cluster/health" headers => { Accept => "application/json" } auth => { user => "AzureDiamond" password => "hunter2" } } } request_timeout => 60 interval => 60 codec => "json" # A hash of request metadata info (timing, response headers, etc.) will be sent here metadata_target => "http_poller_metadata" } } output { stdout { codec => rubydebug } }
Using the HTTP poller with custom a custom CA or self signed cert.
If you have a self signed cert you will need to convert your server’s certificate to a valid# .jks
or .p12
file. An easy way to do it is to run the following one-liner, substituting your server’s URL for the placeholder MYURL
and MYPORT
.
openssl s_client -showcerts -connect MYURL:MYPORT </dev/null 2>/dev/null|openssl x509 -outform PEM > downloaded_cert.pem; keytool -import -alias test -file downloaded_cert.pem -keystore downloaded_truststore.jks
The above snippet will create two files downloaded_cert.pem
and downloaded_truststore.jks
. You will be prompted to set a password for the jks
file during this process. To configure logstash use a config like the one that follows.
http_poller { urls => { myurl => "https://myhostname:1234" } truststore => "/path/to/downloaded_truststore.jks" truststore_password => "mypassword" interval => 30 }
Synopsis
editThis plugin supports the following configuration options:
Required configuration options:
http_poller { interval => ... urls => ... }
Available configuration options:
Setting | Input type | Required | Default value |
---|---|---|---|
No |
|
||
No |
|
||
a valid filesystem path |
No |
||
a valid filesystem path |
No |
||
a valid filesystem path |
No |
||
No |
|
||
No |
|
||
No |
|
||
No |
|
||
Yes |
|||
No |
|
||
a valid filesystem path |
No |
||
No |
|||
No |
|
||
No |
|
||
No |
|
||
No |
|
||
<<,>> |
No |
||
No |
|
||
No |
|
||
No |
|
||
No |
|
||
No |
|||
No |
|||
a valid filesystem path |
No |
||
No |
|||
No |
|
||
No |
|||
Yes |
|||
No |
|
Details
edit
codec
edit- Value type is codec
-
Default value is
"json"
The codec used for input data. Input codecs are a convenient method for decoding your data before it enters the input, without needing a separate filter in your Logstash pipeline.
interval
edit- This is a required setting.
- Value type is number
- There is no default value for this setting.
How often (in seconds) the urls will be called
metadata_target
edit- Value type is string
-
Default value is
"@metadata"
If you’d like to work with the request/response metadata. Set this value to the name of the field you’d like to store a nested hash of metadata.
proxy
edit<li> Value type is <<string,string>> * There is no default value for this setting.
tags
edit- Value type is array
- There is no default value for this setting.
Add any number of arbitrary tags to your event.
This can help with processing later.
target
edit- Value type is string
- There is no default value for this setting.
Define the target field for placing the received data. If this setting is omitted, the data will be stored at the root (top level) of the event.
type
edit- Value type is string
- There is no default value for this setting.
Add a type
field to all events handled by this input.
Types are used mainly for filter activation.
The type is stored as part of the event itself, so you can also use the type to search for it in Kibana.
If you try to set a type on an event that already has one (for example when you send an event from a shipper to an indexer) then a new input will not override the existing type. A type set at the shipper stays with that event for its life even when sent to another Logstash server.