Creates and manages Logging pipeline objects.
Manages a Logging pipeline.
⚠️ Only tokens are accepted for authorization in the logging_pipeline resource. Please ensure you are using tokens as other methods will not be valid.
For re-usability, an array of logs can be defined in a tfvars file or inside the terraform plan, and used as presented below:
The content inside vars.tfvars file:
The content inside the tf plan:
The configuration can then be applied using the following commands:
location
- (Optional)[string] The location of the Logging pipeline. Default: de/txl
One of de/fra
, de/txl
, gb/lhr
, es/vit
, fr/par
. If this is not set and if no value is provided for the IONOS_API_URL
env var, the default location
will be: de/fra
.
name
- (Required)[string] The name of the Logging pipeline.
grafana_address
- (Computed)[string] The address of the client's grafana instance.
log
- (Required)[list] Pipeline logs, a list that contains elements with the following structure:
source
- (Required)[string] The source parser to be used.
tag
- (Required)[string] The tag is used to distinguish different pipelines. Must be unique amongst the pipeline's array items.
protocol
- (Required)[string] "Protocol to use as intake. Possible values are: http, tcp."
public
- (Computed)[bool]
destinations
- (Optional)[list] The configuration of the logs datastore, a list that contains elements with the following structure:
type
- (Optional)[string] The internal output stream to send logs to.
retention_in_days
- (Optional)[int] Defines the number of days a log record should be kept in loki. Works with loki destination type only. Can be one of: 7, 14, 30.
In order to import a Logging pipeline, you can define an empty Logging pipeline resource in the plan:
The resource can be imported using the location
and pipeline_id
, for example: