diff --git a/pipeline/outputs/firehose.md b/pipeline/outputs/firehose.md index 3f10dfaa5..924e84ddb 100644 --- a/pipeline/outputs/firehose.md +++ b/pipeline/outputs/firehose.md @@ -4,43 +4,45 @@ description: Send logs to Amazon Kinesis Firehose # Amazon Kinesis Data Firehose -The Amazon Kinesis Data Firehose output plugin allows to ingest your records into the [Firehose](https://aws.amazon.com/kinesis/data-firehose/) service. +The _Amazon Kinesis Data Firehose_ output plugin lets you ingest your records into the [Firehose](https://aws.amazon.com/kinesis/data-firehose/) service. -This is the documentation for the core Fluent Bit Firehose plugin written in C. It can replace the [aws/amazon-kinesis-firehose-for-fluent-bit](https://github.com/aws/amazon-kinesis-firehose-for-fluent-bit) Golang Fluent Bit plugin released last year. The Golang plugin was named `firehose`; this new high performance and highly efficient firehose plugin is called `kinesis_firehose` to prevent conflicts/confusion. +This is the documentation for the core Fluent Bit Firehose plugin written in C. It can replace the [aws/amazon-kinesis-firehose-for-fluent-bit](https://github.com/aws/amazon-kinesis-firehose-for-fluent-bit) Golang Fluent Bit plugin. The Golang plugin was named `firehose`. This new Firehose plugin is called `kinesis_firehose` to prevent conflicts/confusion. -See [here](https://github.com/fluent/fluent-bit-docs/tree/43c4fe134611da471e706b0edb2f9acd7cdfdbc3/administration/aws-credentials.md) for details on how AWS credentials are fetched. +See [AWS credentials](https://docs.fluentbit.io/manual/administration/aws-credentials) for details on how AWS credentials are fetched. -## Configuration Parameters +## Configuration parameters -| Key | Description | -| :--- | :--- | -| region | The AWS region. | -| delivery\_stream | The name of the Kinesis Firehose Delivery stream that you want log records sent to. | -| time\_key | Add the timestamp to the record under this key. By default the timestamp from Fluent Bit will not be added to records sent to Kinesis. | -| time\_key\_format | strftime compliant format string for the timestamp; for example, the default is '%Y-%m-%dT%H:%M:%S'. Supports millisecond precision with '%3N' and supports nanosecond precision with '%9N' and '%L'; for example, adding '%3N' to support millisecond '%Y-%m-%dT%H:%M:%S.%3N'. This option is used with time\_key. | -| log\_key | By default, the whole log record will be sent to Firehose. If you specify a key name with this option, then only the value of that key will be sent to Firehose. For example, if you are using the Fluentd Docker log driver, you can specify `log_key log` and only the log message will be sent to Firehose. | -| compression | Compression type for Firehose records. Each log record is individually compressed and sent to Firehose. 'gzip' and 'arrow' are the supported values. 'arrow' is only an available if Apache Arrow was enabled at compile time. Defaults to no compression. | -| role\_arn | ARN of an IAM role to assume \(for cross account access\). | -| endpoint | Specify a custom endpoint for the Firehose API. | -| sts\_endpoint | Custom endpoint for the STS API. | -| auto\_retry\_requests | Immediately retry failed requests to AWS services once. This option does not affect the normal Fluent Bit retry mechanism with backoff. Instead, it enables an immediate retry with no delay for networking errors, which may help improve throughput when there are transient/random networking issues. This option defaults to `true`. | -| external\_id | Specify an external ID for the STS API, can be used with the role_arn parameter if your role requires an external ID. | -| profile | AWS profile name to use. Defaults to `default`. | -| workers | The number of [workers](../../administration/multithreading.md#outputs) to perform flush operations for this output. Default: `1`. | +This plugin uses the following configuration parameters: -## Getting Started +| Key | Description | Default | +| :--- | :--- | --------- | +| `region` | The AWS region . | _none_ | +| `delivery_stream` | The name of the Kinesis Firehose Delivery stream that you want log records sent to. | _none_ | +| `time_key` | Add the timestamp to the record under this key. By default, the timestamp from Fluent Bit won't be added to records sent to Kinesis. | _none_ | +| `time_key_format` | strftime compliant format string for the timestamp; for example, the default is `%Y-%m-%dT%H:%M:%S`. Supports millisecond precision with `%3N` and nanosecond precision with `%9N` and `%L`. For example, adding `%3N` to support millisecond `%Y-%m-%dT%H:%M:%S.%3N`. This option is used with `time_key`. | _none_ | +| `log_key` | By default, the whole log record will be sent to Firehose. If you specify a key name with this option, then only the value of that key will be sent to Firehose. For example, if you are using the Fluentd Docker log driver, you can specify `log_key log` and only the log message will be sent to Firehose. | _none_ | +| `compression` | Compression type for Firehose records. Each log record is individually compressed and sent to Firehose. Supported values: `gzip`. `arrow`. `arrow` is only an available if Apache Arrow was enabled at compile time. Defaults to no compression. | _none_ | +| `role_arn` | ARN of an IAM role to assume (for cross account access`). | _none_ | +| `endpoint` | Specify a custom endpoint for the Firehose API. | _none_ | +| `sts_endpoint` | Custom endpoint for the STS API. | _none_ | +| `auto_retry_requests` | Immediately retry failed requests to AWS services once. This option doesn't affect the normal Fluent Bit retry mechanism with backoff. Instead, it enables an immediate retry with no delay for networking errors, which can help improve throughput when there are transient/random networking issues. | `true` | +| `external_id` | Specify an external ID for the STS API. Can be used with the `role_arn` parameter if your role requires an external ID. | _none_ | +| `profile` | AWS profile name to use. | `default` | +| `workers` | The number of [workers](../../administration/multithreading.md#outputs) to perform flush operations for this output. | `1` | -In order to send records into Amazon Kinesis Data Firehose, you can run the plugin from the command line or through the configuration file: +## Get started -### Command Line +To send records into Amazon Kinesis Data Firehose, you can run the plugin from the command line or through the configuration file. -The **firehose** plugin, can read the parameters from the command line through the **-p** argument \(property\), e.g: +### Command line + +The Firehose plugin can read the parameters from the command line through the `-p` argument (property). ```shell fluent-bit -i cpu -o kinesis_firehose -p delivery_stream=my-stream -p region=us-west-2 -m '*' -f 1 ``` -### Configuration File +### Configuration file In your main configuration file append the following: @@ -49,7 +51,7 @@ In your main configuration file append the following: ```yaml pipeline: - + outputs: - name: kinesis_firehose match: '*' @@ -90,7 +92,7 @@ The following AWS IAM permissions are required to use this plugin: ### Worker support -Fluent Bit 1.7 adds a new feature called `workers` which enables outputs to have dedicated threads. This `kinesis_firehose` plugin fully supports workers. +Fluent Bit 1.7 added a new feature called `workers` which enables outputs to have dedicated threads. This `kinesis_firehose` plugin fully supports workers. Example: @@ -99,7 +101,7 @@ Example: ```yaml pipeline: - + outputs: - name: kinesis_firehose match: '*' @@ -125,7 +127,7 @@ pipeline: {% hint style="info" %} -If you enable a single worker, you are enabling a dedicated thread for your Firehose output. We recommend starting with without workers, evaluating the performance, and then adding workers one at a time until you reach your desired/needed throughput. For most users, no workers or a single worker will be sufficient. +If you enable a single worker, you are enabling a dedicated thread for your Firehose output. Fluent Bit recommends starting with without workers, evaluating the performance, and then adding workers one at a time until you reach your desired/needed throughput. For most users, no workers or a single worker will be sufficient. {% endhint %} @@ -141,7 +143,7 @@ Amazon distributes a container image with Fluent Bit and these plugins. [aws-for-fluent-bit](https://gallery.ecr.aws/aws-observability/aws-for-fluent-bit) -Our images are available in Amazon ECR Public Gallery. You can download images with different tags by following command: +Fluent Bit images are available in Amazon ECR Public Gallery. You can download images with different tags by following command: ```shell docker pull public.ecr.aws/aws-observability/aws-for-fluent-bit: @@ -167,10 +169,10 @@ You can check the [Amazon ECR Public official doc](https://docs.aws.amazon.com/A #### Amazon ECR -You can use our SSM Public Parameters to find the Amazon ECR image URI in your region: +You can use Fluent Bit SSM Public Parameters to find the Amazon ECR image URI in your region: ```shell aws ssm get-parameters-by-path --path /aws/service/aws-for-fluent-bit/ ``` -For more see [the AWS for Fluent Bit GitHub repo](https://github.com/aws/aws-for-fluent-bit#public-images). \ No newline at end of file +For more see [the AWS for Fluent Bit GitHub repository](https://github.com/aws/aws-for-fluent-bit#public-images). diff --git a/vale-styles/FluentBit/Headings.yml b/vale-styles/FluentBit/Headings.yml index bdd98e97e..2a3956016 100644 --- a/vale-styles/FluentBit/Headings.yml +++ b/vale-styles/FluentBit/Headings.yml @@ -10,6 +10,7 @@ exceptions: - Amazon - Amazon CloudWatch - Amazon ECR Public Gallery + - Amazon Kinesis Data Firehose - Amazon Kinesis Firehose - Amazon Kinesis Streams - API