You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+45-15Lines changed: 45 additions & 15 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ A Kafka Connect Sink for Splunk features:
8
8
## Requirements
9
9
1. Kafka version 0.10 and above.
10
10
2. Java 8 and above.
11
-
3. A Splunk environment of version 6.4* and above, configured with valid HTTP Event Collector (HEC) tokens. (A minimum Splunk version of 6.5 is required for event annotation)
11
+
3. A Splunk environment of version 6.5 and above, configured with valid HTTP Event Collector (HEC) tokens.
12
12
13
13
* HEC token settings should be the same on all Splunk Indexers and Heavy Forwarders in your environment.
14
14
* Task configuration parameters will vary depending on acknowledgement setting (See the [Configuration](#configuration) section for details).
@@ -25,7 +25,7 @@ Note: The resulting "kafka-connect-splunk-*.tar.gz" package is self-contained. B
25
25
26
26
## Quick Start
27
27
28
-
1.[Start](https://kafka.apache.org/quickstart) your Kafka Cluster and Zookeeper on your local host. Confirm both are running.
28
+
1.[Start](https://kafka.apache.org/quickstart) your Kafka Cluster and confirm it is running.
29
29
2. If this is a new install, create a test topic (eg: `perf`). Inject events into the topic. This can be done using [Kafka data-gen-app](https://github.com/dtregonning/kafka-data-gen) or the Kafka bundle [kafka-console-producer](https://kafka.apache.org/quickstart#quickstart_send).
30
30
3. Untar the package created from the build script: `tar xzvf kafka-connect-splunk-*.tar.gz` (Default target location is /tmp/kafka-connect-splunk-build/kafka-connect-splunk).
31
31
4. Navigate to kafka-connect-splunk directory `cd kafka-connect-splunk`.
@@ -57,7 +57,7 @@ Note: The resulting "kafka-connect-splunk-*.tar.gz" package is self-contained. B
57
57
"splunk.hec.total.channels": "8",
58
58
"splunk.hec.max.batch.size": "1000000",
59
59
"splunk.hec.threads": "2",
60
-
"splunk.hec.event.timeout": "60",
60
+
"splunk.hec.event.timeout": "300",
61
61
"splunk.hec.socket.timeout": "120",
62
62
"splunk.hec.track.data": "true"
63
63
}
@@ -95,7 +95,7 @@ Use the following connector deployment options:
95
95
* Splunk Kafka Connector in a dedicated Kafka Connect Cluster (recommended)
96
96
* Splunk Kafka Connector in an existing Kafka Connect Cluster
97
97
98
-
### Connector in a dedicated Kafka Connect Cluster
98
+
### Connector in a dedicated Kafka Connect Cluster
99
99
Running the Splunk Kafka Connector in a dedicated Kafka Connect Cluster is recommended. Isolating the Splunk connector from other Kafka connectors results in significant performance benefits in high throughput environments.
100
100
101
101
1. Untar the **kafka-connect-splunk-*.tar.gz** package and navigate to the **kafka-connect-splunk** directory.
@@ -143,16 +143,46 @@ Running the Splunk Kafka Connector in a dedicated Kafka Connect Cluster is recom
143
143
> Note: The **KAFKA\_HEAP\_OPTS** environment variable controls how much memory Kafka Connect can use. Set the **KAFKA\_HEAP\_OPTS** with the recommended value stated in the example above.
144
144
145
145
### Connector in an existing Kafka Connect Cluster
146
-
1. Untar the **kafka-connect-splunk-*.tar.gz** installation package and go to the **kafka-connect-splunk** directory.
147
146
148
-
```
149
-
tar xzvf kafka-connect-splunk-*.tar.gz
150
-
cd kafka-connect-splunk
151
-
```
147
+
1. Navigate to Splunkbase and download the latest version of [Splunk Kafka Connect](https://splunkbase.splunk.com/app/3862/)
148
+
149
+
2. Copy downloaded file onto every host into the directory that contains your other connectors or create a folder to store them in. (ex. `/opt/connectors/splunk-kafka-connect`)
150
+
151
+
3. Create a properties file called `kafka-connect.properties.` File should be created in directory `$KAFKA_CONNECT_HOME/config/`.
152
+
Copy the following contents into the file and modify the <BOOTSTRAP_SERVERS> to point to one of your kafka brokers (ex. `localhost:9092`): and
153
+
modify <PLUGIN_PATH> to point to the top level directory of where you are storing your connectors. (ex. `/opt/connectors`)
154
+
155
+
> Note: - If running Kafka Version 0.10.x - PLUGIN_PATH is not a valid configuration property. To make the connector visible to
156
+
Kafka Connect the connectors folder must be added to the classpath. (ex. export `CLASSPATH=/opt/connectors/*`)
152
157
153
-
2. Copy the **conectors/kafka-connect-splunk-*.jar** to the plugin path specified by **plugin.path** in the existing Kafka Connect on every host.
154
-
3. Copy **libs/commons-logging-1.2.jar** to **libs** of the existing Kafka Connect on each host.
> Note - For more information on the worker paramaters please refer to Kafka Connect [documentation](https://kafka.apache.org/documentation/#connect_running).
184
+
185
+
4. Run `$KAFKA_CONNECT_HOME/bin/connect-distributed.sh $KAFKA_CONNECT_HOME/config/kafka-connect.properties` to start Kafka Connect or restart Kafka Connect with existing configuration file.
156
186
157
187
## Security
158
188
The Kafka Connect Splunk Sink supports the following security mechanisms
@@ -399,7 +429,7 @@ Use the below schema to configure Splunk Kafka Connector
399
429
* `splunk.hec.ack.poll.interval` - This setting is only applicable when `splunk.hec.ack.enabled` is set to `true`. Internally it controls the event ACKs polling interval. By default, this setting is 10 seconds.
400
430
* `splunk.hec.ack.poll.threads` - This setting is used for performance tuning and is only applicable when `splunk.hec.ack.enabled` is set to `true`. It controls how many threads should be spawned to poll event ACKs. By default, it is set to `1`.
401
431
> Note: For large Splunk indexer clusters (For example, 100 indexers) you need to increase this number. Recommended increase to speed up ACK polling is 4 threads.
402
-
* `splunk.hec.event.timeout` - This setting is applicable when `splunk.hec.ack.enabled` is set to `true`. When events are POSTed to Splunk and before they are ACKed, this setting determines how long the connector will wait before timing out and resending. By default, it is set to 120 seconds.
432
+
* `splunk.hec.event.timeout` - This setting is applicable when `splunk.hec.ack.enabled` is set to `true`. When events are POSTed to Splunk and before they are ACKed, this setting determines how long the connector will wait before timing out and resending. By default, it is set to 300 seconds.
403
433
404
434
#### Endpoint Parameters
405
435
* `splunk.hec.raw` - Set to `true` in order for Splunk software to ingest data using the the /raw HEC endpoint. Default is `false`, which will use the /event endpoint.
@@ -433,7 +463,7 @@ Use the below schema to configure Splunk Kafka Connector
433
463
"splunk.hec.ack.enabled : "true",
434
464
"splunk.hec.ack.poll.interval" : "20",
435
465
"splunk.hec.ack.poll.threads" : "2",
436
-
"splunk.hec.event.timeout" : "120",
466
+
"splunk.hec.event.timeout" : "300",
437
467
"splunk.hec.raw" : "true",
438
468
"splunk.hec.raw.line.breaker" : "#####"
439
469
}
@@ -454,7 +484,7 @@ Use the below schema to configure Splunk Kafka Connector
0 commit comments