diff --git a/artifacts/logdna-splunk-integration/connect-distrubuted.properties b/artifacts/logdna-splunk-integration/connect-distrubuted.properties index 8b137891..ba32797b 100644 --- a/artifacts/logdna-splunk-integration/connect-distrubuted.properties +++ b/artifacts/logdna-splunk-integration/connect-distrubuted.properties @@ -1 +1,64 @@ +## +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +## +# This file contains some of the configurations for the Kafka Connect distributed worker. This file is intended +# to be used with the examples, and some settings may differ from those used in a production system, especially +# the `bootstrap.servers` and those specifying replication factors. + +# A list of host/port pairs to use for establishing the initial connection to the Kafka cluster. +bootstrap.servers= +security.protocol=SASL_SSL +sasl.mechanism=PLAIN +sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule required username="token" password=""; + +consumer.security.protocol=SASL_SSL +consumer.sasl.mechanism=PLAIN +consumer.sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule required username="token" password=""; + +producer.security.protocol=SASL_SSL +producer.sasl.mechanism=PLAIN +producer.sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule required username="token" password=""; + +plugin.path=/Users/bkadambi/Documents/software/kafka/connectors + + +# The converters specify the format of data in Kafka and how to translate it into Connect data. Every Connect user will +# need to configure these based on the format they want their data in when loaded from or stored into Kafka +# key.converter=org.apache.kafka.connect.json.JsonConverter +# value.converter=org.apache.kafka.connect.json.JsonConverter +key.converter=org.apache.kafka.connect.storage.StringConverter +value.converter=org.apache.kafka.connect.storage.StringConverter +# Converter-specific settings can be passed in by prefixing the Converter's setting with the converter we want to apply +# it to +key.converter.schemas.enable=false +value.converter.schemas.enable=false + +group.id=kafka-connect-splunk-hec-sink +config.storage.topic=__kafka-connect-splunk-task-configs +config.storage.replication.factor=3 + +offset.storage.topic=__kafka-connect-splunk-offsets +offset.storage.replication.factor=3 +offset.storage.partitions=25 + +status.storage.topic=__kafka-connect-splunk-statuses +status.storage.replication.factor=3 +status.storage.partitions=5 + +# Flush much faster than normal, which is useful for testing/debugging +offset.flush.interval.ms=10000 +rest.advertised.host.name=localhost