| from __future__ import absolute_import |
|
|
| import atexit |
| import logging |
| import os |
| import os.path |
| import socket |
| import subprocess |
| import time |
| import uuid |
|
|
| import py |
| from kafka.vendor.six.moves import urllib, range |
| from kafka.vendor.six.moves.urllib.parse import urlparse |
|
|
| from kafka import errors, KafkaAdminClient, KafkaClient, KafkaConsumer, KafkaProducer |
| from kafka.errors import InvalidReplicationFactorError |
| from kafka.protocol.admin import CreateTopicsRequest |
| from kafka.protocol.metadata import MetadataRequest |
| from test.testutil import env_kafka_version, random_string |
| from test.service import ExternalService, SpawnedService |
|
|
| log = logging.getLogger(__name__) |
|
|
|
|
| def get_open_port(): |
| sock = socket.socket() |
| sock.bind(("", 0)) |
| port = sock.getsockname()[1] |
| sock.close() |
| return port |
|
|
|
|
| def gen_ssl_resources(directory): |
| os.system(""" |
| cd {0} |
| echo Generating SSL resources in {0} |
| |
| # Step 1 |
| keytool -keystore kafka.server.keystore.jks -alias localhost -validity 1 \ |
| -genkey -storepass foobar -keypass foobar \ |
| -dname "CN=localhost, OU=kafka-python, O=kafka-python, L=SF, ST=CA, C=US" \ |
| -ext SAN=dns:localhost |
| |
| # Step 2 |
| openssl genrsa -out ca-key 2048 |
| openssl req -new -x509 -key ca-key -out ca-cert -days 1 \ |
| -subj "/C=US/ST=CA/O=MyOrg, Inc./CN=mydomain.com" |
| keytool -keystore kafka.server.truststore.jks -alias CARoot -import \ |
| -file ca-cert -storepass foobar -noprompt |
| |
| # Step 3 |
| keytool -keystore kafka.server.keystore.jks -alias localhost -certreq \ |
| -file cert-file -storepass foobar |
| openssl x509 -req -CA ca-cert -CAkey ca-key -in cert-file -out cert-signed \ |
| -days 1 -CAcreateserial -passin pass:foobar |
| keytool -keystore kafka.server.keystore.jks -alias CARoot -import \ |
| -file ca-cert -storepass foobar -noprompt |
| keytool -keystore kafka.server.keystore.jks -alias localhost -import \ |
| -file cert-signed -storepass foobar -noprompt |
| """.format(directory)) |
|
|
|
|
| class Fixture(object): |
| kafka_version = os.environ.get('KAFKA_VERSION', '0.11.0.2') |
| scala_version = os.environ.get("SCALA_VERSION", '2.8.0') |
| project_root = os.environ.get('PROJECT_ROOT', |
| os.path.abspath(os.path.join(os.path.dirname(__file__), ".."))) |
| kafka_root = os.environ.get("KAFKA_ROOT", |
| os.path.join(project_root, 'servers', kafka_version, "kafka-bin")) |
|
|
| def __init__(self): |
| self.child = None |
|
|
| @classmethod |
| def download_official_distribution(cls, |
| kafka_version=None, |
| scala_version=None, |
| output_dir=None): |
| if not kafka_version: |
| kafka_version = cls.kafka_version |
| if not scala_version: |
| scala_version = cls.scala_version |
| if not output_dir: |
| output_dir = os.path.join(cls.project_root, 'servers', 'dist') |
|
|
| distfile = 'kafka_%s-%s' % (scala_version, kafka_version,) |
| url_base = 'https://archive.apache.org/dist/kafka/%s/' % (kafka_version,) |
| output_file = os.path.join(output_dir, distfile + '.tgz') |
|
|
| if os.path.isfile(output_file): |
| log.info("Found file already on disk: %s", output_file) |
| return output_file |
|
|
| |
| try: |
| url = url_base + distfile + '.tgz' |
| log.info("Attempting to download %s", url) |
| response = urllib.request.urlopen(url) |
| except urllib.error.HTTPError: |
| log.exception("HTTP Error") |
| url = url_base + distfile + '.tar.gz' |
| log.info("Attempting to download %s", url) |
| response = urllib.request.urlopen(url) |
|
|
| log.info("Saving distribution file to %s", output_file) |
| with open(output_file, 'w') as output_file_fd: |
| output_file_fd.write(response.read()) |
|
|
| return output_file |
|
|
| @classmethod |
| def test_resource(cls, filename): |
| return os.path.join(cls.project_root, "servers", cls.kafka_version, "resources", filename) |
|
|
| @classmethod |
| def kafka_run_class_args(cls, *args): |
| result = [os.path.join(cls.kafka_root, 'bin', 'kafka-run-class.sh')] |
| result.extend([str(arg) for arg in args]) |
| return result |
|
|
| def kafka_run_class_env(self): |
| env = os.environ.copy() |
| env['KAFKA_LOG4J_OPTS'] = "-Dlog4j.configuration=file:%s" % \ |
| (self.test_resource("log4j.properties"),) |
| return env |
|
|
| @classmethod |
| def render_template(cls, source_file, target_file, binding): |
| log.info('Rendering %s from template %s', target_file.strpath, source_file) |
| with open(source_file, "r") as handle: |
| template = handle.read() |
| assert len(template) > 0, 'Empty template %s' % (source_file,) |
| with open(target_file.strpath, "w") as handle: |
| handle.write(template.format(**binding)) |
| handle.flush() |
| os.fsync(handle) |
|
|
| |
| |
| dirfd = os.open(os.path.dirname(target_file.strpath), os.O_DIRECTORY) |
| os.fsync(dirfd) |
| os.close(dirfd) |
| log.debug("Template string:") |
| for line in template.splitlines(): |
| log.debug(' ' + line.strip()) |
| log.debug("Rendered template:") |
| with open(target_file.strpath, 'r') as o: |
| for line in o: |
| log.debug(' ' + line.strip()) |
| log.debug("binding:") |
| for key, value in binding.items(): |
| log.debug(" {key}={value}".format(key=key, value=value)) |
|
|
| def dump_logs(self): |
| self.child.dump_logs() |
|
|
|
|
| class ZookeeperFixture(Fixture): |
| @classmethod |
| def instance(cls): |
| if "ZOOKEEPER_URI" in os.environ: |
| parse = urlparse(os.environ["ZOOKEEPER_URI"]) |
| (host, port) = (parse.hostname, parse.port) |
| fixture = ExternalService(host, port) |
| else: |
| (host, port) = ("127.0.0.1", None) |
| fixture = cls(host, port) |
|
|
| fixture.open() |
| return fixture |
|
|
| def __init__(self, host, port, tmp_dir=None): |
| super(ZookeeperFixture, self).__init__() |
| self.host = host |
| self.port = port |
|
|
| self.tmp_dir = tmp_dir |
|
|
| def kafka_run_class_env(self): |
| env = super(ZookeeperFixture, self).kafka_run_class_env() |
| env['LOG_DIR'] = self.tmp_dir.join('logs').strpath |
| return env |
|
|
| def out(self, message): |
| log.info("*** Zookeeper [%s:%s]: %s", self.host, self.port or '(auto)', message) |
|
|
| def open(self): |
| if self.tmp_dir is None: |
| self.tmp_dir = py.path.local.mkdtemp() |
| self.tmp_dir.ensure(dir=True) |
|
|
| self.out("Running local instance...") |
| log.info(" host = %s", self.host) |
| log.info(" port = %s", self.port or '(auto)') |
| log.info(" tmp_dir = %s", self.tmp_dir.strpath) |
|
|
| |
| template = self.test_resource("zookeeper.properties") |
| properties = self.tmp_dir.join("zookeeper.properties") |
| args = self.kafka_run_class_args("org.apache.zookeeper.server.quorum.QuorumPeerMain", |
| properties.strpath) |
| env = self.kafka_run_class_env() |
|
|
| |
| timeout = 5 |
| max_timeout = 120 |
| backoff = 1 |
| end_at = time.time() + max_timeout |
| tries = 1 |
| auto_port = (self.port is None) |
| while time.time() < end_at: |
| if auto_port: |
| self.port = get_open_port() |
| self.out('Attempting to start on port %d (try #%d)' % (self.port, tries)) |
| self.render_template(template, properties, vars(self)) |
| self.child = SpawnedService(args, env) |
| self.child.start() |
| timeout = min(timeout, max(end_at - time.time(), 0)) |
| if self.child.wait_for(r"binding to port", timeout=timeout): |
| break |
| self.child.dump_logs() |
| self.child.stop() |
| timeout *= 2 |
| time.sleep(backoff) |
| tries += 1 |
| backoff += 1 |
| else: |
| raise RuntimeError('Failed to start Zookeeper before max_timeout') |
| self.out("Done!") |
| atexit.register(self.close) |
|
|
| def close(self): |
| if self.child is None: |
| return |
| self.out("Stopping...") |
| self.child.stop() |
| self.child = None |
| self.out("Done!") |
| self.tmp_dir.remove() |
|
|
| def __del__(self): |
| self.close() |
|
|
|
|
| class KafkaFixture(Fixture): |
| broker_user = 'alice' |
| broker_password = 'alice-secret' |
|
|
| @classmethod |
| def instance(cls, broker_id, zookeeper, zk_chroot=None, |
| host=None, port=None, |
| transport='PLAINTEXT', replicas=1, partitions=2, |
| sasl_mechanism=None, auto_create_topic=True, tmp_dir=None): |
|
|
| if zk_chroot is None: |
| zk_chroot = "kafka-python_" + str(uuid.uuid4()).replace("-", "_") |
| if "KAFKA_URI" in os.environ: |
| parse = urlparse(os.environ["KAFKA_URI"]) |
| (host, port) = (parse.hostname, parse.port) |
| fixture = ExternalService(host, port) |
| else: |
| if host is None: |
| host = "localhost" |
| fixture = KafkaFixture(host, port, broker_id, |
| zookeeper, zk_chroot, |
| transport=transport, |
| replicas=replicas, partitions=partitions, |
| sasl_mechanism=sasl_mechanism, |
| auto_create_topic=auto_create_topic, |
| tmp_dir=tmp_dir) |
|
|
| fixture.open() |
| return fixture |
|
|
| def __init__(self, host, port, broker_id, zookeeper, zk_chroot, |
| replicas=1, partitions=2, transport='PLAINTEXT', |
| sasl_mechanism=None, auto_create_topic=True, |
| tmp_dir=None): |
| super(KafkaFixture, self).__init__() |
|
|
| self.host = host |
| self.port = port |
|
|
| self.broker_id = broker_id |
| self.auto_create_topic = auto_create_topic |
| self.transport = transport.upper() |
| if sasl_mechanism is not None: |
| self.sasl_mechanism = sasl_mechanism.upper() |
| else: |
| self.sasl_mechanism = None |
| self.ssl_dir = self.test_resource('ssl') |
|
|
| |
| |
| |
| self.start_pattern = r"\[Kafka ?Server (id=)?%d\],? started" % (broker_id,) |
| |
| self.scram_pattern = r"Removing Produce quota for user %s" % (self.broker_user) |
|
|
| self.zookeeper = zookeeper |
| self.zk_chroot = zk_chroot |
| |
| self.zk_host = self.zookeeper.host |
| self.zk_port = self.zookeeper.port |
|
|
| self.replicas = replicas |
| self.partitions = partitions |
|
|
| self.tmp_dir = tmp_dir |
| self.running = False |
|
|
| self._client = None |
| self.sasl_config = '' |
| self.jaas_config = '' |
|
|
| def _sasl_config(self): |
| if not self.sasl_enabled: |
| return '' |
|
|
| sasl_config = ( |
| 'sasl.enabled.mechanisms={mechanism}\n' |
| 'sasl.mechanism.inter.broker.protocol={mechanism}\n' |
| ) |
| return sasl_config.format(mechanism=self.sasl_mechanism) |
|
|
| def _jaas_config(self): |
| if not self.sasl_enabled: |
| return '' |
|
|
| elif self.sasl_mechanism == 'PLAIN': |
| jaas_config = ( |
| 'org.apache.kafka.common.security.plain.PlainLoginModule required\n' |
| ' username="{user}" password="{password}" user_{user}="{password}";\n' |
| ) |
| elif self.sasl_mechanism in ("SCRAM-SHA-256", "SCRAM-SHA-512"): |
| jaas_config = ( |
| 'org.apache.kafka.common.security.scram.ScramLoginModule required\n' |
| ' username="{user}" password="{password}";\n' |
| ) |
| else: |
| raise ValueError("SASL mechanism {} currently not supported".format(self.sasl_mechanism)) |
| return jaas_config.format(user=self.broker_user, password=self.broker_password) |
|
|
| def _add_scram_user(self): |
| self.out("Adding SCRAM credentials for user {} to zookeeper.".format(self.broker_user)) |
| args = self.kafka_run_class_args( |
| "kafka.admin.ConfigCommand", |
| "--zookeeper", |
| "%s:%d/%s" % (self.zookeeper.host, |
| self.zookeeper.port, |
| self.zk_chroot), |
| "--alter", |
| "--entity-type", "users", |
| "--entity-name", self.broker_user, |
| "--add-config", |
| "{}=[password={}]".format(self.sasl_mechanism, self.broker_password), |
| ) |
| env = self.kafka_run_class_env() |
| proc = subprocess.Popen(args, env=env, stdout=subprocess.PIPE, stderr=subprocess.PIPE) |
|
|
| stdout, stderr = proc.communicate() |
|
|
| if proc.returncode != 0: |
| self.out("Failed to save credentials to zookeeper!") |
| self.out(stdout) |
| self.out(stderr) |
| raise RuntimeError("Failed to save credentials to zookeeper!") |
| self.out("User created.") |
|
|
| @property |
| def sasl_enabled(self): |
| return self.sasl_mechanism is not None |
|
|
| def bootstrap_server(self): |
| return '%s:%d' % (self.host, self.port) |
|
|
| def kafka_run_class_env(self): |
| env = super(KafkaFixture, self).kafka_run_class_env() |
| env['LOG_DIR'] = self.tmp_dir.join('logs').strpath |
| return env |
|
|
| def out(self, message): |
| log.info("*** Kafka [%s:%s]: %s", self.host, self.port or '(auto)', message) |
|
|
| def _create_zk_chroot(self): |
| self.out("Creating Zookeeper chroot node...") |
| args = self.kafka_run_class_args("org.apache.zookeeper.ZooKeeperMain", |
| "-server", |
| "%s:%d" % (self.zookeeper.host, |
| self.zookeeper.port), |
| "create", |
| "/%s" % (self.zk_chroot,), |
| "kafka-python") |
| env = self.kafka_run_class_env() |
| proc = subprocess.Popen(args, env=env, stdout=subprocess.PIPE, stderr=subprocess.PIPE) |
|
|
| stdout, stderr = proc.communicate() |
|
|
| if proc.returncode != 0: |
| self.out("Failed to create Zookeeper chroot node") |
| self.out(stdout) |
| self.out(stderr) |
| raise RuntimeError("Failed to create Zookeeper chroot node") |
| self.out("Kafka chroot created in Zookeeper!") |
|
|
| def start(self): |
| |
| properties = self.tmp_dir.join("kafka.properties") |
| jaas_conf = self.tmp_dir.join("kafka_server_jaas.conf") |
| properties_template = self.test_resource("kafka.properties") |
| jaas_conf_template = self.test_resource("kafka_server_jaas.conf") |
|
|
| args = self.kafka_run_class_args("kafka.Kafka", properties.strpath) |
| env = self.kafka_run_class_env() |
| if self.sasl_enabled: |
| opts = env.get('KAFKA_OPTS', '').strip() |
| opts += ' -Djava.security.auth.login.config={}'.format(jaas_conf.strpath) |
| env['KAFKA_OPTS'] = opts |
| self.render_template(jaas_conf_template, jaas_conf, vars(self)) |
|
|
| timeout = 5 |
| max_timeout = 120 |
| backoff = 1 |
| end_at = time.time() + max_timeout |
| tries = 1 |
| auto_port = (self.port is None) |
| while time.time() < end_at: |
| |
| |
| |
| if auto_port: |
| self.port = get_open_port() |
| self.out('Attempting to start on port %d (try #%d)' % (self.port, tries)) |
| self.render_template(properties_template, properties, vars(self)) |
|
|
| self.child = SpawnedService(args, env) |
| self.child.start() |
| timeout = min(timeout, max(end_at - time.time(), 0)) |
| if self._broker_ready(timeout) and self._scram_user_present(timeout): |
| break |
|
|
| self.child.dump_logs() |
| self.child.stop() |
|
|
| timeout *= 2 |
| time.sleep(backoff) |
| tries += 1 |
| backoff += 1 |
| else: |
| raise RuntimeError('Failed to start KafkaInstance before max_timeout') |
|
|
| (self._client,) = self.get_clients(1, client_id='_internal_client') |
|
|
| self.out("Done!") |
| self.running = True |
|
|
| def _broker_ready(self, timeout): |
| return self.child.wait_for(self.start_pattern, timeout=timeout) |
|
|
| def _scram_user_present(self, timeout): |
| |
| if not self.sasl_enabled or not self.sasl_mechanism.startswith('SCRAM-SHA-'): |
| return True |
| return self.child.wait_for(self.scram_pattern, timeout=timeout) |
|
|
| def open(self): |
| if self.running: |
| self.out("Instance already running") |
| return |
|
|
| |
| if self.tmp_dir is None: |
| self.tmp_dir = py.path.local.mkdtemp() |
| self.tmp_dir.ensure(dir=True) |
| self.tmp_dir.ensure('logs', dir=True) |
| self.tmp_dir.ensure('data', dir=True) |
|
|
| self.out("Running local instance...") |
| log.info(" host = %s", self.host) |
| log.info(" port = %s", self.port or '(auto)') |
| log.info(" transport = %s", self.transport) |
| log.info(" sasl_mechanism = %s", self.sasl_mechanism) |
| log.info(" broker_id = %s", self.broker_id) |
| log.info(" zk_host = %s", self.zookeeper.host) |
| log.info(" zk_port = %s", self.zookeeper.port) |
| log.info(" zk_chroot = %s", self.zk_chroot) |
| log.info(" replicas = %s", self.replicas) |
| log.info(" partitions = %s", self.partitions) |
| log.info(" tmp_dir = %s", self.tmp_dir.strpath) |
|
|
| self._create_zk_chroot() |
| self.sasl_config = self._sasl_config() |
| self.jaas_config = self._jaas_config() |
| |
| if self.sasl_enabled and self.sasl_mechanism.startswith("SCRAM-SHA") and self.broker_id == 0: |
| self._add_scram_user() |
| self.start() |
|
|
| atexit.register(self.close) |
|
|
| def __del__(self): |
| self.close() |
|
|
| def stop(self): |
| if not self.running: |
| self.out("Instance already stopped") |
| return |
|
|
| self.out("Stopping...") |
| self.child.stop() |
| self.child = None |
| self.running = False |
| self.out("Stopped!") |
|
|
| def close(self): |
| self.stop() |
| if self.tmp_dir is not None: |
| self.tmp_dir.remove() |
| self.tmp_dir = None |
| self.out("Done!") |
|
|
| def dump_logs(self): |
| super(KafkaFixture, self).dump_logs() |
| self.zookeeper.dump_logs() |
|
|
| def _send_request(self, request, timeout=None): |
| def _failure(error): |
| raise error |
| retries = 10 |
| while True: |
| node_id = self._client.least_loaded_node() |
| for connect_retry in range(40): |
| self._client.maybe_connect(node_id) |
| if self._client.connected(node_id): |
| break |
| self._client.poll(timeout_ms=100) |
| else: |
| raise RuntimeError('Could not connect to broker with node id %d' % (node_id,)) |
|
|
| try: |
| future = self._client.send(node_id, request) |
| future.error_on_callbacks = True |
| future.add_errback(_failure) |
| self._client.poll(future=future, timeout_ms=timeout) |
| return future.value |
| except Exception as exc: |
| time.sleep(1) |
| retries -= 1 |
| if retries == 0: |
| raise exc |
| else: |
| pass |
|
|
| def _create_topic(self, topic_name, num_partitions=None, replication_factor=None, timeout_ms=10000): |
| if num_partitions is None: |
| num_partitions = self.partitions |
| if replication_factor is None: |
| replication_factor = self.replicas |
|
|
| |
| if self.auto_create_topic and num_partitions == self.partitions and replication_factor == self.replicas: |
| self._create_topic_via_metadata(topic_name, timeout_ms) |
| elif env_kafka_version() >= (0, 10, 1, 0): |
| try: |
| self._create_topic_via_admin_api(topic_name, num_partitions, replication_factor, timeout_ms) |
| except InvalidReplicationFactorError: |
| |
| |
| time.sleep(0.5) |
| self._create_topic_via_admin_api(topic_name, num_partitions, replication_factor, timeout_ms) |
| else: |
| self._create_topic_via_cli(topic_name, num_partitions, replication_factor) |
|
|
| def _create_topic_via_metadata(self, topic_name, timeout_ms=10000): |
| self._send_request(MetadataRequest[0]([topic_name]), timeout_ms) |
|
|
| def _create_topic_via_admin_api(self, topic_name, num_partitions, replication_factor, timeout_ms=10000): |
| request = CreateTopicsRequest[0]([(topic_name, num_partitions, |
| replication_factor, [], [])], timeout_ms) |
| response = self._send_request(request, timeout=timeout_ms) |
| for topic_result in response.topic_errors: |
| error_code = topic_result[1] |
| if error_code != 0: |
| raise errors.for_code(error_code) |
|
|
| def _create_topic_via_cli(self, topic_name, num_partitions, replication_factor): |
| args = self.kafka_run_class_args('kafka.admin.TopicCommand', |
| '--zookeeper', '%s:%s/%s' % (self.zookeeper.host, |
| self.zookeeper.port, |
| self.zk_chroot), |
| '--create', |
| '--topic', topic_name, |
| '--partitions', self.partitions \ |
| if num_partitions is None else num_partitions, |
| '--replication-factor', self.replicas \ |
| if replication_factor is None \ |
| else replication_factor) |
| if env_kafka_version() >= (0, 10): |
| args.append('--if-not-exists') |
| env = self.kafka_run_class_env() |
| proc = subprocess.Popen(args, env=env, stdout=subprocess.PIPE, stderr=subprocess.PIPE) |
| stdout, stderr = proc.communicate() |
| if proc.returncode != 0: |
| if 'kafka.common.TopicExistsException' not in stdout: |
| self.out("Failed to create topic %s" % (topic_name,)) |
| self.out(stdout) |
| self.out(stderr) |
| raise RuntimeError("Failed to create topic %s" % (topic_name,)) |
|
|
| def get_topic_names(self): |
| args = self.kafka_run_class_args('kafka.admin.TopicCommand', |
| '--zookeeper', '%s:%s/%s' % (self.zookeeper.host, |
| self.zookeeper.port, |
| self.zk_chroot), |
| '--list' |
| ) |
| env = self.kafka_run_class_env() |
| env.pop('KAFKA_LOG4J_OPTS') |
| proc = subprocess.Popen(args, env=env, stdout=subprocess.PIPE, stderr=subprocess.PIPE) |
| stdout, stderr = proc.communicate() |
| if proc.returncode != 0: |
| self.out("Failed to list topics!") |
| self.out(stdout) |
| self.out(stderr) |
| raise RuntimeError("Failed to list topics!") |
| return stdout.decode().splitlines(False) |
|
|
| def create_topics(self, topic_names, num_partitions=None, replication_factor=None): |
| for topic_name in topic_names: |
| self._create_topic(topic_name, num_partitions, replication_factor) |
|
|
| def _enrich_client_params(self, params, **defaults): |
| params = params.copy() |
| for key, value in defaults.items(): |
| params.setdefault(key, value) |
| params.setdefault('bootstrap_servers', self.bootstrap_server()) |
| if self.sasl_enabled: |
| params.setdefault('sasl_mechanism', self.sasl_mechanism) |
| params.setdefault('security_protocol', self.transport) |
| if self.sasl_mechanism in ('PLAIN', 'SCRAM-SHA-256', 'SCRAM-SHA-512'): |
| params.setdefault('sasl_plain_username', self.broker_user) |
| params.setdefault('sasl_plain_password', self.broker_password) |
| return params |
|
|
| @staticmethod |
| def _create_many_clients(cnt, cls, *args, **params): |
| client_id = params['client_id'] |
| for _ in range(cnt): |
| params['client_id'] = '%s_%s' % (client_id, random_string(4)) |
| yield cls(*args, **params) |
|
|
| def get_clients(self, cnt=1, **params): |
| params = self._enrich_client_params(params, client_id='client') |
| for client in self._create_many_clients(cnt, KafkaClient, **params): |
| yield client |
|
|
| def get_admin_clients(self, cnt, **params): |
| params = self._enrich_client_params(params, client_id='admin_client') |
| for client in self._create_many_clients(cnt, KafkaAdminClient, **params): |
| yield client |
|
|
| def get_consumers(self, cnt, topics, **params): |
| params = self._enrich_client_params( |
| params, client_id='consumer', heartbeat_interval_ms=500, auto_offset_reset='earliest' |
| ) |
| for client in self._create_many_clients(cnt, KafkaConsumer, *topics, **params): |
| yield client |
|
|
| def get_producers(self, cnt, **params): |
| params = self._enrich_client_params(params, client_id='producer') |
| for client in self._create_many_clients(cnt, KafkaProducer, **params): |
| yield client |
|
|