diff options
author | Taras <voyn1991@gmail.com> | 2017-10-12 11:07:14 +0300 |
---|---|---|
committer | Taras <voyn1991@gmail.com> | 2017-10-12 11:10:44 +0300 |
commit | d10051bb09942bfd48c6f262a8cdbf5651963c2e (patch) | |
tree | 1f6a262324e3ced9c8655d71fbed5e9ed8a0234d /kafka/producer | |
parent | e992fbfad926486766ff7b63a499f9cf29984fec (diff) | |
download | kafka-python-d10051bb09942bfd48c6f262a8cdbf5651963c2e.tar.gz |
Added minor fixes for PR review
Diffstat (limited to 'kafka/producer')
-rw-r--r-- | kafka/producer/kafka.py | 2 | ||||
-rw-r--r-- | kafka/producer/record_accumulator.py | 18 |
2 files changed, 4 insertions, 16 deletions
diff --git a/kafka/producer/kafka.py b/kafka/producer/kafka.py index a53ac49..5638b61 100644 --- a/kafka/producer/kafka.py +++ b/kafka/producer/kafka.py @@ -370,7 +370,7 @@ class KafkaProducer(object): else: checker, compression_attrs = self._COMPRESSORS[ct] assert checker(), "Libraries for {} compression codec not found".format(ct) - self.config['compression_type'] = compression_attrs + self.config['compression_attrs'] = compression_attrs message_version = self._max_usable_produce_magic() self._accumulator = RecordAccumulator(message_version=message_version, metrics=self._metrics, **self.config) diff --git a/kafka/producer/record_accumulator.py b/kafka/producer/record_accumulator.py index 0c0ce27..716ae65 100644 --- a/kafka/producer/record_accumulator.py +++ b/kafka/producer/record_accumulator.py @@ -149,7 +149,7 @@ class RecordAccumulator(object): will block up to max_block_ms, raising an exception on timeout. In the current implementation, this setting is an approximation. Default: 33554432 (32MB) - compression_type (int): The compression type for all data generated by + compression_attrs (int): The compression type for all data generated by the producer. Valid values are gzip(1), snappy(2), lz4(3), or none(0). Compression is of full batches of data, so the efficacy of batching @@ -168,7 +168,7 @@ class RecordAccumulator(object): DEFAULT_CONFIG = { 'buffer_memory': 33554432, 'batch_size': 16384, - 'compression_type': None, + 'compression_attrs': 0, 'linger_ms': 0, 'retry_backoff_ms': 100, 'message_version': 0, @@ -176,24 +176,12 @@ class RecordAccumulator(object): 'metric_group_prefix': 'producer-metrics', } - _COMPRESSORS = { - 'gzip': LegacyRecordBatchBuilder.CODEC_GZIP, - 'snappy': LegacyRecordBatchBuilder.CODEC_SNAPPY, - 'lz4': LegacyRecordBatchBuilder.CODEC_LZ4, - None: LegacyRecordBatchBuilder.CODEC_NONE - } - def __init__(self, **configs): self.config = copy.copy(self.DEFAULT_CONFIG) for key in self.config: if key in configs: self.config[key] = configs.pop(key) - # Convert compression type to INT presentation. Mostly for unit tests, - # as Producer should pass already converted values. - ct = self.config["compression_type"] - self.config["compression_type"] = self._COMPRESSORS.get(ct, ct) - self._closed = False self._flushes_in_progress = AtomicInteger() self._appends_in_progress = AtomicInteger() @@ -269,7 +257,7 @@ class RecordAccumulator(object): records = MemoryRecordsBuilder( self.config['message_version'], - self.config['compression_type'], + self.config['compression_attrs'], self.config['batch_size'] ) |