summaryrefslogtreecommitdiff
path: root/kafka
Commit message (Collapse)AuthorAgeFilesLines
* Remove deprecated `ConnectionError`Remove-deprecated-ConnectionErrorJeff Widman2019-05-241-4/+0
| | | | | | | This has been deprecated for a bit in favor of `KafkaConnectionError` because it conflicts with Python's built-in `ConnectionError`. Time to remove it as part of cleaning up our old deprecated code.
* Update docs for api_version_auto_timeout_ms (#1812)Jeff Widman2019-05-242-2/+2
| | | | | | | | | The docs for `api_version_auto_timeout_ms` mention setting `api_version='auto'` but that value has been deprecated for years in favor of `api_version=None`. Updating the docs for now, and will remove support for `'auto'` in next major version bump.
* Fix typo in _fetch_all_topic_metadata function (#1809)Brian Sang2019-05-231-1/+1
|
* Make partitions_for_topic a read-through cache (#1781)Brian Sang2019-05-221-8/+25
| | | If the cluster metadata object has no info about the topic, then issue a blocking metadata call to fetch it.
* Remove unused imports (#1808)Jeff Widman2019-05-221-2/+2
|
* Use futures to parallelize calls to _send_request_to_node() (#1807)Lou-Cipher2019-05-211-34/+75
| | | | | Use `futures` to parallelize calls to `_send_request_to_node()` This allows queries that need to go to multiple brokers to be run in parallel.
* Update link to kafka docsJeff Widman2019-05-171-1/+1
| | | Now that the old zookeeper consumer has been completely deprecated/removed, these are no longer the "new consumer configs" but rather simply the "consumer configs"
* A little python cleanup (#1805)Jeff Widman2019-05-171-4/+2
| | | | | | 1. Remove unused variable: `partitions_for_topic` 2. No need to cast to list as `sorted()` already returns a list 3. Using `enumerate()` is cleaner than `range(len())` and handles assigning `member`
* Bump version for development of next releaseDana Powers2019-04-031-1/+1
|
* Release 1.4.61.4.6Dana Powers2019-04-021-1/+1
|
* Do not call state_change_callback with lock (#1775)Dana Powers2019-04-022-21/+29
|
* Additional BrokerConnection locks to synchronize protocol/IFR state (#1768)Dana Powers2019-04-021-61/+85
|
* Return connection state explicitly after close in connect() (#1778)Dana Powers2019-04-021-1/+3
|
* Reset reconnect backoff on SSL connection (#1777)Dana Powers2019-04-021-0/+1
|
* Fix possible AttribueError during conn._close_socket (#1776)Dana Powers2019-04-011-1/+1
|
* Dont treat popped conn.close() as failure in state change callback (#1773)Dana Powers2019-04-011-3/+10
|
* Avoid race condition on client._conns in send() (#1772)Dana Powers2019-03-311-2/+3
| | | There was a very small possibility that between checking `self._can_send_request(node_id)` and grabbing the connection object via `self._conns[node_id]` that the connection could get closed / recycled / removed from _conns and cause a KeyError. This PR should prevent such a KeyError. In the case where the connection is disconnected by the time we call send(), we should expect conn.send() simply to fail the request.
* lock client.check_version (#1771)Dana Powers2019-03-311-0/+5
|
* Dont wakeup during maybe_refresh_metadata -- it is only called by poll() (#1769)Dana Powers2019-03-301-4/+4
|
* Revert 703f0659 / fix 0.8.2 protocol quick detection (#1763)Dana Powers2019-03-272-6/+9
|
* Send pending requests before waiting for responses (#1762)Dana Powers2019-03-271-2/+4
|
* Dont do client wakeup when sending from sender thread (#1761)Dana Powers2019-03-242-6/+10
|
* Update sasl configuration docstringsDana Powers2019-03-235-24/+24
|
* Support SASL OAuthBearer Authentication (#1750)Phong Pham2019-03-227-6/+117
|
* Maintain shadow cluster metadata for bootstrapping (#1753)Dana Powers2019-03-212-35/+21
|
* Allow configuration of SSL Ciphers (#1755)Dana Powers2019-03-214-1/+28
|
* Wrap SSL sockets after connecting (#1754)Dana Powers2019-03-211-19/+11
|
* Fix race condition in protocol.send_bytes (#1752)Filip Stefanak2019-03-211-1/+2
|
* Bump version for developmentDana Powers2019-03-141-1/+1
|
* Release 1.4.51.4.5Dana Powers2019-03-141-1/+1
|
* Error if connections_max_idle_ms not larger than request_timeout_ms (#1688)Jeff Widman2019-03-141-3/+7
|
* Retry bootstrapping after backoff when necessary (#1736)Dana Powers2019-03-142-84/+89
| | | | | | | The current client attempts to bootstrap once during initialization, but if it fails there is no second attempt and the client will be inoperable. This can happen, for example, if an entire cluster is down at the time a long-running client starts execution. This commit attempts to fix this by removing the synchronous bootstrapping from `KafkaClient` init, and instead merges bootstrap metadata with the cluster metadata. The Java client uses a similar approach. This allows us to continue falling back to bootstrap data when necessary throughout the life of a long-running consumer or producer. Fix #1670
* Fix default protocol parser versionDana Powers2019-03-131-0/+3
|
* Minor updates to client_async.pyDana Powers2019-03-131-4/+4
|
* Recheck connecting nodes sooner when refreshing metadata (#1737)Dana Powers2019-03-131-3/+1
|
* Don't recheck version if api_versions data is already cached (#1738)Dana Powers2019-03-131-0/+3
| | | I noticed during local testing that version probing was happening twice when connecting to newer broker versions. This was because we call check_version() once explicitly, and then again implicitly within get_api_versions(). But once we have _api_versions data cached, we can just return it and avoid probing versions a second time.
* Attempt to join heartbeat thread during close() (#1735)Dana Powers2019-03-131-3/+6
| | | | | | | Underlying issue here is a race on consumer.close() between the client, the connections/sockets, and the heartbeat thread. Although the heartbeat thread is signaled to close, we do not block for it. So when we go on to close the client and its underlying connections, if the heartbeat is still doing work it can cause errors/crashes if it attempts to access the now closed objects (selectors and/or sockets, primarily). So this commit adds a blocking thread join to the heartbeat close. This may cause some additional blocking time on consumer.close() while the heartbeat thread finishes. But it should be small in average case and in the worst case will be no longer than the heartbeat_timeout_ms (though if we timeout the join, race errors may still occur). Fix #1666
* 1701 use last offset from fetch v4 if available (#1724)Keith So2019-03-133-0/+28
|
* Catch thrown OSError by python 3.7 when creating a connection (#1694)Daniel Johansson2019-03-121-0/+3
|
* Ignore lookup_coordinator result in commit_offsets_async (#1712)Faqa2019-03-121-1/+2
|
* Synchronize puts to KafkaConsumer protocol buffer during async sendsDana Powers2019-03-121-21/+36
|
* Do network connections and writes in KafkaClient.poll() (#1729)Dana Powers2019-03-084-49/+76
| | | | | | * Add BrokerConnection.send_pending_requests to support async network sends * Send network requests during KafkaClient.poll() rather than in KafkaClient.send() * Dont acquire lock during KafkaClient.send if node is connected / ready * Move all network connection IO into KafkaClient.poll()
* Do not require client lock for read-only operations (#1730)Dana Powers2019-03-061-50/+50
| | | In an effort to reduce the surface area of lock coordination, and thereby hopefully reduce lock contention, I think we can remove locking from the read-only KafkaClient methods: connected, is_disconnected, in_flight_request_count, and least_loaded_node . Given that the read data could change after the lock is released but before the caller uses it, the value of acquiring a lock here does not seem high to me.
* Make NotEnoughReplicasError/NotEnoughReplicasAfterAppendError retriable (#1722)le-linh2019-03-031-0/+2
|
* Remove unused importJeff Widman2019-01-281-1/+0
|
* Improve KafkaConsumer join group / only enable Heartbeat Thread during ↵Dana Powers2019-01-151-11/+23
| | | | stable group (#1695)
* Remove unused `skip_double_compressed_messages`Jeff Widman2019-01-132-16/+0
| | | | | | | | | | This `skip_double_compressed_messages` flag was added in https://github.com/dpkp/kafka-python/pull/755 in order to fix https://github.com/dpkp/kafka-python/issues/718. However, grep'ing through the code, it looks like it this is no longer used anywhere and doesn't do anything. So removing it.
* Timeout all unconnected conns (incl SSL) after request_timeout_msDana Powers2019-01-131-6/+8
|
* Fix `AttributeError` caused by `getattr()`Jeff Widman2019-01-071-1/+2
| | | | | | | `getattr(object, 'x', object.y)` will evaluate the default argument `object.y` regardless of whether `'x'` exists. For details see: https://stackoverflow.com/q/31443989/770425
* Fix SSL connection testing in Python 3.7Ben Weir2019-01-031-0/+7
|