Clickhouse-connect

Latest version: v0.8.16

Safety actively analyzes 723650 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 14 of 20

0.5.9

Bug Fixes
- Large query results using `zstd` compression incorrectly buffered all incoming data at the start of the query,
consuming an excessive amount of memory. This has been fixed. https://github.com/ClickHouse/clickhouse-connect/issues/122
Big thanks to [Denny Crane](https://github.com/den-crane) for his detailed investigation of the problem. Note that
this affected large queries using the default `compress=True` client setting, as ClickHouse would prefer `zstd` compression
in those cases.
- Fixed an issue where a small query_limit would break client initialization due to an incomplete read of the `system.settings`
table. https://github.com/ClickHouse/clickhouse-connect/issues/123

Improvement
- Stream error handling has been improved so exceptions thrown while consuming a stream should be correctly propagated.
This includes unexpected stream closures by the ClickHouse server. Errors inserted into the HTTP response by ClickHouse
during a query should also be reported as part of a StreamFailureError

0.5.8

Bug Fix
- Return empty dataframe instead of empty list when no records returned from `query_df` method Fixes
https://github.com/ClickHouse/clickhouse-connect/issues/118

Default parameter change
- The client `query_limit` now defaults to 0 (unlimited rows returned), since the previous default of 5000 was unintuitive
and led to confusion when limited results were returned.

New Feature
- Allow client side control of datetime.datetime timezones for query results. The client `query` methods for native
Python results now accept two new parameters: `query_tz` is the timezone to be assigned for any DateTime or DateTime64
objects in the results, while timezones can be set per column using the `column_tzs` dictionary of column names to
timezones. See the [test file](https://github.com/ClickHouse/clickhouse-connect/blob/main/tests/integration_tests/test_timezones.py)
for simple examples. This is a workaround for https://github.com/ClickHouse/clickhouse-connect/issues/120 and the
underlying ClickHouse issue https://github.com/ClickHouse/ClickHouse/issues/40397 Note that this issue only affects DateTime
columns, not DateTime64, although the query context parameters will override the returned DateTime64 timezone as well.

0.5.7

Bug Fix
- Http proxies did not work after removing the requests library. https://github.com/ClickHouse/clickhouse-connect/issues/114.
This should be fixed. Note that socks proxies are still not supported directly, but can be added by creating a correctly
configured urllib3 SOCKSProxyManager and using it as the `pool_mgr` argument to teh `clickhouse_connect.create_client` method.

0.5.6

Bug Fix
- Dataframe inserts would incorrectly modify null-like elements of the inserted dataframe. https://github.com/ClickHouse/clickhouse-connect/issues/112.
This should be fixed

0.5.5

Bug Fix
- Queries of LowCardinality columns using pandas or numpy query methods would result in an exception. https://github.com/ClickHouse/clickhouse-connect/issues/108
This has been fixed.

0.5.4

New Features
* Several streaming query methods have been added to the core ClickHouse Connect client. Each of these methods returns a StreamContext object, which must be used as a Python `with` Context to stream data (this ensures the underlying
streaming response is properly closed/consumed.) For simple examples, see the basic [tests](https://github.com/ClickHouse/clickhouse-connect/blob/main/tests/integration_tests/test_streaming.py).
* `query_column_block_stream` -- returns a generator of blocks in column oriented (Native) format. Fastest method for retrieving data in native Python format
* `query_row_block_stream` -- returns a generator of blocks in row oriented format. Used for processing data in a "batch" of rows at time while limiting memory usage
* `query_rows_stream` -- returns a convenience generator to process rows one at a time (data is still loaded in ClickHouse blocks to preserve memory)
* `query_np_stream` -- returns a generator where each ClickHouse data block is transformed into a Numpy array
* `query_df_stream` -- returns a generator where each ClickHouse data block is transformed into a Pandas Dataframe
* The `client_name` is now reported in a standardized way to ClickHouse (as the `http_user_agent`). For better tracking of your
Python application, use the new `product_name` common setting or set `client_name` `get_client` parameter to identify your product
as `<your-product-name>/<product-version>`.

Performance Improvements
* C/Cython optimizations for transforming ClickHouse data to Python types have been improved, and additional datatypes have been
optimized in Cython. The performance increase over the previous 0.5.x version is approximately 10% for "normal" read queries.
* Transformation of Numpy arrays and Pandas Dataframes has been completely rewritten to avoid an intermediate conversion to
Python types. As a result, querying in Numpy format, and especially Pandas format, has been **significantly** improved -- from 2x
for small datasets to 5x or more for very large Pandas DataFrames (even without streaming). Queries including Numpy datetime64 or
Pandas Timestamp objects have particularly benefited from the new implementation.

Bug Fixes
* The default `maxsize` for concurrent HTTP connections to a single host was accidentally dropped in the 0.5.x release. It
has been restored to 8 for better performance when using multiple client objects.
* A single low level retry has been restored for HTTP connections on ConnectionReset or RemoteDisconnected exceptions. This
should reduce connection errors related to ClickHouse closing expired KeepAlive connections.

Internal Changes
* As noted above, streaming, contexts and exception handling have been tightened up to avoid leaving HTTP responses open
when querying streams.
* Previous versions used `threading.local()` variables to store context information during query processing. The architecture
has been changed to pass the relevant Query or Insert Context to transformation methods instead of relying on thread local
variables. This is significantly safer in an environment where multiple queries can conceivably be open at the same on the
same thread (for example, if using async functions).
* Per query formatting logic has moved from `ClickHouseType` to the `QueryContext`.
* `ClickHouseType` methods have been renamed to remove outdated references to `native` format (everything is native now)
* Upgraded Cython Build to 3.0.11alpha release

Page 14 of 20

© 2025 Safety CLI Cybersecurity Inc. All Rights Reserved.