settings.md 70.5 KB
Newer Older
1
# Settings {#settings}
2

3
## distributed\_product\_mode {#distributed-product-mode}
4

5
Changes the behavior of [distributed subqueries](../../sql-reference/operators/in.md).
6

7
ClickHouse applies this setting when the query contains the product of distributed tables, i.e. when the query for a distributed table contains a non-GLOBAL subquery for the distributed table.
8 9 10

Restrictions:

11 12 13
-   Only applied for IN and JOIN subqueries.
-   Only if the FROM section uses a distributed table containing more than one shard.
-   If the subquery concerns a distributed table containing more than one shard.
14
-   Not used for a table-valued [remote](../../sql-reference/table-functions/remote.md) function.
15

16
Possible values:
17

18 19 20 21
-   `deny` — Default value. Prohibits using these types of subqueries (returns the “Double-distributed in/JOIN subqueries is denied” exception).
-   `local` — Replaces the database and table in the subquery with local ones for the destination server (shard), leaving the normal `IN`/`JOIN.`
-   `global` — Replaces the `IN`/`JOIN` query with `GLOBAL IN`/`GLOBAL JOIN.`
-   `allow` — Allows the use of these types of subqueries.
22

23
## enable\_optimize\_predicate\_expression {#enable-optimize-predicate-expression}
24

25
Turns on predicate pushdown in `SELECT` queries.
26

27
Predicate pushdown may significantly reduce network traffic for distributed queries.
28 29 30

Possible values:

31 32
-   0 — Disabled.
-   1 — Enabled.
33

F
filimonov 已提交
34
Default value: 1.
35

S
Fixes  
Sergei Bocharov 已提交
36
Usage
37

38
Consider the following queries:
39

40 41
1.  `SELECT count() FROM test_table WHERE date = '2018-10-10'`
2.  `SELECT count() FROM (SELECT * FROM test_table) WHERE date = '2018-10-10'`
42

43
If `enable_optimize_predicate_expression = 1`, then the execution time of these queries is equal because ClickHouse applies `WHERE` to the subquery when processing it.
44

45
If `enable_optimize_predicate_expression = 0`, then the execution time of the second query is much longer, because the `WHERE` clause applies to all the data after the subquery finishes.
46

47
## fallback\_to\_stale\_replicas\_for\_distributed\_queries {#settings-fallback_to_stale_replicas_for_distributed_queries}
48

49
Forces a query to an out-of-date replica if updated data is not available. See [Replication](../../engines/table-engines/mergetree-family/replication.md).
50 51 52

ClickHouse selects the most relevant from the outdated replicas of the table.

53
Used when performing `SELECT` from a distributed table that points to replicated tables.
54 55 56

By default, 1 (enabled).

57
## force\_index\_by\_date {#settings-force_index_by_date}
58

59
Disables query execution if the index can’t be used by date.
60 61 62

Works with tables in the MergeTree family.

63
If `force_index_by_date=1`, ClickHouse checks whether the query has a date key condition that can be used for restricting data ranges. If there is no suitable condition, it throws an exception. However, it does not check whether the condition reduces the amount of data to read. For example, the condition `Date != ' 2000-01-01 '` is acceptable even when it matches all the data in the table (i.e., running the query requires a full scan). For more information about ranges of data in MergeTree tables, see [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md).
64

65
## force\_primary\_key {#force-primary-key}
66 67 68 69 70

Disables query execution if indexing by the primary key is not possible.

Works with tables in the MergeTree family.

71
If `force_primary_key=1`, ClickHouse checks to see if the query has a primary key condition that can be used for restricting data ranges. If there is no suitable condition, it throws an exception. However, it does not check whether the condition reduces the amount of data to read. For more information about data ranges in MergeTree tables, see [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md).
72

73
## format\_schema {#format-schema}
A
Alexander Tokmakov 已提交
74

75
This parameter is useful when you are using formats that require a schema definition, such as [Cap’n Proto](https://capnproto.org/) or [Protobuf](https://developers.google.com/protocol-buffers/). The value depends on the format.
76

77
## fsync\_metadata {#fsync-metadata}
78

79
Enables or disables [fsync](http://pubs.opengroup.org/onlinepubs/9699919799/functions/fsync.html) when writing `.sql` files. Enabled by default.
80

A
alexey-milovidov 已提交
81
It makes sense to disable it if the server has millions of tiny tables that are constantly being created and destroyed.
82

83
## enable\_http\_compression {#settings-enable_http_compression}
84

85
Enables or disables data compression in the response to an HTTP request.
86 87 88 89 90

For more information, read the [HTTP interface description](../../interfaces/http.md).

Possible values:

91 92
-   0 — Disabled.
-   1 — Enabled.
93 94 95

Default value: 0.

96
## http\_zlib\_compression\_level {#settings-http_zlib_compression_level}
97

98
Sets the level of data compression in the response to an HTTP request if [enable\_http\_compression = 1](#settings-enable_http_compression).
99

100
Possible values: Numbers from 1 to 9.
101 102 103

Default value: 3.

104
## http\_native\_compression\_disable\_checksumming\_on\_decompress {#settings-http_native_compression_disable_checksumming_on_decompress}
105

106
Enables or disables checksum verification when decompressing the HTTP POST data from the client. Used only for ClickHouse native compression format (not used with `gzip` or `deflate`).
107 108 109 110 111

For more information, read the [HTTP interface description](../../interfaces/http.md).

Possible values:

112 113
-   0 — Disabled.
-   1 — Enabled.
114 115 116

Default value: 0.

117
## send\_progress\_in\_http\_headers {#settings-send_progress_in_http_headers}
118

119
Enables or disables `X-ClickHouse-Progress` HTTP response headers in `clickhouse-server` responses.
120 121 122 123 124

For more information, read the [HTTP interface description](../../interfaces/http.md).

Possible values:

125 126
-   0 — Disabled.
-   1 — Enabled.
127 128 129

Default value: 0.

130
## max\_http\_get\_redirects {#setting-max_http_get_redirects}
131

132
Limits the maximum number of HTTP GET redirect hops for [URL](../../engines/table-engines/special/url.md)-engine tables. The setting applies to both types of tables: those created by the [CREATE TABLE](../../sql-reference/statements/create/table.md) query and by the [url](../../sql-reference/table-functions/url.md) table function.
133 134 135

Possible values:

136 137
-   Any positive integer number of hops.
-   0 — No hops allowed.
138 139 140

Default value: 0.

141
## input\_format\_allow\_errors\_num {#settings-input_format_allow_errors_num}
142 143 144 145 146

Sets the maximum number of acceptable errors when reading from text formats (CSV, TSV, etc.).

The default value is 0.

A
Alexander Tokmakov 已提交
147
Always pair it with `input_format_allow_errors_ratio`.
148 149 150

If an error occurred while reading rows but the error counter is still less than `input_format_allow_errors_num`, ClickHouse ignores the row and moves on to the next one.

A
Alexander Tokmakov 已提交
151
If both `input_format_allow_errors_num` and `input_format_allow_errors_ratio` are exceeded, ClickHouse throws an exception.
152

153
## input\_format\_allow\_errors\_ratio {#settings-input_format_allow_errors_ratio}
154 155 156 157 158 159

Sets the maximum percentage of errors allowed when reading from text formats (CSV, TSV, etc.).
The percentage of errors is set as a floating-point number between 0 and 1.

The default value is 0.

A
Alexander Tokmakov 已提交
160
Always pair it with `input_format_allow_errors_num`.
161 162 163

If an error occurred while reading rows but the error counter is still less than `input_format_allow_errors_ratio`, ClickHouse ignores the row and moves on to the next one.

A
Alexander Tokmakov 已提交
164
If both `input_format_allow_errors_num` and `input_format_allow_errors_ratio` are exceeded, ClickHouse throws an exception.
165

166
## input\_format\_values\_interpret\_expressions {#settings-input_format_values_interpret_expressions}
167

168
Enables or disables the full SQL parser if the fast stream parser can’t parse the data. This setting is used only for the [Values](../../interfaces/formats.md#data-format-values) format at the data insertion. For more information about syntax parsing, see the [Syntax](../../sql-reference/syntax.md) section.
169 170 171

Possible values:

172
-   0 — Disabled.
173

174
    In this case, you must provide formatted data. See the [Formats](../../interfaces/formats.md) section.
175

176
-   1 — Enabled.
177

178
    In this case, you can use an SQL expression as a value, but data insertion is much slower this way. If you insert only formatted data, then ClickHouse behaves as if the setting value is 0.
179 180 181

Default value: 1.

S
Fixes  
Sergei Bocharov 已提交
182
Example of Use
183

184
Insert the [DateTime](../../sql-reference/data-types/datetime.md) type value with the different settings.
185

186
``` sql
187 188
SET input_format_values_interpret_expressions = 0;
INSERT INTO datetime_t VALUES (now())
189
```
S
Fixes  
Sergei Bocharov 已提交
190

191
``` text
192 193 194 195
Exception on client:
Code: 27. DB::Exception: Cannot parse input: expected ) before: now()): (at row 1)
```

196
``` sql
197 198
SET input_format_values_interpret_expressions = 1;
INSERT INTO datetime_t VALUES (now())
199
```
S
Fixes  
Sergei Bocharov 已提交
200

201
``` text
202 203 204
Ok.
```

205
The last query is equivalent to the following:
206

207
``` sql
208 209
SET input_format_values_interpret_expressions = 0;
INSERT INTO datetime_t SELECT now()
210
```
S
Fixes  
Sergei Bocharov 已提交
211

212
``` text
213 214 215
Ok.
```

216
## input\_format\_values\_deduce\_templates\_of\_expressions {#settings-input_format_values_deduce_templates_of_expressions}
S
Fixes  
Sergei Bocharov 已提交
217

218
Enables or disables template deduction for SQL expressions in [Values](../../interfaces/formats.md#data-format-values) format. It allows parsing and interpreting expressions in `Values` much faster if expressions in consecutive rows have the same structure. ClickHouse tries to deduce template of an expression, parse the following rows using this template and evaluate the expression on a batch of successfully parsed rows.
B
BayoNet 已提交
219 220 221 222 223 224 225 226 227

Possible values:

-   0 — Disabled.
-   1 — Enabled.

Default value: 1.

For the following query:
S
Fixes  
Sergei Bocharov 已提交
228

229
``` sql
A
Alexander Tokmakov 已提交
230
INSERT INTO test VALUES (lower('Hello')), (lower('world')), (lower('INSERT')), (upper('Values')), ...
231
```
S
Fixes  
Sergei Bocharov 已提交
232

B
BayoNet 已提交
233 234 235
-   If `input_format_values_interpret_expressions=1` and `format_values_deduce_templates_of_expressions=0`, expressions are interpreted separately for each row (this is very slow for large number of rows).
-   If `input_format_values_interpret_expressions=0` and `format_values_deduce_templates_of_expressions=1`, expressions in the first, second and third rows are parsed using template `lower(String)` and interpreted together, expression in the forth row is parsed with another template (`upper(String)`).
-   If `input_format_values_interpret_expressions=1` and `format_values_deduce_templates_of_expressions=1`, the same as in previous case, but also allows fallback to interpreting expressions separately if it’s not possible to deduce template.
A
Alexander Tokmakov 已提交
236

237
## input\_format\_values\_accurate\_types\_of\_literals {#settings-input-format-values-accurate-types-of-literals}
S
Fixes  
Sergei Bocharov 已提交
238

B
BayoNet 已提交
239
This setting is used only when `input_format_values_deduce_templates_of_expressions = 1`. It can happen, that expressions for some column have the same structure, but contain numeric literals of different types, e.g.
240 241

``` sql
A
Alexander Tokmakov 已提交
242 243 244 245
(..., abs(0), ...),             -- UInt64 literal
(..., abs(3.141592654), ...),   -- Float64 literal
(..., abs(-1), ...),            -- Int64 literal
```
S
Fixes  
Sergei Bocharov 已提交
246

B
BayoNet 已提交
247 248 249 250 251 252 253 254 255 256 257
Possible values:

-   0 — Disabled.

    In this case, ClickHouse may use a more general type for some literals (e.g., `Float64` or `Int64` instead of `UInt64` for `42`), but it may cause overflow and precision issues.

-   1 — Enabled.

    In this case, ClickHouse checks the actual type of literal and uses an expression template of the corresponding type. In some cases, it may significantly slow down expression evaluation in `Values`.

Default value: 1.
A
Alexander Tokmakov 已提交
258

259
## input\_format\_defaults\_for\_omitted\_fields {#session_settings-input_format_defaults_for_omitted_fields}
260

261
When performing `INSERT` queries, replace omitted input column values with default values of the respective columns. This option only applies to [JSONEachRow](../../interfaces/formats.md#jsoneachrow), [CSV](../../interfaces/formats.md#csv) and [TabSeparated](../../interfaces/formats.md#tabseparated) formats.
262 263

!!! note "Note"
264
    When this option is enabled, extended table metadata are sent from server to client. It consumes additional computing resources on the server and can reduce performance.
265

266 267
Possible values:

268 269
-   0 — Disabled.
-   1 — Enabled.
270

271
Default value: 1.
272

273
## input\_format\_tsv\_empty\_as\_default {#settings-input-format-tsv-empty-as-default}
A
fixes  
Alexander Tokmakov 已提交
274 275 276 277 278

When enabled, replace empty input fields in TSV with default values. For complex default expressions `input_format_defaults_for_omitted_fields` must be enabled too.

Disabled by default.

279
## input\_format\_null\_as\_default {#settings-input-format-null-as-default}
A
Alexander Tokmakov 已提交
280

281
Enables or disables using default values if input data contain `NULL`, but data type of the corresponding column in not `Nullable(T)` (for text input formats).
A
Alexander Tokmakov 已提交
282

283
## input\_format\_skip\_unknown\_fields {#settings-input-format-skip-unknown-fields}
284

285
Enables or disables skipping insertion of extra data.
286

287
When writing data, ClickHouse throws an exception if input data contain columns that do not exist in the target table. If skipping is enabled, ClickHouse doesn’t insert extra data and doesn’t throw an exception.
288

289 290
Supported formats:

291 292 293 294
-   [JSONEachRow](../../interfaces/formats.md#jsoneachrow)
-   [CSVWithNames](../../interfaces/formats.md#csvwithnames)
-   [TabSeparatedWithNames](../../interfaces/formats.md#tabseparatedwithnames)
-   [TSKV](../../interfaces/formats.md#tskv)
295 296

Possible values:
297

298 299
-   0 — Disabled.
-   1 — Enabled.
300

301 302
Default value: 0.

303
## input\_format\_import\_nested\_json {#settings-input_format_import_nested_json}
304

305
Enables or disables the insertion of JSON data with nested objects.
306 307 308

Supported formats:

309
-   [JSONEachRow](../../interfaces/formats.md#jsoneachrow)
310 311 312

Possible values:

313 314
-   0 — Disabled.
-   1 — Enabled.
315 316 317

Default value: 0.

S
Fixes  
Sergei Bocharov 已提交
318
See also:
319

320
-   [Usage of Nested Structures](../../interfaces/formats.md#jsoneachrow-nested) with the `JSONEachRow` format.
321

322
## input\_format\_with\_names\_use\_header {#settings-input-format-with-names-use-header}
323 324

Enables or disables checking the column order when inserting data.
B
BayoNet 已提交
325

326 327 328
To improve insert performance, we recommend disabling this check if you are sure that the column order of the input data is the same as in the target table.

Supported formats:
329

330 331
-   [CSVWithNames](../../interfaces/formats.md#csvwithnames)
-   [TabSeparatedWithNames](../../interfaces/formats.md#tabseparatedwithnames)
332 333 334

Possible values:

335 336
-   0 — Disabled.
-   1 — Enabled.
337 338

Default value: 1.
339

340
## date\_time\_input\_format {#settings-date_time_input_format}
341

342
Allows choosing a parser of the text representation of date and time.
343

344
The setting doesn’t apply to [date and time functions](../../sql-reference/functions/date-time-functions.md).
345 346 347

Possible values:

348
-   `'best_effort'` — Enables extended parsing.
349

350
    ClickHouse can parse the basic `YYYY-MM-DD HH:MM:SS` format and all [ISO 8601](https://en.wikipedia.org/wiki/ISO_8601) date and time formats. For example, `'2018-06-08T01:02:03.000Z'`.
351

352
-   `'basic'` — Use basic parser.
353

354
    ClickHouse can parse only the basic `YYYY-MM-DD HH:MM:SS` format. For example, `'2019-08-20 10:18:56'`.
355 356

Default value: `'basic'`.
357

S
Fixes  
Sergei Bocharov 已提交
358
See also:
359

360 361
-   [DateTime data type.](../../sql-reference/data-types/datetime.md)
-   [Functions for working with dates and times.](../../sql-reference/functions/date-time-functions.md)
362

363
## join\_default\_strictness {#settings-join_default_strictness}
B
BayoNet 已提交
364

365
Sets default strictness for [JOIN clauses](../../sql-reference/statements/select/join.md#select-join).
B
BayoNet 已提交
366

367
Possible values:
B
BayoNet 已提交
368

369 370 371 372
-   `ALL` — If the right table has several matching rows, ClickHouse creates a [Cartesian product](https://en.wikipedia.org/wiki/Cartesian_product) from matching rows. This is the normal `JOIN` behaviour from standard SQL.
-   `ANY` — If the right table has several matching rows, only the first one found is joined. If the right table has only one matching row, the results of `ANY` and `ALL` are the same.
-   `ASOF` — For joining sequences with an uncertain match.
-   `Empty string` — If `ALL` or `ANY` is not specified in the query, ClickHouse throws an exception.
B
BayoNet 已提交
373

374
Default value: `ALL`.
B
BayoNet 已提交
375

376
## join\_any\_take\_last\_row {#settings-join_any_take_last_row}
377

378
Changes behaviour of join operations with `ANY` strictness.
379

380
!!! warning "Attention"
381
    This setting applies only for `JOIN` operations with [Join](../../engines/table-engines/special/join.md) engine tables.
382 383 384

Possible values:

385 386
-   0 — If the right table has more than one matching row, only the first one found is joined.
-   1 — If the right table has more than one matching row, only the last one found is joined.
387

B
BayoNet 已提交
388
Default value: 0.
389

S
Fixes  
Sergei Bocharov 已提交
390
See also:
391

392
-   [JOIN clause](../../sql-reference/statements/select/join.md#select-join)
393
-   [Join table engine](../../engines/table-engines/special/join.md)
394
-   [join\_default\_strictness](#settings-join_default_strictness)
B
BayoNet 已提交
395

396
## join\_use\_nulls {#join_use_nulls}
397

398
Sets the type of [JOIN](../../sql-reference/statements/select/join.md) behavior. When merging tables, empty cells may appear. ClickHouse fills them differently based on this setting.
399

400
Possible values:
401

402
-   0 — The empty cells are filled with the default value of the corresponding field type.
403
-   1 — `JOIN` behaves the same way as in standard SQL. The type of the corresponding field is converted to [Nullable](../../sql-reference/data-types/nullable.md#data_type-nullable), and empty cells are filled with [NULL](../../sql-reference/syntax.md).
404

405 406
Default value: 0.

407
## partial\_merge\_join\_optimizations {#partial_merge_join_optimizations}
408 409 410

Disables optimizations in partial merge join algorithm for [JOIN](../../sql-reference/statements/select/join.md) queries.

411
By default, this setting enables improvements that could lead to wrong results. If you see suspicious results in your queries, disable optimizations by this setting. Optimizations can be different in different versions of the ClickHouse server.
412 413 414 415 416 417 418 419

Possible values:

-   0 — Optimizations disabled.
-   1 — Optimizations enabled.

Default value: 1.

420
## partial\_merge\_join\_rows\_in\_right\_blocks {#partial_merge_join_rows_in_right_blocks}
421 422 423 424 425

Limits sizes of right-hand join data blocks in partial merge join algorithm for [JOIN](../../sql-reference/statements/select/join.md) queries.

ClickHouse server:

426 427 428
1.  Splits right-hand join data into blocks with up to the specified number of rows.
2.  Indexes each block with their minimum and maximum values
3.  Unloads prepared blocks to disk if possible.
429 430 431

Possible values:

432
-   Any positive integer. Recommended range of values: \[1000, 100000\].
433 434 435

Default value: 65536.

436
## join\_on\_disk\_max\_files\_to\_merge {#join_on_disk_max_files_to_merge}
437

438
Limits the number of files allowed for parallel sorting in MergeJoin operations when they are executed on disk.
439 440 441 442 443

The bigger the value of the setting, the more RAM used and the less disk I/O needed.

Possible values:

444
-   Any positive integer, starting from 2.
445 446 447

Default value: 64.

448
## any\_join\_distinct\_right\_table\_keys {#any_join_distinct_right_table_keys}
449 450 451 452 453 454 455 456

Enables legacy ClickHouse server behavior in `ANY INNER|LEFT JOIN` operations.

!!! note "Warning"
    Use this setting only for the purpose of backward compatibility if your use cases depend on legacy `JOIN` behavior.

When the legacy behavior enabled:

457 458
-   Results of `t1 ANY LEFT JOIN t2` and `t2 ANY RIGHT JOIN t1` operations are not equal because ClickHouse uses the logic with many-to-one left-to-right table keys mapping.
-   Results of `ANY INNER JOIN` operations contain all rows from the left table like the `SEMI LEFT JOIN` operations do.
459 460 461

When the legacy behavior disabled:

462 463
-   Results of `t1 ANY LEFT JOIN t2` and `t2 ANY RIGHT JOIN t1` operations are equal because ClickHouse uses the logic which provides one-to-many keys mapping in `ANY RIGHT JOIN` operations.
-   Results of `ANY INNER JOIN` operations contain one row per key from both left and right tables.
464 465 466

Possible values:

467 468
-   0 — Legacy behavior is disabled.
-   1 — Legacy behavior is enabled.
469 470 471 472 473 474 475

Default value: 0.

See also:

-   [JOIN strictness](../../sql-reference/statements/select/join.md#select-join-strictness)

476
## temporary\_files\_codec {#temporary_files_codec}
477 478 479

Sets compression codec for temporary files used in sorting and joining operations on disk.

480
Possible values:
481

482 483
-   LZ4 — [LZ4](https://en.wikipedia.org/wiki/LZ4_(compression_algorithm)) compression is applied.
-   NONE — No compression is applied.
484 485 486

Default value: LZ4.

487
## max\_block\_size {#setting-max_block_size}
488

489
In ClickHouse, data is processed by blocks (sets of column parts). The internal processing cycles for a single block are efficient enough, but there are noticeable expenditures on each block. The `max_block_size` setting is a recommendation for what size of the block (in a count of rows) to load from tables. The block size shouldn’t be too small, so that the expenditures on each block are still noticeable, but not too large so that the query with LIMIT that is completed after the first block is processed quickly. The goal is to avoid consuming too much memory when extracting a large number of columns in multiple threads and to preserve at least some cache locality.
490

491
Default value: 65,536.
492 493 494

Blocks the size of `max_block_size` are not always loaded from the table. If it is obvious that less data needs to be retrieved, a smaller block is processed.

495
## preferred\_block\_size\_bytes {#preferred-block-size-bytes}
496 497 498

Used for the same purpose as `max_block_size`, but it sets the recommended block size in bytes by adapting it to the number of rows in the block.
However, the block size cannot be more than `max_block_size` rows.
499
By default: 1,000,000. It only works when reading from MergeTree engines.
500

501
## merge\_tree\_min\_rows\_for\_concurrent\_read {#setting-merge-tree-min-rows-for-concurrent-read}
B
BayoNet 已提交
502

503
If the number of rows to be read from a file of a [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md) table exceeds `merge_tree_min_rows_for_concurrent_read` then ClickHouse tries to perform a concurrent reading from this file on several threads.
B
BayoNet 已提交
504

505
Possible values:
B
BayoNet 已提交
506

507
-   Any positive integer.
508 509 510

Default value: 163840.

511
## merge\_tree\_min\_bytes\_for\_concurrent\_read {#setting-merge-tree-min-bytes-for-concurrent-read}
512

513
If the number of bytes to read from one file of a [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md)-engine table exceeds `merge_tree_min_bytes_for_concurrent_read`, then ClickHouse tries to concurrently read from this file in several threads.
514

515
Possible value:
516

517
-   Any positive integer.
B
BayoNet 已提交
518

519
Default value: 251658240.
B
BayoNet 已提交
520

521
## merge\_tree\_min\_rows\_for\_seek {#setting-merge-tree-min-rows-for-seek}
B
BayoNet 已提交
522

523
If the distance between two data blocks to be read in one file is less than `merge_tree_min_rows_for_seek` rows, then ClickHouse does not seek through the file but reads the data sequentially.
B
BayoNet 已提交
524

525
Possible values:
B
BayoNet 已提交
526

527
-   Any positive integer.
528 529 530

Default value: 0.

531
## merge\_tree\_min\_bytes\_for\_seek {#setting-merge-tree-min-bytes-for-seek}
532

533
If the distance between two data blocks to be read in one file is less than `merge_tree_min_bytes_for_seek` bytes, then ClickHouse sequentially reads a range of file that contains both blocks, thus avoiding extra seek.
534 535 536

Possible values:

537
-   Any positive integer.
538 539

Default value: 0.
B
BayoNet 已提交
540

541
## merge\_tree\_coarse\_index\_granularity {#setting-merge-tree-coarse-index-granularity}
B
BayoNet 已提交
542

543
When searching for data, ClickHouse checks the data marks in the index file. If ClickHouse finds that required keys are in some range, it divides this range into `merge_tree_coarse_index_granularity` subranges and searches the required keys there recursively.
B
BayoNet 已提交
544

545
Possible values:
B
BayoNet 已提交
546

547
-   Any positive even integer.
B
BayoNet 已提交
548

549
Default value: 8.
B
BayoNet 已提交
550

551
## merge\_tree\_max\_rows\_to\_use\_cache {#setting-merge-tree-max-rows-to-use-cache}
B
BayoNet 已提交
552

553
If ClickHouse should read more than `merge_tree_max_rows_to_use_cache` rows in one query, it doesn’t use the cache of uncompressed blocks.
B
BayoNet 已提交
554

555
The cache of uncompressed blocks stores data extracted for queries. ClickHouse uses this cache to speed up responses to repeated small queries. This setting protects the cache from trashing by queries that read a large amount of data. The [uncompressed\_cache\_size](../../operations/server-configuration-parameters/settings.md#server-settings-uncompressed_cache_size) server setting defines the size of the cache of uncompressed blocks.
556 557 558

Possible values:

559
-   Any positive integer.
560 561

Default value: 128 ✕ 8192.
B
BayoNet 已提交
562

563
## merge\_tree\_max\_bytes\_to\_use\_cache {#setting-merge-tree-max-bytes-to-use-cache}
564

565
If ClickHouse should read more than `merge_tree_max_bytes_to_use_cache` bytes in one query, it doesn’t use the cache of uncompressed blocks.
566

567
The cache of uncompressed blocks stores data extracted for queries. ClickHouse uses this cache to speed up responses to repeated small queries. This setting protects the cache from trashing by queries that read a large amount of data. The [uncompressed\_cache\_size](../../operations/server-configuration-parameters/settings.md#server-settings-uncompressed_cache_size) server setting defines the size of the cache of uncompressed blocks.
568

569
Possible value:
570

571
-   Any positive integer.
B
BayoNet 已提交
572

573
Default value: 2013265920.
B
BayoNet 已提交
574

575
## min\_bytes\_to\_use\_direct\_io {#settings-min-bytes-to-use-direct-io}
576

577
The minimum data volume required for using direct I/O access to the storage disk.
578

579
ClickHouse uses this setting when reading data from tables. If the total storage volume of all the data to be read exceeds `min_bytes_to_use_direct_io` bytes, then ClickHouse reads the data from the storage disk with the `O_DIRECT` option.
580

S
Fixes  
Sergei Bocharov 已提交
581
Possible values:
582

583 584
-   0 — Direct I/O is disabled.
-   Positive integer.
585

S
Fixes  
Sergei Bocharov 已提交
586
Default value: 0.
587

588
## log\_queries {#settings-log-queries}
589

590
Setting up query logging.
591

592
Queries sent to ClickHouse with this setup are logged according to the rules in the [query\_log](../../operations/server-configuration-parameters/settings.md#server_configuration_parameters-query-log) server configuration parameter.
593

S
Fixes  
Sergei Bocharov 已提交
594
Example:
595

596
``` text
S
Fixes  
Sergei Bocharov 已提交
597 598
log_queries=1
```
599

600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617
## log\_queries\_min\_type {#settings-log-queries-min-type}

`query_log` minimal type to log.

Possible values:
- `QUERY_START` (`=1`)
- `QUERY_FINISH` (`=2`)
- `EXCEPTION_BEFORE_START` (`=3`)
- `EXCEPTION_WHILE_PROCESSING` (`=4`)

Default value: `QUERY_START`.

Can be used to limit which entiries will goes to `query_log`, say you are interesting only in errors, then you can use `EXCEPTION_WHILE_PROCESSING`:

``` text
log_queries_min_type='EXCEPTION_WHILE_PROCESSING'
```

618
## log\_query\_threads {#settings-log-query-threads}
619 620 621

Setting up query threads logging.

622
Queries’ threads runned by ClickHouse with this setup are logged according to the rules in the [query\_thread\_log](../../operations/server-configuration-parameters/settings.md#server_configuration_parameters-query_thread_log) server configuration parameter.
623

S
Fixes  
Sergei Bocharov 已提交
624
Example:
625

626
``` text
S
Fixes  
Sergei Bocharov 已提交
627 628
log_query_threads=1
```
629

630
## max\_insert\_block\_size {#settings-max_insert_block_size}
631 632 633 634

The size of blocks to form for insertion into a table.
This setting only applies in cases when the server forms the blocks.
For example, for an INSERT via the HTTP interface, the server parses the data format and forms blocks of the specified size.
635 636
But when using clickhouse-client, the client parses the data itself, and the ‘max\_insert\_block\_size’ setting on the server doesn’t affect the size of the inserted blocks.
The setting also doesn’t have a purpose when using INSERT SELECT, since data is inserted using the same blocks that are formed after SELECT.
637

638
Default value: 1,048,576.
639

640
The default is slightly more than `max_block_size`. The reason for this is because certain table engines (`*MergeTree`) form a data part on the disk for each inserted block, which is a fairly large entity. Similarly, `*MergeTree` tables sort data during insertion and a large enough block size allow sorting more data in RAM.
641

642
## min\_insert\_block\_size\_rows {#min-insert-block-size-rows}
643 644 645 646 647

Sets minimum number of rows in block which can be inserted into a table by an `INSERT` query. Smaller-sized blocks are squashed into bigger ones.

Possible values:

648 649
-   Positive integer.
-   0 — Squashing disabled.
650 651 652

Default value: 1048576.

653
## min\_insert\_block\_size\_bytes {#min-insert-block-size-bytes}
654 655 656 657 658

Sets minimum number of bytes in block which can be inserted into a table by an `INSERT` query. Smaller-sized blocks are squashed into bigger ones.

Possible values:

659 660
-   Positive integer.
-   0 — Squashing disabled.
661 662 663

Default value: 268435456.

664
## max\_replica\_delay\_for\_distributed\_queries {#settings-max_replica_delay_for_distributed_queries}
665

666
Disables lagging replicas for distributed queries. See [Replication](../../engines/table-engines/mergetree-family/replication.md).
667 668 669

Sets the time in seconds. If a replica lags more than the set value, this replica is not used.

670
Default value: 300.
671

672
Used when performing `SELECT` from a distributed table that points to replicated tables.
673

674
## max\_threads {#settings-max_threads}
675

676
The maximum number of query processing threads, excluding threads for retrieving data from remote servers (see the ‘max\_distributed\_connections’ parameter).
677 678

This parameter applies to threads that perform the same stages of the query processing pipeline in parallel.
679
For example, when reading from a table, if it is possible to evaluate expressions with functions, filter with WHERE and pre-aggregate for GROUP BY in parallel using at least ‘max\_threads’ number of threads, then ‘max\_threads’ are used.
680

A
alexey-milovidov 已提交
681
Default value: the number of physical CPU cores.
682 683 684

If less than one SELECT query is normally run on a server at a time, set this parameter to a value slightly less than the actual number of processor cores.

685
For queries that are completed quickly because of a LIMIT, you can set a lower ‘max\_threads’. For example, if the necessary number of entries are located in every block and max\_threads = 8, then 8 blocks are retrieved, although it would have been enough to read just one.
686 687 688

The smaller the `max_threads` value, the less memory is consumed.

689
## max\_insert\_threads {#settings-max-insert-threads}
D
Denis Zhuravlev 已提交
690 691 692 693 694

The maximum number of threads to execute the `INSERT SELECT` query.

Possible values:

695 696
-   0 (or 1) — `INSERT SELECT` no parallel execution.
-   Positive integer. Bigger than 1.
D
Denis Zhuravlev 已提交
697 698 699

Default value: 0.

700
Parallel `INSERT SELECT` has effect only if the `SELECT` part is executed in parallel, see [max\_threads](#settings-max_threads) setting.
D
Denis Zhuravlev 已提交
701 702
Higher values will lead to higher memory usage.

703
## max\_compress\_block\_size {#max-compress-block-size}
704

705
The maximum size of blocks of uncompressed data before compressing for writing to a table. By default, 1,048,576 (1 MiB). If the size is reduced, the compression rate is significantly reduced, the compression and decompression speed increases slightly due to cache locality, and memory consumption is reduced. There usually isn’t any reason to change this setting.
706

707
Don’t confuse blocks for compression (a chunk of memory consisting of bytes) with blocks for query processing (a set of rows from a table).
708

709
## min\_compress\_block\_size {#min-compress-block-size}
710

711
For [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md)" tables. In order to reduce latency when processing queries, a block is compressed when writing the next mark if its size is at least ‘min\_compress\_block\_size’. By default, 65,536.
712

713
The actual size of the block, if the uncompressed data is less than ‘max\_compress\_block\_size’, is no less than this value and no less than the volume of data for one mark.
714

715
Let’s look at an example. Assume that ‘index\_granularity’ was set to 8192 during table creation.
716

717
We are writing a UInt32-type column (4 bytes per value). When writing 8192 rows, the total will be 32 KB of data. Since min\_compress\_block\_size = 65,536, a compressed block will be formed for every two marks.
718

719
We are writing a URL column with the String type (average size of 60 bytes per value). When writing 8192 rows, the average will be slightly less than 500 KB of data. Since this is more than 65,536, a compressed block will be formed for each mark. In this case, when reading data from the disk in the range of a single mark, extra data won’t be decompressed.
720

721
There usually isn’t any reason to change this setting.
722

723
## max\_query\_size {#settings-max_query_size}
724 725 726 727

The maximum part of a query that can be taken to RAM for parsing with the SQL parser.
The INSERT query also contains data for INSERT that is processed by a separate stream parser (that consumes O(1) RAM), which is not included in this restriction.

728
Default value: 256 KiB.
729

B
BayoNet 已提交
730 731 732 733 734 735
## max\_parser\_depth {#max_parser_depth}

Limits maximum recursion depth in the recursive descent parser. Allows to control stack size.

Possible values:

736 737
-   Positive integer.
-   0 — Recursion depth is unlimited.
B
BayoNet 已提交
738 739 740

Default value: 1000.

741
## interactive\_delay {#interactive-delay}
742

743
The interval in microseconds for checking whether request execution has been cancelled and sending the progress.
744

745
Default value: 100,000 (checks for cancelling and sends the progress ten times per second).
746

747
## connect\_timeout, receive\_timeout, send\_timeout {#connect-timeout-receive-timeout-send-timeout}
748 749 750

Timeouts in seconds on the socket used for communicating with the client.

751
Default value: 10, 300, 300.
752

753
## cancel\_http\_readonly\_queries\_on\_client\_close {#cancel-http-readonly-queries-on-client-close}
754

755
Cancels HTTP read-only queries (e.g. SELECT) when a client closes the connection without waiting for the response.
756 757 758

Default value: 0

759
## poll\_interval {#poll-interval}
760 761 762

Lock in a wait loop for the specified number of seconds.

763
Default value: 10.
764

765
## max\_distributed\_connections {#max-distributed-connections}
766 767 768

The maximum number of simultaneous connections with remote servers for distributed processing of a single query to a single Distributed table. We recommend setting a value no less than the number of servers in the cluster.

769
Default value: 1024.
770 771 772

The following parameters are only used when creating Distributed tables (and when launching a server), so there is no reason to change them at runtime.

773
## distributed\_connections\_pool\_size {#distributed-connections-pool-size}
774 775 776

The maximum number of simultaneous connections with remote servers for distributed processing of all queries to a single Distributed table. We recommend setting a value no less than the number of servers in the cluster.

777
Default value: 1024.
778

779
## connect\_timeout\_with\_failover\_ms {#connect-timeout-with-failover-ms}
780

781
The timeout in milliseconds for connecting to a remote server for a Distributed table engine, if the ‘shard’ and ‘replica’ sections are used in the cluster definition.
782 783
If unsuccessful, several attempts are made to connect to various replicas.

784
Default value: 50.
785

786
## connections\_with\_failover\_max\_tries {#connections-with-failover-max-tries}
787

788
The maximum number of connection attempts with each replica for the Distributed table engine.
789

790
Default value: 3.
791

792
## extremes {#extremes}
793 794

Whether to count extreme values (the minimums and maximums in columns of a query result). Accepts 0 or 1. By default, 0 (disabled).
795
For more information, see the section “Extreme values”.
796

797
## use\_uncompressed\_cache {#setting-use_uncompressed_cache}
798

799
Whether to use a cache of uncompressed blocks. Accepts 0 or 1. By default, 0 (disabled).
800
Using the uncompressed cache (only for tables in the MergeTree family) can significantly reduce latency and increase throughput when working with a large number of short queries. Enable this setting for users who send frequent short requests. Also pay attention to the [uncompressed\_cache\_size](../../operations/server-configuration-parameters/settings.md#server-settings-uncompressed_cache_size) configuration parameter (only set in the config file) – the size of uncompressed cache blocks. By default, it is 8 GiB. The uncompressed cache is filled in as needed and the least-used data is automatically deleted.
801

802
For queries that read at least a somewhat large volume of data (one million rows or more), the uncompressed cache is disabled automatically to save space for truly small queries. This means that you can keep the ‘use\_uncompressed\_cache’ setting always set to 1.
803

804
## replace\_running\_query {#replace-running-query}
805

806 807
When using the HTTP interface, the ‘query\_id’ parameter can be passed. This is any string that serves as the query identifier.
If a query from the same user with the same ‘query\_id’ already exists at this time, the behaviour depends on the ‘replace\_running\_query’ parameter.
808

809
`0` (default) – Throw an exception (don’t allow the query to run if a query with the same ‘query\_id’ is already running).
810 811 812

`1` – Cancel the old query and start running the new one.

813
Yandex.Metrica uses this parameter set to 1 for implementing suggestions for segmentation conditions. After entering the next character, if the old query hasn’t finished yet, it should be cancelled.
814

815
## stream\_flush\_interval\_ms {#stream-flush-interval-ms}
816

817
Works for tables with streaming in the case of a timeout, or when a thread generates [max\_insert\_block\_size](#settings-max_insert_block_size) rows.
818 819 820 821 822

The default value is 7500.

The smaller the value, the more often data is flushed into the table. Setting the value too low leads to poor performance.

823
## load\_balancing {#settings-load_balancing}
824

825
Specifies the algorithm of replicas selection that is used for distributed query processing.
826

827
ClickHouse supports the following algorithms of choosing replicas:
828

829 830 831 832
-   [Random](#load_balancing-random) (by default)
-   [Nearest hostname](#load_balancing-nearest_hostname)
-   [In order](#load_balancing-in_order)
-   [First or random](#load_balancing-first_or_random)
A
Azat Khuzhin 已提交
833
-   [Round robin](#load_balancing-round_robin)
834

835 836 837 838
See also:

-   [distributed\_replica\_max\_ignored\_errors](#settings-distributed_replica_max_ignored_errors)

I
Ivan Blinkov 已提交
839
### Random (by Default) {#load_balancing-random}
840

841
``` sql
842 843
load_balancing = random
```
844

845
The number of errors is counted for each replica. The query is sent to the replica with the fewest errors, and if there are several of these, to anyone of them.
846 847
Disadvantages: Server proximity is not accounted for; if the replicas have different data, you will also get different data.

848
### Nearest Hostname {#load_balancing-nearest_hostname}
849

850
``` sql
851 852
load_balancing = nearest_hostname
```
853

854
The number of errors is counted for each replica. Every 5 minutes, the number of errors is integrally divided by 2. Thus, the number of errors is calculated for a recent time with exponential smoothing. If there is one replica with a minimal number of errors (i.e. errors occurred recently on the other replicas), the query is sent to it. If there are multiple replicas with the same minimal number of errors, the query is sent to the replica with a hostname that is most similar to the server’s hostname in the config file (for the number of different characters in identical positions, up to the minimum length of both hostnames).
855 856

For instance, example01-01-1 and example01-01-2.yandex.ru are different in one position, while example01-01-1 and example01-02-2 differ in two places.
857
This method might seem primitive, but it doesn’t require external data about network topology, and it doesn’t compare IP addresses, which would be complicated for our IPv6 addresses.
858 859 860 861

Thus, if there are equivalent replicas, the closest one by name is preferred.
We can also assume that when sending a query to the same server, in the absence of failures, a distributed query will also go to the same servers. So even if different data is placed on the replicas, the query will return mostly the same results.

862
### In Order {#load_balancing-in_order}
863

864
``` sql
865 866
load_balancing = in_order
```
867

868
Replicas with the same number of errors are accessed in the same order as they are specified in the configuration.
869 870
This method is appropriate when you know exactly which replica is preferable.

871
### First or Random {#load_balancing-first_or_random}
872

873
``` sql
874 875 876
load_balancing = first_or_random
```

877
This algorithm chooses the first replica in the set or a random replica if the first is unavailable. It’s effective in cross-replication topology setups, but useless in other configurations.
878

879
The `first_or_random` algorithm solves the problem of the `in_order` algorithm. With `in_order`, if one replica goes down, the next one gets a double load while the remaining replicas handle the usual amount of traffic. When using the `first_or_random` algorithm, the load is evenly distributed among replicas that are still available.
880

881
### Round Robin {#load_balancing-round_robin}
A
Azat Khuzhin 已提交
882 883 884 885 886 887 888

``` sql
load_balancing = round_robin
```

This algorithm uses round robin policy across replicas with the same number of errors (only the queries with `round_robin` policy is accounted).

889
## prefer\_localhost\_replica {#settings-prefer-localhost-replica}
890 891 892 893 894

Enables/disables preferable using the localhost replica when processing distributed queries.

Possible values:

895 896
-   1 — ClickHouse always sends a query to the localhost replica if it exists.
-   0 — ClickHouse uses the balancing strategy specified by the [load\_balancing](#settings-load_balancing) setting.
897 898 899

Default value: 1.

900
!!! warning "Warning"
901
    Disable this setting if you use [max\_parallel\_replicas](#settings-max_parallel_replicas).
902

903
## totals\_mode {#totals-mode}
904

905 906
How to calculate TOTALS when HAVING is present, as well as when max\_rows\_to\_group\_by and group\_by\_overflow\_mode = ‘any’ are present.
See the section “WITH TOTALS modifier”.
907

908
## totals\_auto\_threshold {#totals-auto-threshold}
909

910
The threshold for `totals_mode = 'auto'`.
911
See the section “WITH TOTALS modifier”.
912

913
## max\_parallel\_replicas {#settings-max_parallel_replicas}
914 915 916 917 918

The maximum number of replicas for each shard when executing a query.
For consistency (to get different parts of the same data split), this option only works when the sampling key is set.
Replica lag is not controlled.

919
## compile {#compile}
920 921 922

Enable compilation of queries. By default, 0 (disabled).

923
The compilation is only used for part of the query-processing pipeline: for the first stage of aggregation (GROUP BY).
924
If this portion of the pipeline was compiled, the query may run faster due to deployment of short cycles and inlining aggregate function calls. The maximum performance improvement (up to four times faster in rare cases) is seen for queries with multiple simple aggregate functions. Typically, the performance gain is insignificant. In very rare cases, it may slow down query execution.
925

926
## min\_count\_to\_compile {#min-count-to-compile}
927 928

How many times to potentially use a compiled chunk of code before running compilation. By default, 3.
929 930
For testing, the value can be set to 0: compilation runs synchronously and the query waits for the end of the compilation process before continuing execution. For all other cases, use values ​​starting with 1. Compilation normally takes about 5-10 seconds.
If the value is 1 or more, compilation occurs asynchronously in a separate thread. The result will be used as soon as it is ready, including queries that are currently running.
931 932

Compiled code is required for each different combination of aggregate functions used in the query and the type of keys in the GROUP BY clause.
933
The results of the compilation are saved in the build directory in the form of .so files. There is no restriction on the number of compilation results since they don’t use very much space. Old results will be used after server restarts, except in the case of a server upgrade – in this case, the old results are deleted.
934

935
## output\_format\_json\_quote\_64bit\_integers {#session_settings-output_format_json_quote_64bit_integers}
936

937
If the value is true, integers appear in quotes when using JSON\* Int64 and UInt64 formats (for compatibility with most JavaScript implementations); otherwise, integers are output without the quotes.
I
Ivan Zhukov 已提交
938

939
## format\_csv\_delimiter {#settings-format_csv_delimiter}
I
Ivan Zhukov 已提交
940

941 942
The character interpreted as a delimiter in the CSV data. By default, the delimiter is `,`.

943
## input\_format\_csv\_unquoted\_null\_literal\_as\_null {#settings-input_format_csv_unquoted_null_literal_as_null}
A
Alexander Tokmakov 已提交
944 945 946

For CSV input format enables or disables parsing of unquoted `NULL` as literal (synonym for `\N`).

947
## output\_format\_csv\_crlf\_end\_of\_line {#settings-output-format-csv-crlf-end-of-line}
M
Mikhail Korotov 已提交
948

949
Use DOS/Windows-style line separator (CRLF) in CSV instead of Unix style (LF).
M
Mikhail Korotov 已提交
950

951
## output\_format\_tsv\_crlf\_end\_of\_line {#settings-output-format-tsv-crlf-end-of-line}
M
millb 已提交
952

953
Use DOC/Windows-style line separator (CRLF) in TSV instead of Unix style (LF).
M
millb 已提交
954

955
## insert\_quorum {#settings-insert_quorum}
956

957
Enables the quorum writes.
958

959 960
-   If `insert_quorum < 2`, the quorum writes are disabled.
-   If `insert_quorum >= 2`, the quorum writes are enabled.
961

962
Default value: 0.
963

S
Fixes  
Sergei Bocharov 已提交
964
Quorum writes
965 966 967 968 969

`INSERT` succeeds only when ClickHouse manages to correctly write data to the `insert_quorum` of replicas during the `insert_quorum_timeout`. If for any reason the number of replicas with successful writes does not reach the `insert_quorum`, the write is considered failed and ClickHouse will delete the inserted block from all the replicas where data has already been written.

All the replicas in the quorum are consistent, i.e., they contain data from all previous `INSERT` queries. The `INSERT` sequence is linearized.

970
When reading the data written from the `insert_quorum`, you can use the [select\_sequential\_consistency](#settings-select_sequential_consistency) option.
971

S
Fixes  
Sergei Bocharov 已提交
972
ClickHouse generates an exception
973

974 975
-   If the number of available replicas at the time of the query is less than the `insert_quorum`.
-   At an attempt to write data when the previous block has not yet been inserted in the `insert_quorum` of replicas. This situation may occur if the user tries to perform an `INSERT` before the previous one with the `insert_quorum` is completed.
976

S
Fixes  
Sergei Bocharov 已提交
977
See also:
978

979 980
-   [insert\_quorum\_timeout](#settings-insert_quorum_timeout)
-   [select\_sequential\_consistency](#settings-select_sequential_consistency)
981

982
## insert\_quorum\_timeout {#settings-insert_quorum_timeout}
983

984
Write to quorum timeout in seconds. If the timeout has passed and no write has taken place yet, ClickHouse will generate an exception and the client must repeat the query to write the same block to the same or any other replica.
985

986
Default value: 60 seconds.
987

S
Fixes  
Sergei Bocharov 已提交
988
See also:
989

990 991
-   [insert\_quorum](#settings-insert_quorum)
-   [select\_sequential\_consistency](#settings-select_sequential_consistency)
992

993
## select\_sequential\_consistency {#settings-select_sequential_consistency}
994

995 996 997
Enables or disables sequential consistency for `SELECT` queries:

Possible values:
998

999 1000
-   0 — Disabled.
-   1 — Enabled.
1001

1002
Default value: 0.
1003

S
Fixes  
Sergei Bocharov 已提交
1004
Usage
1005

1006 1007
When sequential consistency is enabled, ClickHouse allows the client to execute the `SELECT` query only for those replicas that contain data from all previous `INSERT` queries executed with `insert_quorum`. If the client refers to a partial replica, ClickHouse will generate an exception. The SELECT query will not include data that has not yet been written to the quorum of replicas.

S
Fixes  
Sergei Bocharov 已提交
1008
See also:
1009

1010 1011
-   [insert\_quorum](#settings-insert_quorum)
-   [insert\_quorum\_timeout](#settings-insert_quorum_timeout)
1012

1013
## insert\_deduplicate {#settings-insert-deduplicate}
D
Denis Zhuravlev 已提交
1014

1015
Enables or disables block deduplication of `INSERT` (for Replicated\* tables).
D
Denis Zhuravlev 已提交
1016 1017 1018

Possible values:

1019 1020
-   0 — Disabled.
-   1 — Enabled.
D
Denis Zhuravlev 已提交
1021 1022 1023

Default value: 1.

F
Fan() 已提交
1024 1025
By default, blocks inserted into replicated tables by the `INSERT` statement are deduplicated (see [Data Replication](../../engines/table-engines/mergetree-family/replication.md)).

1026
## deduplicate\_blocks\_in\_dependent\_materialized\_views {#settings-deduplicate-blocks-in-dependent-materialized-views}
D
Denis Zhuravlev 已提交
1027

1028
Enables or disables the deduplication check for materialized views that receive data from Replicated\* tables.
D
Denis Zhuravlev 已提交
1029 1030 1031

Possible values:

1032 1033
      0 — Disabled.
      1 — Enabled.
D
Denis Zhuravlev 已提交
1034 1035 1036 1037 1038

Default value: 0.

Usage

1039 1040
By default, deduplication is not performed for materialized views but is done upstream, in the source table.
If an INSERTed block is skipped due to deduplication in the source table, there will be no insertion into attached materialized views. This behaviour exists to enable insertion of highly aggregated data into materialized views, for cases where inserted blocks are the same after materialized view aggregation but derived from different INSERTs into the source table.
1041
At the same time, this behaviour “breaks” `INSERT` idempotency. If an `INSERT` into the main table was successful and `INSERT` into a materialized view failed (e.g. because of communication failure with Zookeeper) a client will get an error and can retry the operation. However, the materialized view won’t receive the second insert because it will be discarded by deduplication in the main (source) table. The setting `deduplicate_blocks_in_dependent_materialized_views` allows for changing this behaviour. On retry, a materialized view will receive the repeat insert and will perform deduplication check by itself,
1042
ignoring check result for the source table, and will insert rows lost because of the first failure.
D
Denis Zhuravlev 已提交
1043

1044
## max\_network\_bytes {#settings-max-network-bytes}
1045

1046
Limits the data volume (in bytes) that is received or transmitted over the network when executing a query. This setting applies to every individual query.
1047 1048 1049

Possible values:

1050 1051
-   Positive integer.
-   0 — Data volume control is disabled.
1052 1053 1054

Default value: 0.

1055
## max\_network\_bandwidth {#settings-max-network-bandwidth}
1056

1057
Limits the speed of the data exchange over the network in bytes per second. This setting applies to every query.
1058 1059 1060

Possible values:

1061 1062
-   Positive integer.
-   0 — Bandwidth control is disabled.
1063 1064 1065

Default value: 0.

1066
## max\_network\_bandwidth\_for\_user {#settings-max-network-bandwidth-for-user}
1067

1068
Limits the speed of the data exchange over the network in bytes per second. This setting applies to all concurrently running queries performed by a single user.
1069 1070 1071

Possible values:

1072 1073
-   Positive integer.
-   0 — Control of the data speed is disabled.
1074 1075 1076

Default value: 0.

1077
## max\_network\_bandwidth\_for\_all\_users {#settings-max-network-bandwidth-for-all-users}
1078

1079
Limits the speed that data is exchanged at over the network in bytes per second. This setting applies to all concurrently running queries on the server.
1080 1081 1082

Possible values:

1083 1084
-   Positive integer.
-   0 — Control of the data speed is disabled.
1085 1086 1087

Default value: 0.

1088
## count\_distinct\_implementation {#settings-count_distinct_implementation}
1089

1090
Specifies which of the `uniq*` functions should be used to perform the [COUNT(DISTINCT …)](../../sql-reference/aggregate-functions/reference/count.md#agg_function-count) construction.
1091 1092 1093

Possible values:

1094 1095 1096 1097 1098
-   [uniq](../../sql-reference/aggregate-functions/reference/uniq.md#agg_function-uniq)
-   [uniqCombined](../../sql-reference/aggregate-functions/reference/uniqcombined.md#agg_function-uniqcombined)
-   [uniqCombined64](../../sql-reference/aggregate-functions/reference/uniqcombined64.md#agg_function-uniqcombined64)
-   [uniqHLL12](../../sql-reference/aggregate-functions/reference/uniqhll12.md#agg_function-uniqhll12)
-   [uniqExact](../../sql-reference/aggregate-functions/reference/uniqexact.md#agg_function-uniqexact)
1099 1100

Default value: `uniqExact`.
I
Ivan Blinkov 已提交
1101

1102
## skip\_unavailable\_shards {#settings-skip_unavailable_shards}
1103

1104
Enables or disables silently skipping of unavailable shards.
1105

1106
Shard is considered unavailable if all its replicas are unavailable. A replica is unavailable in the following cases:
1107

1108
-   ClickHouse can’t connect to replica for any reason.
1109

1110
    When connecting to a replica, ClickHouse performs several attempts. If all these attempts fail, the replica is considered unavailable.
1111

1112
-   Replica can’t be resolved through DNS.
1113

1114
    If replica’s hostname can’t be resolved through DNS, it can indicate the following situations:
1115

1116
    -   Replica’s host has no DNS record. It can occur in systems with dynamic DNS, for example, [Kubernetes](https://kubernetes.io), where nodes can be unresolvable during downtime, and this is not an error.
1117

1118
    -   Configuration error. ClickHouse configuration file contains a wrong hostname.
1119 1120 1121

Possible values:

1122
-   1 — skipping enabled.
1123

1124
    If a shard is unavailable, ClickHouse returns a result based on partial data and doesn’t report node availability issues.
1125

1126
-   0 — skipping disabled.
1127

1128
    If a shard is unavailable, ClickHouse throws an exception.
1129

1130 1131
Default value: 0.

1132
## optimize\_skip\_unused\_shards {#optimize-skip-unused-shards}
1133

1134 1135 1136 1137
Enables or disables skipping of unused shards for [SELECT](../../sql-reference/statements/select/index.md) queries that have sharding key condition in `WHERE/PREWHERE` (assuming that the data is distributed by sharding key, otherwise does nothing).

Possible values:

1138 1139
-   0 — Disabled.
-   1 — Enabled.
1140 1141 1142

Default value: 0

1143 1144
## optimize\_skip\_unused\_shards\_nesting {#optimize-skip-unused-shards-nesting}

A
Azat Khuzhin 已提交
1145
Controls [`optimize_skip_unused_shards`](#optimize-skip-unused-shards) (hence still requires [`optimize_skip_unused_shards`](#optimize-skip-unused-shards)) depends on the nesting level of the distributed query (case when you have `Distributed` table that look into another `Distributed` table).
1146 1147 1148

Possible values:

A
Azat Khuzhin 已提交
1149 1150 1151
-   0 — Disabled, `optimize_skip_unused_shards` works always.
-   1 — Enables `optimize_skip_unused_shards` only for the first level.
-   2 — Enables `optimize_skip_unused_shards` up to the second level.
1152 1153 1154

Default value: 0

1155
## force\_optimize\_skip\_unused\_shards {#force-optimize-skip-unused-shards}
1156

1157
Enables or disables query execution if [optimize\_skip\_unused\_shards](#optimize-skip-unused-shards) is enabled and skipping of unused shards is not possible. If the skipping is not possible and the setting is enabled, an exception will be thrown.
1158 1159 1160

Possible values:

1161
-   0 — Disabled. ClickHouse doesn’t throw an exception.
1162 1163
-   1 — Enabled. Query execution is disabled only if the table has a sharding key.
-   2 — Enabled. Query execution is disabled regardless of whether a sharding key is defined for the table.
1164 1165 1166

Default value: 0

1167
## force\_optimize\_skip\_unused\_shards\_nesting {#settings-force_optimize_skip_unused_shards_nesting}
1168

A
Azat Khuzhin 已提交
1169
Controls [`force_optimize_skip_unused_shards`](#force-optimize-skip-unused-shards) (hence still requires [`force_optimize_skip_unused_shards`](#force-optimize-skip-unused-shards)) depends on the nesting level of the distributed query (case when you have `Distributed` table that look into another `Distributed` table).
1170 1171 1172

Possible values:

A
Azat Khuzhin 已提交
1173 1174 1175
-   0 - Disabled, `force_optimize_skip_unused_shards` works always.
-   1 — Enables `force_optimize_skip_unused_shards` only for the first level.
-   2 — Enables `force_optimize_skip_unused_shards` up to the second level.
1176

A
Azat Khuzhin 已提交
1177
Default value: 0
1178

1179
## optimize\_throw\_if\_noop {#setting-optimize_throw_if_noop}
1180

1181
Enables or disables throwing an exception if an [OPTIMIZE](../../sql-reference/statements/misc.md#misc_operations-optimize) query didn’t perform a merge.
1182

1183
By default, `OPTIMIZE` returns successfully even if it didn’t do anything. This setting lets you differentiate these situations and get the reason in an exception message.
1184 1185 1186

Possible values:

1187 1188
-   1 — Throwing an exception is enabled.
-   0 — Throwing an exception is disabled.
1189 1190

Default value: 0.
1191

1192
## distributed\_replica\_error\_half\_life {#settings-distributed_replica_error_half_life}
V
Vasily Nemkov 已提交
1193

1194 1195
-   Type: seconds
-   Default value: 60 seconds
V
Vasily Nemkov 已提交
1196

1197
Controls how fast errors in distributed tables are zeroed. If a replica is unavailable for some time, accumulates 5 errors, and distributed\_replica\_error\_half\_life is set to 1 second, then the replica is considered normal 3 seconds after last error.
V
Vasily Nemkov 已提交
1198

S
Fixes  
Sergei Bocharov 已提交
1199
See also:
V
Vasily Nemkov 已提交
1200

1201
-   [load\_balancing](#load_balancing-round_robin)
1202
-   [Table engine Distributed](../../engines/table-engines/special/distributed.md)
1203
-   [distributed\_replica\_error\_cap](#settings-distributed_replica_error_cap)
1204
-   [distributed\_replica\_max\_ignored\_errors](#settings-distributed_replica_max_ignored_errors)
V
Vasily Nemkov 已提交
1205

1206
## distributed\_replica\_error\_cap {#settings-distributed_replica_error_cap}
V
Vasily Nemkov 已提交
1207

1208 1209
-   Type: unsigned int
-   Default value: 1000
V
Vasily Nemkov 已提交
1210

F
FeehanG 已提交
1211
Error count of each replica is capped at this value, preventing a single replica from accumulating too many errors.
V
Vasily Nemkov 已提交
1212

S
Fixes  
Sergei Bocharov 已提交
1213
See also:
V
Vasily Nemkov 已提交
1214

1215
-   [load\_balancing](#load_balancing-round_robin)
1216
-   [Table engine Distributed](../../engines/table-engines/special/distributed.md)
1217
-   [distributed\_replica\_error\_half\_life](#settings-distributed_replica_error_half_life)
1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232
-   [distributed\_replica\_max\_ignored\_errors](#settings-distributed_replica_max_ignored_errors)

## distributed\_replica\_max\_ignored\_errors {#settings-distributed_replica_max_ignored_errors}

-   Type: unsigned int
-   Default value: 0

Number of errors that will be ignored while choosing replicas (according to `load_balancing` algorithm).

See also:

-   [load\_balancing](#load_balancing-round_robin)
-   [Table engine Distributed](../../engines/table-engines/special/distributed.md)
-   [distributed\_replica\_error\_cap](#settings-distributed_replica_error_cap)
-   [distributed\_replica\_error\_half\_life](#settings-distributed_replica_error_half_life)
1233

1234
## distributed\_directory\_monitor\_sleep\_time\_ms {#distributed_directory_monitor_sleep_time_ms}
1235

1236
Base interval for the [Distributed](../../engines/table-engines/special/distributed.md) table engine to send data. The actual interval grows exponentially in the event of errors.
1237 1238 1239

Possible values:

1240
-   A positive integer number of milliseconds.
1241 1242 1243

Default value: 100 milliseconds.

1244
## distributed\_directory\_monitor\_max\_sleep\_time\_ms {#distributed_directory_monitor_max_sleep_time_ms}
1245

1246
Maximum interval for the [Distributed](../../engines/table-engines/special/distributed.md) table engine to send data. Limits exponential growth of the interval set in the [distributed\_directory\_monitor\_sleep\_time\_ms](#distributed_directory_monitor_sleep_time_ms) setting.
1247 1248 1249

Possible values:

1250
-   A positive integer number of milliseconds.
1251 1252 1253

Default value: 30000 milliseconds (30 seconds).

1254
## distributed\_directory\_monitor\_batch\_inserts {#distributed_directory_monitor_batch_inserts}
1255 1256 1257

Enables/disables sending of inserted data in batches.

1258
When batch sending is enabled, the [Distributed](../../engines/table-engines/special/distributed.md) table engine tries to send multiple files of inserted data in one operation instead of sending them separately. Batch sending improves cluster performance by better-utilizing server and network resources.
1259 1260 1261

Possible values:

1262 1263
-   1 — Enabled.
-   0 — Disabled.
1264

1265
Default value: 0.
1266

1267
## os\_thread\_priority {#setting-os-thread-priority}
1268

1269
Sets the priority ([nice](https://en.wikipedia.org/wiki/Nice_(Unix))) for threads that execute queries. The OS scheduler considers this priority when choosing the next thread to run on each available CPU core.
1270 1271

!!! warning "Warning"
1272
    To use this setting, you need to set the `CAP_SYS_NICE` capability. The `clickhouse-server` package sets it up during installation. Some virtual environments don’t allow you to set the `CAP_SYS_NICE` capability. In this case, `clickhouse-server` shows a message about it at the start.
1273 1274 1275

Possible values:

1276
-   You can set values in the range `[-20, 19]`.
1277

1278
Lower values mean higher priority. Threads with low `nice` priority values are executed more frequently than threads with high values. High values are preferable for long-running non-interactive queries because it allows them to quickly give up resources in favour of short interactive queries when they arrive.
1279 1280 1281

Default value: 0.

1282
## query\_profiler\_real\_time\_period\_ns {#query_profiler_real_time_period_ns}
1283

1284
Sets the period for a real clock timer of the [query profiler](../../operations/optimizing-performance/sampling-query-profiler.md). Real clock timer counts wall-clock time.
1285 1286 1287

Possible values:

1288
-   Positive integer number, in nanoseconds.
1289

1290
    Recommended values:
1291

1292 1293
            - 10000000 (100 times a second) nanoseconds and less for single queries.
            - 1000000000 (once a second) for cluster-wide profiling.
1294

1295
-   0 for turning off the timer.
1296

1297
Type: [UInt64](../../sql-reference/data-types/int-uint.md).
1298

1299
Default value: 1000000000 nanoseconds (once a second).
1300

S
Fixes  
Sergei Bocharov 已提交
1301
See also:
B
BayoNet 已提交
1302

1303
-   System table [trace\_log](../../operations/system-tables/trace_log.md#system_tables-trace_log)
1304

1305
## query\_profiler\_cpu\_time\_period\_ns {#query_profiler_cpu_time_period_ns}
1306

1307
Sets the period for a CPU clock timer of the [query profiler](../../operations/optimizing-performance/sampling-query-profiler.md). This timer counts only CPU time.
1308 1309 1310

Possible values:

1311
-   A positive integer number of nanoseconds.
1312

1313
    Recommended values:
1314

1315 1316
            - 10000000 (100 times a second) nanoseconds and more for single queries.
            - 1000000000 (once a second) for cluster-wide profiling.
1317

1318
-   0 for turning off the timer.
1319

1320
Type: [UInt64](../../sql-reference/data-types/int-uint.md).
1321 1322 1323

Default value: 1000000000 nanoseconds.

S
Fixes  
Sergei Bocharov 已提交
1324
See also:
B
BayoNet 已提交
1325

1326
-   System table [trace\_log](../../operations/system-tables/trace_log.md#system_tables-trace_log)
B
BayoNet 已提交
1327

1328
## allow\_introspection\_functions {#settings-allow_introspection_functions}
1329

1330
Enables of disables [introspections functions](../../sql-reference/functions/introspection.md) for query profiling.
1331 1332 1333

Possible values:

1334 1335
-   1 — Introspection functions enabled.
-   0 — Introspection functions disabled.
1336 1337 1338

Default value: 0.

1339 1340
**See Also**

1341 1342
-   [Sampling Query Profiler](../../operations/optimizing-performance/sampling-query-profiler.md)
-   System table [trace\_log](../../operations/system-tables/trace_log.md#system_tables-trace_log)
1343

1344
## input\_format\_parallel\_parsing {#input-format-parallel-parsing}
N
docs  
Nikita Mikhaylov 已提交
1345

1346 1347
-   Type: bool
-   Default value: True
N
docs  
Nikita Mikhaylov 已提交
1348

N
Nikita Mikhaylov 已提交
1349
Enable order-preserving parallel parsing of data formats. Supported only for TSV, TKSV, CSV and JSONEachRow formats.
N
docs  
Nikita Mikhaylov 已提交
1350

1351
## min\_chunk\_bytes\_for\_parallel\_parsing {#min-chunk-bytes-for-parallel-parsing}
N
docs  
Nikita Mikhaylov 已提交
1352

1353 1354
-   Type: unsigned int
-   Default value: 1 MiB
N
docs  
Nikita Mikhaylov 已提交
1355

A
Alexander Kuzmenkov 已提交
1356
The minimum chunk size in bytes, which each thread will parse in parallel.
1357

1358
## output\_format\_avro\_codec {#settings-output_format_avro_codec}
A
Andrew Onyshchuk 已提交
1359 1360 1361 1362 1363 1364 1365

Sets the compression codec used for output Avro file.

Type: string

Possible values:

1366 1367 1368
-   `null` — No compression
-   `deflate` — Compress with Deflate (zlib)
-   `snappy` — Compress with [Snappy](https://google.github.io/snappy/)
A
Andrew Onyshchuk 已提交
1369 1370 1371

Default value: `snappy` (if available) or `deflate`.

1372
## output\_format\_avro\_sync\_interval {#settings-output_format_avro_sync_interval}
A
Andrew Onyshchuk 已提交
1373 1374 1375 1376 1377 1378 1379 1380 1381

Sets minimum data size (in bytes) between synchronization markers for output Avro file.

Type: unsigned int

Possible values: 32 (32 bytes) - 1073741824 (1 GiB)

Default value: 32768 (32 KiB)

1382
## format\_avro\_schema\_registry\_url {#format_avro_schema_registry_url}
A
Andrew Onyshchuk 已提交
1383

1384
Sets [Confluent Schema Registry](https://docs.confluent.io/current/schema-registry/index.html) URL to use with [AvroConfluent](../../interfaces/formats.md#data-format-avro-confluent) format.
A
Andrew Onyshchuk 已提交
1385

1386
Default value: `Empty`.
A
Andrew Onyshchuk 已提交
1387

1388
## background\_pool\_size {#background_pool_size}
1389

1390 1391
Sets the number of threads performing background operations in table engines (for example, merges in [MergeTree engine](../../engines/table-engines/mergetree-family/index.md) tables). This setting is applied from `default` profile at ClickHouse server start and can’t be changed in a user session. By adjusting this setting, you manage CPU and disk load. Smaller pool size utilizes less CPU and disk resources, but background processes advance slower which might eventually impact query performance.

A
alesapin 已提交
1392
Before changing it, please also take a look at related [MergeTree settings](../../operations/server-configuration-parameters/settings.md#server_configuration_parameters-merge_tree), such as `number_of_free_entries_in_pool_to_lower_max_size_of_merge` and `number_of_free_entries_in_pool_to_execute_mutation`.
1393 1394 1395 1396 1397 1398 1399

Possible values:

-   Any positive integer.

Default value: 16.

1400
## background\_buffer\_flush\_schedule\_pool\_size {#background_buffer_flush_schedule_pool_size}
1401

1402
Sets the number of threads performing background flush in [Buffer](../../engines/table-engines/special/buffer.md)-engine tables. This setting is applied at ClickHouse server start and can’t be changed in a user session.
1403

1404 1405 1406 1407 1408 1409
Possible values:

-   Any positive integer.

Default value: 16.

1410
## background\_move\_pool\_size {#background_move_pool_size}
1411 1412 1413 1414 1415 1416 1417 1418 1419

Sets the number of threads performing background moves of data parts for [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md#table_engine-mergetree-multiple-volumes)-engine tables. This setting is applied at ClickHouse server start and can’t be changed in a user session.

Possible values:

-   Any positive integer.

Default value: 8.

1420
## background\_schedule\_pool\_size {#background_schedule_pool_size}
1421

1422
Sets the number of threads performing background tasks for [replicated](../../engines/table-engines/mergetree-family/replication.md) tables, [Kafka](../../engines/table-engines/integrations/kafka.md) streaming, [DNS cache updates](../../operations/server-configuration-parameters/settings.md#server-settings-dns-cache-update-period). This setting is applied at ClickHouse server start and can’t be changed in a user session.
1423 1424 1425 1426 1427 1428 1429

Possible values:

-   Any positive integer.

Default value: 16.

1430
## always\_fetch\_merged\_part {#always_fetch_merged_part}
1431

1432
Prohibits data parts merging in [Replicated\*MergeTree](../../engines/table-engines/mergetree-family/replication.md)-engine tables.
1433 1434 1435 1436 1437 1438

When merging is prohibited, the replica never merges parts and always downloads merged parts from other replicas. If there is no required data yet, the replica waits for it. CPU and disk load on the replica server decreases, but the network load on cluster increases. This setting can be useful on servers with relatively weak CPUs or slow disks, such as servers for backups storage.

Possible values:

-   0 — `Replicated*MergeTree`-engine tables merge data parts at the replica.
1439
-   1 — `Replicated*MergeTree`-engine tables don’t merge data parts at the replica. The tables download merged data parts from other replicas.
1440 1441 1442

Default value: 0.

1443
**See Also**
1444 1445 1446

-   [Data Replication](../../engines/table-engines/mergetree-family/replication.md)

1447
## background\_distributed\_schedule\_pool\_size {#background_distributed_schedule_pool_size}
1448 1449 1450 1451 1452 1453 1454

Sets the number of threads performing background tasks for [distributed](../../engines/table-engines/special/distributed.md) sends. This setting is applied at ClickHouse server start and can’t be changed in a user session.

Possible values:

-   Any positive integer.

1455 1456
Default value: 16.

1457
## transform\_null\_in {#transform_null_in}
B
BayoNet 已提交
1458 1459 1460

Enables equality of [NULL](../../sql-reference/syntax.md#null-literal) values for [IN](../../sql-reference/operators/in.md) operator.

1461
By default, `NULL` values can’t be compared because `NULL` means undefined value. Thus, comparison `expr = NULL` must always return `false`. With this setting `NULL = NULL` returns `true` for `IN` operator.
B
BayoNet 已提交
1462 1463 1464 1465 1466 1467 1468 1469

Possible values:

-   0 — Comparison of `NULL` values in `IN` operator returns `false`.
-   1 — Comparison of `NULL` values in `IN` operator returns `true`.

Default value: 0.

1470
**Example**
B
BayoNet 已提交
1471 1472 1473

Consider the `null_in` table:

1474
``` text
B
BayoNet 已提交
1475 1476 1477 1478 1479 1480 1481 1482 1483
┌──idx─┬─────i─┐
│    1 │     1 │
│    2 │  NULL │
│    3 │     3 │
└──────┴───────┘
```

Query:

1484
``` sql
B
BayoNet 已提交
1485 1486 1487 1488 1489
SELECT idx, i FROM null_in WHERE i IN (1, NULL) SETTINGS transform_null_in = 0;
```

Result:

1490
``` text
B
BayoNet 已提交
1491 1492 1493 1494 1495 1496 1497
┌──idx─┬────i─┐
│    1 │    1 │
└──────┴──────┘
```

Query:

1498
``` sql
B
BayoNet 已提交
1499 1500 1501 1502 1503
SELECT idx, i FROM null_in WHERE i IN (1, NULL) SETTINGS transform_null_in = 1;
```

Result:

1504
``` text
B
BayoNet 已提交
1505 1506 1507 1508 1509 1510
┌──idx─┬─────i─┐
│    1 │     1 │
│    2 │  NULL │
└──────┴───────┘
```

1511
**See Also**
B
BayoNet 已提交
1512 1513 1514

-   [NULL Processing in IN Operators](../../sql-reference/operators/in.md#in-null-processing)

1515
## low\_cardinality\_max\_dictionary\_size {#low_cardinality_max_dictionary_size}
B
BayoNet 已提交
1516

1517
Sets a maximum size in rows of a shared global dictionary for the [LowCardinality](../../sql-reference/data-types/lowcardinality.md) data type that can be written to a storage file system. This setting prevents issues with RAM in case of unlimited dictionary growth. All the data that can’t be encoded due to maximum dictionary size limitation ClickHouse writes in an ordinary method.
B
BayoNet 已提交
1518 1519 1520 1521 1522 1523 1524

Possible values:

-   Any positive integer.

Default value: 8192.

1525
## low\_cardinality\_use\_single\_dictionary\_for\_part {#low_cardinality_use_single_dictionary_for_part}
B
BayoNet 已提交
1526 1527 1528 1529 1530 1531 1532

Turns on or turns off using of single dictionary for the data part.

By default, ClickHouse server monitors the size of dictionaries and if a dictionary overflows then the server starts to write the next one. To prohibit creating several dictionaries set `low_cardinality_use_single_dictionary_for_part = 1`.

Possible values:

1533 1534
-   1 — Creating several dictionaries for the data part is prohibited.
-   0 — Creating several dictionaries for the data part is not prohibited.
B
BayoNet 已提交
1535 1536 1537

Default value: 0.

1538
## low\_cardinality\_allow\_in\_native\_format {#low_cardinality_allow_in_native_format}
B
BayoNet 已提交
1539 1540 1541 1542 1543

Allows or restricts using the [LowCardinality](../../sql-reference/data-types/lowcardinality.md) data type with the [Native](../../interfaces/formats.md#native) format.

If usage of `LowCardinality` is restricted, ClickHouse server converts `LowCardinality`-columns to ordinary ones for `SELECT` queries, and convert ordinary columns to `LowCardinality`-columns for `INSERT` queries.

1544
This setting is required mainly for third-party clients which don’t support `LowCardinality` data type.
B
BayoNet 已提交
1545 1546 1547

Possible values:

1548 1549
-   1 — Usage of `LowCardinality` is not restricted.
-   0 — Usage of `LowCardinality` is restricted.
B
BayoNet 已提交
1550 1551 1552

Default value: 1.

1553
## allow\_suspicious\_low\_cardinality\_types {#allow_suspicious_low_cardinality_types}
B
BayoNet 已提交
1554 1555 1556 1557 1558

Allows or restricts using [LowCardinality](../../sql-reference/data-types/lowcardinality.md) with data types with fixed size of 8 bytes or less: numeric data types and `FixedString(8_bytes_or_less)`.

For small fixed values using of `LowCardinality` is usually inefficient, because ClickHouse stores a numeric index for each row. As a result:

1559 1560 1561
-   Disk space usage can rise.
-   RAM consumption can be higher, depending on a dictionary size.
-   Some functions can work slower due to extra coding/encoding operations.
B
BayoNet 已提交
1562 1563 1564 1565 1566

Merge times in [MergeTree](../../engines/table-engines/mergetree-family/mergetree.md)-engine tables can grow due to all the reasons described above.

Possible values:

1567 1568
-   1 — Usage of `LowCardinality` is not restricted.
-   0 — Usage of `LowCardinality` is restricted.
B
BayoNet 已提交
1569 1570 1571

Default value: 0.

1572
## min\_insert\_block\_size\_rows\_for\_materialized\_views {#min-insert-block-size-rows-for-materialized-views}
E
Evgenia Sudarikova 已提交
1573

1574
Sets minimum number of rows in block which can be inserted into a table by an `INSERT` query. Smaller-sized blocks are squashed into bigger ones. This setting is applied only for blocks inserted into [materialized view](../../sql-reference/statements/create/view.md). By adjusting this setting, you control blocks squashing while pushing to materialized view and avoid excessive memory usage.
E
Evgenia Sudarikova 已提交
1575 1576 1577 1578 1579 1580 1581 1582 1583 1584

Possible values:

-   Any positive integer.
-   0 — Squashing disabled.

Default value: 1048576.

**See Also**

1585
-   [min\_insert\_block\_size\_rows](#min-insert-block-size-rows)
E
Evgenia Sudarikova 已提交
1586

1587
## min\_insert\_block\_size\_bytes\_for\_materialized\_views {#min-insert-block-size-bytes-for-materialized-views}
E
Evgenia Sudarikova 已提交
1588

1589
Sets minimum number of bytes in block which can be inserted into a table by an `INSERT` query. Smaller-sized blocks are squashed into bigger ones. This setting is applied only for blocks inserted into [materialized view](../../sql-reference/statements/create/view.md). By adjusting this setting, you control blocks squashing while pushing to materialized view and avoid excessive memory usage.
E
Evgenia Sudarikova 已提交
1590 1591 1592 1593 1594 1595 1596 1597 1598 1599

Possible values:

-   Any positive integer.
-   0 — Squashing disabled.

Default value: 268435456.

**See also**

1600
-   [min\_insert\_block\_size\_bytes](#min-insert-block-size-bytes)
E
Evgenia Sudarikova 已提交
1601

I
Ivan Blinkov 已提交
1602
[Original article](https://clickhouse.tech/docs/en/operations/settings/settings/) <!-- hide -->