Managed Service for ClickHouse API, gRPC: ClusterService.Create
- gRPC request
- CreateClusterRequest
- ConfigSpec
- Clickhouse
- ClickhouseConfig
- MergeTree
- Compression
- ExternalDictionary
- Structure
- Id
- Key
- Attribute
- Layout
- Range
- HttpSource
- Header
- MysqlSource
- Replica
- ClickhouseSource
- MongodbSource
- PostgresqlSource
- GraphiteRollup
- Pattern
- Retention
- Kafka
- KafkaTopic
- Rabbitmq
- QueryMaskingRule
- QueryCache
- Resources
- Zookeeper
- Access
- CloudStorage
- DatabaseSpec
- UserSpec
- Permission
- UserSettings
- UserQuota
- HostSpec
- MaintenanceWindow
- AnytimeMaintenanceWindow
- WeeklyMaintenanceWindow
- operation.Operation
- CreateClusterMetadata
- Cluster
- Monitoring
- ClusterConfig
- Clickhouse
- ClickhouseConfigSet
- ClickhouseConfig
- MergeTree
- Compression
- ExternalDictionary
- Structure
- Id
- Key
- Attribute
- Layout
- Range
- HttpSource
- Header
- MysqlSource
- Replica
- ClickhouseSource
- MongodbSource
- PostgresqlSource
- GraphiteRollup
- Pattern
- Retention
- Kafka
- KafkaTopic
- Rabbitmq
- QueryMaskingRule
- QueryCache
- Resources
- Zookeeper
- Access
- CloudStorage
- MaintenanceWindow
- AnytimeMaintenanceWindow
- WeeklyMaintenanceWindow
- MaintenanceOperation
Creates a ClickHouse cluster in the specified folder.
gRPC request
rpc Create (CreateClusterRequest) returns (operation.Operation)
CreateClusterRequest
{
"folderId": "string",
"name": "string",
"description": "string",
"labels": "string",
"environment": "Environment",
"configSpec": {
"version": "string",
"clickhouse": {
"config": {
"logLevel": "LogLevel",
"mergeTree": {
"replicatedDeduplicationWindow": "google.protobuf.Int64Value",
"replicatedDeduplicationWindowSeconds": "google.protobuf.Int64Value",
"partsToDelayInsert": "google.protobuf.Int64Value",
"partsToThrowInsert": "google.protobuf.Int64Value",
"inactivePartsToDelayInsert": "google.protobuf.Int64Value",
"inactivePartsToThrowInsert": "google.protobuf.Int64Value",
"maxReplicatedMergesInQueue": "google.protobuf.Int64Value",
"numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge": "google.protobuf.Int64Value",
"maxBytesToMergeAtMinSpaceInPool": "google.protobuf.Int64Value",
"maxBytesToMergeAtMaxSpaceInPool": "google.protobuf.Int64Value",
"minBytesForWidePart": "google.protobuf.Int64Value",
"minRowsForWidePart": "google.protobuf.Int64Value",
"ttlOnlyDropParts": "google.protobuf.BoolValue",
"allowRemoteFsZeroCopyReplication": "google.protobuf.BoolValue",
"mergeWithTtlTimeout": "google.protobuf.Int64Value",
"mergeWithRecompressionTtlTimeout": "google.protobuf.Int64Value",
"maxPartsInTotal": "google.protobuf.Int64Value",
"maxNumberOfMergesWithTtlInPool": "google.protobuf.Int64Value",
"cleanupDelayPeriod": "google.protobuf.Int64Value",
"numberOfFreeEntriesInPoolToExecuteMutation": "google.protobuf.Int64Value",
"maxAvgPartSizeForTooManyParts": "google.protobuf.Int64Value",
"minAgeToForceMergeSeconds": "google.protobuf.Int64Value",
"minAgeToForceMergeOnPartitionOnly": "google.protobuf.BoolValue",
"mergeSelectingSleepMs": "google.protobuf.Int64Value",
"mergeMaxBlockSize": "google.protobuf.Int64Value",
"checkSampleColumnIsCorrect": "google.protobuf.BoolValue",
"maxMergeSelectingSleepMs": "google.protobuf.Int64Value",
"maxCleanupDelayPeriod": "google.protobuf.Int64Value"
},
"compression": [
{
"method": "Method",
"minPartSize": "int64",
"minPartSizeRatio": "double",
"level": "google.protobuf.Int64Value"
}
],
"dictionaries": [
{
"name": "string",
"structure": {
"id": {
"name": "string"
},
"key": {
"attributes": [
{
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
}
]
},
"rangeMin": {
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
},
"rangeMax": {
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
},
"attributes": [
{
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
}
]
},
"layout": {
"type": "Type",
"sizeInCells": "int64",
"maxArraySize": "int64"
},
// Includes only one of the fields `fixedLifetime`, `lifetimeRange`
"fixedLifetime": "int64",
"lifetimeRange": {
"min": "int64",
"max": "int64"
},
// end of the list of possible fields
// Includes only one of the fields `httpSource`, `mysqlSource`, `clickhouseSource`, `mongodbSource`, `postgresqlSource`
"httpSource": {
"url": "string",
"format": "string",
"headers": [
{
"name": "string",
"value": "string"
}
]
},
"mysqlSource": {
"db": "string",
"table": "string",
"port": "int64",
"user": "string",
"password": "string",
"replicas": [
{
"host": "string",
"priority": "int64",
"port": "int64",
"user": "string",
"password": "string"
}
],
"where": "string",
"invalidateQuery": "string",
"closeConnection": "google.protobuf.BoolValue",
"shareConnection": "google.protobuf.BoolValue"
},
"clickhouseSource": {
"db": "string",
"table": "string",
"host": "string",
"port": "int64",
"user": "string",
"password": "string",
"where": "string",
"secure": "google.protobuf.BoolValue"
},
"mongodbSource": {
"db": "string",
"collection": "string",
"host": "string",
"port": "int64",
"user": "string",
"password": "string",
"options": "string"
},
"postgresqlSource": {
"db": "string",
"table": "string",
"hosts": [
"string"
],
"port": "int64",
"user": "string",
"password": "string",
"invalidateQuery": "string",
"sslMode": "SslMode"
}
// end of the list of possible fields
}
],
"graphiteRollup": [
{
"name": "string",
"patterns": [
{
"regexp": "string",
"function": "string",
"retention": [
{
"age": "int64",
"precision": "int64"
}
]
}
],
"pathColumnName": "string",
"timeColumnName": "string",
"valueColumnName": "string",
"versionColumnName": "string"
}
],
"kafka": {
"securityProtocol": "SecurityProtocol",
"saslMechanism": "SaslMechanism",
"saslUsername": "string",
"saslPassword": "string",
"enableSslCertificateVerification": "google.protobuf.BoolValue",
"maxPollIntervalMs": "google.protobuf.Int64Value",
"sessionTimeoutMs": "google.protobuf.Int64Value",
"debug": "Debug",
"autoOffsetReset": "AutoOffsetReset"
},
"kafkaTopics": [
{
"name": "string",
"settings": {
"securityProtocol": "SecurityProtocol",
"saslMechanism": "SaslMechanism",
"saslUsername": "string",
"saslPassword": "string",
"enableSslCertificateVerification": "google.protobuf.BoolValue",
"maxPollIntervalMs": "google.protobuf.Int64Value",
"sessionTimeoutMs": "google.protobuf.Int64Value",
"debug": "Debug",
"autoOffsetReset": "AutoOffsetReset"
}
}
],
"rabbitmq": {
"username": "string",
"password": "string",
"vhost": "string"
},
"maxConnections": "google.protobuf.Int64Value",
"maxConcurrentQueries": "google.protobuf.Int64Value",
"keepAliveTimeout": "google.protobuf.Int64Value",
"uncompressedCacheSize": "google.protobuf.Int64Value",
"markCacheSize": "google.protobuf.Int64Value",
"maxTableSizeToDrop": "google.protobuf.Int64Value",
"maxPartitionSizeToDrop": "google.protobuf.Int64Value",
"builtinDictionariesReloadInterval": "google.protobuf.Int64Value",
"timezone": "string",
"geobaseEnabled": "google.protobuf.BoolValue",
"geobaseUri": "string",
"queryLogRetentionSize": "google.protobuf.Int64Value",
"queryLogRetentionTime": "google.protobuf.Int64Value",
"queryThreadLogEnabled": "google.protobuf.BoolValue",
"queryThreadLogRetentionSize": "google.protobuf.Int64Value",
"queryThreadLogRetentionTime": "google.protobuf.Int64Value",
"partLogRetentionSize": "google.protobuf.Int64Value",
"partLogRetentionTime": "google.protobuf.Int64Value",
"metricLogEnabled": "google.protobuf.BoolValue",
"metricLogRetentionSize": "google.protobuf.Int64Value",
"metricLogRetentionTime": "google.protobuf.Int64Value",
"traceLogEnabled": "google.protobuf.BoolValue",
"traceLogRetentionSize": "google.protobuf.Int64Value",
"traceLogRetentionTime": "google.protobuf.Int64Value",
"textLogEnabled": "google.protobuf.BoolValue",
"textLogRetentionSize": "google.protobuf.Int64Value",
"textLogRetentionTime": "google.protobuf.Int64Value",
"textLogLevel": "LogLevel",
"opentelemetrySpanLogEnabled": "google.protobuf.BoolValue",
"opentelemetrySpanLogRetentionSize": "google.protobuf.Int64Value",
"opentelemetrySpanLogRetentionTime": "google.protobuf.Int64Value",
"queryViewsLogEnabled": "google.protobuf.BoolValue",
"queryViewsLogRetentionSize": "google.protobuf.Int64Value",
"queryViewsLogRetentionTime": "google.protobuf.Int64Value",
"asynchronousMetricLogEnabled": "google.protobuf.BoolValue",
"asynchronousMetricLogRetentionSize": "google.protobuf.Int64Value",
"asynchronousMetricLogRetentionTime": "google.protobuf.Int64Value",
"sessionLogEnabled": "google.protobuf.BoolValue",
"sessionLogRetentionSize": "google.protobuf.Int64Value",
"sessionLogRetentionTime": "google.protobuf.Int64Value",
"zookeeperLogEnabled": "google.protobuf.BoolValue",
"zookeeperLogRetentionSize": "google.protobuf.Int64Value",
"zookeeperLogRetentionTime": "google.protobuf.Int64Value",
"asynchronousInsertLogEnabled": "google.protobuf.BoolValue",
"asynchronousInsertLogRetentionSize": "google.protobuf.Int64Value",
"asynchronousInsertLogRetentionTime": "google.protobuf.Int64Value",
"backgroundPoolSize": "google.protobuf.Int64Value",
"backgroundMergesMutationsConcurrencyRatio": "google.protobuf.Int64Value",
"backgroundSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundFetchesPoolSize": "google.protobuf.Int64Value",
"backgroundMovePoolSize": "google.protobuf.Int64Value",
"backgroundDistributedSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundBufferFlushSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundMessageBrokerSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundCommonPoolSize": "google.protobuf.Int64Value",
"defaultDatabase": "google.protobuf.StringValue",
"totalMemoryProfilerStep": "google.protobuf.Int64Value",
"totalMemoryTrackerSampleProbability": "google.protobuf.DoubleValue",
"queryMaskingRules": [
{
"name": "string",
"regexp": "string",
"replace": "string"
}
],
"dictionariesLazyLoad": "google.protobuf.BoolValue",
"queryCache": {
"maxSizeInBytes": "google.protobuf.Int64Value",
"maxEntries": "google.protobuf.Int64Value",
"maxEntrySizeInBytes": "google.protobuf.Int64Value",
"maxEntrySizeInRows": "google.protobuf.Int64Value"
}
},
"resources": {
"resourcePresetId": "string",
"diskSize": "int64",
"diskTypeId": "string"
}
},
"zookeeper": {
"resources": {
"resourcePresetId": "string",
"diskSize": "int64",
"diskTypeId": "string"
}
},
"backupWindowStart": "google.type.TimeOfDay",
"access": {
"dataLens": "bool",
"webSql": "bool",
"metrika": "bool",
"serverless": "bool",
"dataTransfer": "bool",
"yandexQuery": "bool"
},
"cloudStorage": {
"enabled": "bool",
"moveFactor": "google.protobuf.DoubleValue",
"dataCacheEnabled": "google.protobuf.BoolValue",
"dataCacheMaxSize": "google.protobuf.Int64Value",
"preferNotToMerge": "google.protobuf.BoolValue"
},
"sqlDatabaseManagement": "google.protobuf.BoolValue",
"sqlUserManagement": "google.protobuf.BoolValue",
"adminPassword": "string",
"embeddedKeeper": "google.protobuf.BoolValue",
"backupRetainPeriodDays": "google.protobuf.Int64Value"
},
"databaseSpecs": [
{
"name": "string"
}
],
"userSpecs": [
{
"name": "string",
"password": "string",
"permissions": [
{
"databaseName": "string"
}
],
"settings": {
"readonly": "google.protobuf.Int64Value",
"allowDdl": "google.protobuf.BoolValue",
"allowIntrospectionFunctions": "google.protobuf.BoolValue",
"connectTimeout": "google.protobuf.Int64Value",
"connectTimeoutWithFailover": "google.protobuf.Int64Value",
"receiveTimeout": "google.protobuf.Int64Value",
"sendTimeout": "google.protobuf.Int64Value",
"timeoutBeforeCheckingExecutionSpeed": "google.protobuf.Int64Value",
"insertQuorum": "google.protobuf.Int64Value",
"insertQuorumTimeout": "google.protobuf.Int64Value",
"insertQuorumParallel": "google.protobuf.BoolValue",
"insertNullAsDefault": "google.protobuf.BoolValue",
"selectSequentialConsistency": "google.protobuf.BoolValue",
"deduplicateBlocksInDependentMaterializedViews": "google.protobuf.BoolValue",
"replicationAlterPartitionsSync": "google.protobuf.Int64Value",
"maxReplicaDelayForDistributedQueries": "google.protobuf.Int64Value",
"fallbackToStaleReplicasForDistributedQueries": "google.protobuf.BoolValue",
"distributedProductMode": "DistributedProductMode",
"distributedAggregationMemoryEfficient": "google.protobuf.BoolValue",
"distributedDdlTaskTimeout": "google.protobuf.Int64Value",
"skipUnavailableShards": "google.protobuf.BoolValue",
"compileExpressions": "google.protobuf.BoolValue",
"minCountToCompileExpression": "google.protobuf.Int64Value",
"maxBlockSize": "google.protobuf.Int64Value",
"minInsertBlockSizeRows": "google.protobuf.Int64Value",
"minInsertBlockSizeBytes": "google.protobuf.Int64Value",
"maxInsertBlockSize": "google.protobuf.Int64Value",
"minBytesToUseDirectIo": "google.protobuf.Int64Value",
"useUncompressedCache": "google.protobuf.BoolValue",
"mergeTreeMaxRowsToUseCache": "google.protobuf.Int64Value",
"mergeTreeMaxBytesToUseCache": "google.protobuf.Int64Value",
"mergeTreeMinRowsForConcurrentRead": "google.protobuf.Int64Value",
"mergeTreeMinBytesForConcurrentRead": "google.protobuf.Int64Value",
"maxBytesBeforeExternalGroupBy": "google.protobuf.Int64Value",
"maxBytesBeforeExternalSort": "google.protobuf.Int64Value",
"groupByTwoLevelThreshold": "google.protobuf.Int64Value",
"groupByTwoLevelThresholdBytes": "google.protobuf.Int64Value",
"priority": "google.protobuf.Int64Value",
"maxThreads": "google.protobuf.Int64Value",
"maxMemoryUsage": "google.protobuf.Int64Value",
"maxMemoryUsageForUser": "google.protobuf.Int64Value",
"maxNetworkBandwidth": "google.protobuf.Int64Value",
"maxNetworkBandwidthForUser": "google.protobuf.Int64Value",
"maxPartitionsPerInsertBlock": "google.protobuf.Int64Value",
"maxConcurrentQueriesForUser": "google.protobuf.Int64Value",
"forceIndexByDate": "google.protobuf.BoolValue",
"forcePrimaryKey": "google.protobuf.BoolValue",
"maxRowsToRead": "google.protobuf.Int64Value",
"maxBytesToRead": "google.protobuf.Int64Value",
"readOverflowMode": "OverflowMode",
"maxRowsToGroupBy": "google.protobuf.Int64Value",
"groupByOverflowMode": "GroupByOverflowMode",
"maxRowsToSort": "google.protobuf.Int64Value",
"maxBytesToSort": "google.protobuf.Int64Value",
"sortOverflowMode": "OverflowMode",
"maxResultRows": "google.protobuf.Int64Value",
"maxResultBytes": "google.protobuf.Int64Value",
"resultOverflowMode": "OverflowMode",
"maxRowsInDistinct": "google.protobuf.Int64Value",
"maxBytesInDistinct": "google.protobuf.Int64Value",
"distinctOverflowMode": "OverflowMode",
"maxRowsToTransfer": "google.protobuf.Int64Value",
"maxBytesToTransfer": "google.protobuf.Int64Value",
"transferOverflowMode": "OverflowMode",
"maxExecutionTime": "google.protobuf.Int64Value",
"timeoutOverflowMode": "OverflowMode",
"maxRowsInSet": "google.protobuf.Int64Value",
"maxBytesInSet": "google.protobuf.Int64Value",
"setOverflowMode": "OverflowMode",
"maxRowsInJoin": "google.protobuf.Int64Value",
"maxBytesInJoin": "google.protobuf.Int64Value",
"joinOverflowMode": "OverflowMode",
"joinAlgorithm": [
"JoinAlgorithm"
],
"anyJoinDistinctRightTableKeys": "google.protobuf.BoolValue",
"maxColumnsToRead": "google.protobuf.Int64Value",
"maxTemporaryColumns": "google.protobuf.Int64Value",
"maxTemporaryNonConstColumns": "google.protobuf.Int64Value",
"maxQuerySize": "google.protobuf.Int64Value",
"maxAstDepth": "google.protobuf.Int64Value",
"maxAstElements": "google.protobuf.Int64Value",
"maxExpandedAstElements": "google.protobuf.Int64Value",
"minExecutionSpeed": "google.protobuf.Int64Value",
"minExecutionSpeedBytes": "google.protobuf.Int64Value",
"countDistinctImplementation": "CountDistinctImplementation",
"inputFormatValuesInterpretExpressions": "google.protobuf.BoolValue",
"inputFormatDefaultsForOmittedFields": "google.protobuf.BoolValue",
"inputFormatNullAsDefault": "google.protobuf.BoolValue",
"dateTimeInputFormat": "DateTimeInputFormat",
"inputFormatWithNamesUseHeader": "google.protobuf.BoolValue",
"outputFormatJsonQuote_64bitIntegers": "google.protobuf.BoolValue",
"outputFormatJsonQuoteDenormals": "google.protobuf.BoolValue",
"dateTimeOutputFormat": "DateTimeOutputFormat",
"lowCardinalityAllowInNativeFormat": "google.protobuf.BoolValue",
"allowSuspiciousLowCardinalityTypes": "google.protobuf.BoolValue",
"emptyResultForAggregationByEmptySet": "google.protobuf.BoolValue",
"httpConnectionTimeout": "google.protobuf.Int64Value",
"httpReceiveTimeout": "google.protobuf.Int64Value",
"httpSendTimeout": "google.protobuf.Int64Value",
"enableHttpCompression": "google.protobuf.BoolValue",
"sendProgressInHttpHeaders": "google.protobuf.BoolValue",
"httpHeadersProgressInterval": "google.protobuf.Int64Value",
"addHttpCorsHeader": "google.protobuf.BoolValue",
"cancelHttpReadonlyQueriesOnClientClose": "google.protobuf.BoolValue",
"maxHttpGetRedirects": "google.protobuf.Int64Value",
"joinedSubqueryRequiresAlias": "google.protobuf.BoolValue",
"joinUseNulls": "google.protobuf.BoolValue",
"transformNullIn": "google.protobuf.BoolValue",
"quotaMode": "QuotaMode",
"flattenNested": "google.protobuf.BoolValue",
"formatRegexp": "string",
"formatRegexpEscapingRule": "FormatRegexpEscapingRule",
"formatRegexpSkipUnmatched": "google.protobuf.BoolValue",
"asyncInsert": "google.protobuf.BoolValue",
"asyncInsertThreads": "google.protobuf.Int64Value",
"waitForAsyncInsert": "google.protobuf.BoolValue",
"waitForAsyncInsertTimeout": "google.protobuf.Int64Value",
"asyncInsertMaxDataSize": "google.protobuf.Int64Value",
"asyncInsertBusyTimeout": "google.protobuf.Int64Value",
"asyncInsertStaleTimeout": "google.protobuf.Int64Value",
"memoryProfilerStep": "google.protobuf.Int64Value",
"memoryProfilerSampleProbability": "google.protobuf.DoubleValue",
"maxFinalThreads": "google.protobuf.Int64Value",
"inputFormatParallelParsing": "google.protobuf.BoolValue",
"inputFormatImportNestedJson": "google.protobuf.BoolValue",
"localFilesystemReadMethod": "LocalFilesystemReadMethod",
"maxReadBufferSize": "google.protobuf.Int64Value",
"insertKeeperMaxRetries": "google.protobuf.Int64Value",
"maxTemporaryDataOnDiskSizeForUser": "google.protobuf.Int64Value",
"maxTemporaryDataOnDiskSizeForQuery": "google.protobuf.Int64Value",
"maxParserDepth": "google.protobuf.Int64Value",
"remoteFilesystemReadMethod": "RemoteFilesystemReadMethod",
"memoryOvercommitRatioDenominator": "google.protobuf.Int64Value",
"memoryOvercommitRatioDenominatorForUser": "google.protobuf.Int64Value",
"memoryUsageOvercommitMaxWaitMicroseconds": "google.protobuf.Int64Value",
"logQueryThreads": "google.protobuf.BoolValue",
"maxInsertThreads": "google.protobuf.Int64Value",
"useHedgedRequests": "google.protobuf.BoolValue",
"idleConnectionTimeout": "google.protobuf.Int64Value",
"hedgedConnectionTimeoutMs": "google.protobuf.Int64Value",
"loadBalancing": "LoadBalancing",
"preferLocalhostReplica": "google.protobuf.BoolValue",
"compile": "google.protobuf.BoolValue",
"minCountToCompile": "google.protobuf.Int64Value"
},
"quotas": [
{
"intervalDuration": "google.protobuf.Int64Value",
"queries": "google.protobuf.Int64Value",
"errors": "google.protobuf.Int64Value",
"resultRows": "google.protobuf.Int64Value",
"readRows": "google.protobuf.Int64Value",
"executionTime": "google.protobuf.Int64Value"
}
]
}
],
"hostSpecs": [
{
"zoneId": "string",
"type": "Type",
"subnetId": "string",
"assignPublicIp": "bool",
"shardName": "string"
}
],
"networkId": "string",
"shardName": "string",
"serviceAccountId": "string",
"securityGroupIds": [
"string"
],
"deletionProtection": "bool",
"maintenanceWindow": {
// Includes only one of the fields `anytime`, `weeklyMaintenanceWindow`
"anytime": "AnytimeMaintenanceWindow",
"weeklyMaintenanceWindow": {
"day": "WeekDay",
"hour": "int64"
}
// end of the list of possible fields
}
}
Field |
Description |
folderId |
string Required field. ID of the folder to create the ClickHouse cluster in. |
name |
string Required field. Name of the ClickHouse cluster. The name must be unique within the folder. |
description |
string Description of the ClickHouse cluster. |
labels |
string Custom labels for the ClickHouse cluster as |
environment |
enum Environment Required field. Deployment environment of the ClickHouse cluster.
|
configSpec |
Required field. Configuration and resources for hosts that should be created for the ClickHouse cluster. |
databaseSpecs[] |
Descriptions of databases to be created in the ClickHouse cluster. |
userSpecs[] |
Descriptions of database users to be created in the ClickHouse cluster. |
hostSpecs[] |
Individual configurations for hosts that should be created for the ClickHouse cluster. |
networkId |
string Required field. ID of the network to create the cluster in. |
shardName |
string Name of the first shard in cluster. If not set, defaults to the value 'shard1'. |
serviceAccountId |
string ID of the service account used for access to Object Storage. |
securityGroupIds[] |
string User security groups |
deletionProtection |
bool Deletion Protection inhibits deletion of the cluster |
maintenanceWindow |
Window of maintenance operations. |
ConfigSpec
Field |
Description |
version |
string Version of the ClickHouse server software. |
clickhouse |
Configuration and resources for a ClickHouse server. |
zookeeper |
Configuration and resources for a ZooKeeper server. |
backupWindowStart |
Time to start the daily backup, in the UTC timezone. |
access |
Access policy for external services. If you want a specific service to access the ClickHouse cluster, then set the necessary values in this policy. |
cloudStorage |
|
sqlDatabaseManagement |
Whether database management through SQL commands is enabled. |
sqlUserManagement |
Whether user management through SQL commands is enabled. |
adminPassword |
string Password for user 'admin' that has SQL user management access. |
embeddedKeeper |
Whether cluster should use embedded Keeper instead of Zookeeper |
backupRetainPeriodDays |
Retain period of automatically created backup in days |
Clickhouse
Field |
Description |
config |
Configuration for a ClickHouse server. |
resources |
Resources allocated to ClickHouse hosts. |
ClickhouseConfig
ClickHouse configuration options. Detailed description for each set of options
is available in ClickHouse documentation
Any options not listed here are not supported.
Field |
Description |
logLevel |
enum LogLevel Logging level for the ClickHouse cluster. Possible values: TRACE, DEBUG, INFORMATION, WARNING, ERROR.
|
mergeTree |
Settings for the MergeTree engine. |
compression[] |
Compression settings for the ClickHouse cluster. |
dictionaries[] |
Configuration of external dictionaries to be used by the ClickHouse cluster. |
graphiteRollup[] |
Settings for thinning Graphite data. |
kafka |
|
kafkaTopics[] |
|
rabbitmq |
|
maxConnections |
Maximum number of inbound connections. |
maxConcurrentQueries |
Maximum number of simultaneously processed requests. |
keepAliveTimeout |
Number of milliseconds that ClickHouse waits for incoming requests before closing the connection. |
uncompressedCacheSize |
Cache size (in bytes) for uncompressed data used by MergeTree tables. |
markCacheSize |
Approximate size (in bytes) of the cache of "marks" used by MergeTree tables. |
maxTableSizeToDrop |
Maximum size of the table that can be deleted using a DROP query. |
maxPartitionSizeToDrop |
Maximum size of the partition that can be deleted using a DROP query. |
builtinDictionariesReloadInterval |
The setting is deprecated and has no effect. |
timezone |
string The server's time zone to be used in DateTime fields conversions. Specified as an IANA identifier. |
geobaseEnabled |
Enable or disable geobase. |
geobaseUri |
string Address of the archive with the user geobase in Object Storage. |
queryLogRetentionSize |
The maximum size that query_log can grow to before old data will be removed. If set to 0, automatic removal of |
queryLogRetentionTime |
The maximum time that query_log records will be retained before removal. If set to 0, automatic removal of |
queryThreadLogEnabled |
Whether query_thread_log system table is enabled. |
queryThreadLogRetentionSize |
The maximum size that query_thread_log can grow to before old data will be removed. If set to 0, automatic removal of |
queryThreadLogRetentionTime |
The maximum time that query_thread_log records will be retained before removal. If set to 0, automatic removal of |
partLogRetentionSize |
The maximum size that part_log can grow to before old data will be removed. If set to 0, automatic removal of |
partLogRetentionTime |
The maximum time that part_log records will be retained before removal. If set to 0, automatic removal of |
metricLogEnabled |
Whether metric_log system table is enabled. |
metricLogRetentionSize |
The maximum size that metric_log can grow to before old data will be removed. If set to 0, automatic removal of |
metricLogRetentionTime |
The maximum time that metric_log records will be retained before removal. If set to 0, automatic removal of |
traceLogEnabled |
Whether trace_log system table is enabled. |
traceLogRetentionSize |
The maximum size that trace_log can grow to before old data will be removed. If set to 0, automatic removal of |
traceLogRetentionTime |
The maximum time that trace_log records will be retained before removal. If set to 0, automatic removal of |
textLogEnabled |
Whether text_log system table is enabled. |
textLogRetentionSize |
The maximum size that text_log can grow to before old data will be removed. If set to 0, automatic removal of |
textLogRetentionTime |
The maximum time that text_log records will be retained before removal. If set to 0, automatic removal of |
textLogLevel |
enum LogLevel Logging level for text_log system table. Possible values: TRACE, DEBUG, INFORMATION, WARNING, ERROR.
|
opentelemetrySpanLogEnabled |
Enable or disable opentelemetry_span_log system table. Default value: false. |
opentelemetrySpanLogRetentionSize |
The maximum size that opentelemetry_span_log can grow to before old data will be removed. If set to 0 (default), |
opentelemetrySpanLogRetentionTime |
The maximum time that opentelemetry_span_log records will be retained before removal. If set to 0, |
queryViewsLogEnabled |
Enable or disable query_views_log system table. Default value: false. |
queryViewsLogRetentionSize |
The maximum size that query_views_log can grow to before old data will be removed. If set to 0 (default), |
queryViewsLogRetentionTime |
The maximum time that query_views_log records will be retained before removal. If set to 0, |
asynchronousMetricLogEnabled |
Enable or disable asynchronous_metric_log system table. Default value: false. |
asynchronousMetricLogRetentionSize |
The maximum size that asynchronous_metric_log can grow to before old data will be removed. If set to 0 (default), |
asynchronousMetricLogRetentionTime |
The maximum time that asynchronous_metric_log records will be retained before removal. If set to 0, |
sessionLogEnabled |
Enable or disable session_log system table. Default value: false. |
sessionLogRetentionSize |
The maximum size that session_log can grow to before old data will be removed. If set to 0 (default), |
sessionLogRetentionTime |
The maximum time that session_log records will be retained before removal. If set to 0, |
zookeeperLogEnabled |
Enable or disable zookeeper_log system table. Default value: false. |
zookeeperLogRetentionSize |
The maximum size that zookeeper_log can grow to before old data will be removed. If set to 0 (default), |
zookeeperLogRetentionTime |
The maximum time that zookeeper_log records will be retained before removal. If set to 0, |
asynchronousInsertLogEnabled |
Enable or disable asynchronous_insert_log system table. Default value: false. |
asynchronousInsertLogRetentionSize |
The maximum size that asynchronous_insert_log can grow to before old data will be removed. If set to 0 (default), |
asynchronousInsertLogRetentionTime |
The maximum time that asynchronous_insert_log records will be retained before removal. If set to 0, |
backgroundPoolSize |
|
backgroundMergesMutationsConcurrencyRatio |
Sets a ratio between the number of threads and the number of background merges and mutations that can be executed concurrently. For example, if the ratio equals to 2 and background_pool_size is set to 16 then ClickHouse can execute 32 background merges concurrently. This is possible, because background operations could be suspended and postponed. This is needed to give small merges more execution priority. You can only increase this ratio at runtime. To lower it you have to restart the server. The same as for background_pool_size setting background_merges_mutations_concurrency_ratio could be applied from the default profile for backward compatibility. |
backgroundSchedulePoolSize |
|
backgroundFetchesPoolSize |
Sets the number of threads performing background fetches for tables with ReplicatedMergeTree engines. Default value: 8. More info see in ClickHouse documentation |
backgroundMovePoolSize |
|
backgroundDistributedSchedulePoolSize |
|
backgroundBufferFlushSchedulePoolSize |
|
backgroundMessageBrokerSchedulePoolSize |
|
backgroundCommonPoolSize |
The maximum number of threads that will be used for performing a variety of operations (mostly garbage collection) for *MergeTree-engine tables in a background. |
defaultDatabase |
The default database. To get a list of cluster databases, see Yandex Managed ClickHouse documentation. |
totalMemoryProfilerStep |
Sets the memory size (in bytes) for a stack trace at every peak allocation step. Default value: 4194304. More info see in ClickHouse documentation |
totalMemoryTrackerSampleProbability |
|
queryMaskingRules[] |
Regexp-based rules, which will be applied to queries as well as all log messages before storing them in server logs, system.query_log, system.text_log, system.processes tables, and in logs sent to the client. That allows preventing sensitive data leakage from SQL queries (like names, emails, personal identifiers or credit card numbers) to logs. |
dictionariesLazyLoad |
Lazy loading of dictionaries. |
queryCache |
Query cache |
MergeTree
Options specific to the MergeTree table engine.
Field |
Description |
replicatedDeduplicationWindow |
Number of blocks of hashes to keep in ZooKeeper. |
replicatedDeduplicationWindowSeconds |
Period of time to keep blocks of hashes for. |
partsToDelayInsert |
If table contains at least that many active parts in single partition, artificially slow down insert into table. |
partsToThrowInsert |
If more than this number active parts in single partition, throw 'Too many parts ...' exception. |
inactivePartsToDelayInsert |
|
inactivePartsToThrowInsert |
|
maxReplicatedMergesInQueue |
How many tasks of merging and mutating parts are allowed simultaneously in ReplicatedMergeTree queue. |
numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge |
If there is less than specified number of free entries in background pool (or replicated queue), start to lower |
maxBytesToMergeAtMinSpaceInPool |
Maximum in total size of parts to merge, when there are minimum free threads in background pool (or entries |
maxBytesToMergeAtMaxSpaceInPool |
|
minBytesForWidePart |
Minimum number of bytes in a data part that can be stored in Wide format. More info see in ClickHouse documentation |
minRowsForWidePart |
Minimum number of rows in a data part that can be stored in Wide format. More info see in ClickHouse documentation |
ttlOnlyDropParts |
Enables or disables complete dropping of data parts where all rows are expired in MergeTree tables. More info see in ClickHouse documentation |
allowRemoteFsZeroCopyReplication |
|
mergeWithTtlTimeout |
|
mergeWithRecompressionTtlTimeout |
|
maxPartsInTotal |
|
maxNumberOfMergesWithTtlInPool |
|
cleanupDelayPeriod |
|
numberOfFreeEntriesInPoolToExecuteMutation |
|
maxAvgPartSizeForTooManyParts |
The 'too many parts' check according to 'parts_to_delay_insert' and 'parts_to_throw_insert' will be active only if the average part size (in the relevant partition) is not larger than the specified threshold. If it is larger than the specified threshold, the INSERTs will be neither delayed or rejected. This allows to have hundreds of terabytes in a single table on a single server if the parts are successfully merged to larger parts. This does not affect the thresholds on inactive parts or total parts. |
minAgeToForceMergeSeconds |
Merge parts if every part in the range is older than the value of min_age_to_force_merge_seconds. |
minAgeToForceMergeOnPartitionOnly |
Whether min_age_to_force_merge_seconds should be applied only on the entire partition and not on subset. |
mergeSelectingSleepMs |
Sleep time for merge selecting when no part is selected. A lower setting triggers selecting tasks in background_schedule_pool frequently, which results in a large number of requests to ClickHouse Keeper in large-scale clusters. |
mergeMaxBlockSize |
The number of rows that are read from the merged parts into memory. |
checkSampleColumnIsCorrect |
Enables the check at table creation, that the data type of a column for sampling or sampling expression is correct. The data type must be one of unsigned integer types |
maxMergeSelectingSleepMs |
Maximum sleep time for merge selecting, a lower setting will trigger selecting tasks in background_schedule_pool frequently which result in large amount of requests to zookeeper in large-scale clusters. |
maxCleanupDelayPeriod |
Maximum period to clean old queue logs, blocks hashes and parts. |
Compression
Field |
Description |
method |
enum Method Compression method to use for the specified combination of
|
minPartSize |
int64 Minimum size of a part of a table. |
minPartSizeRatio |
double Minimum ratio of a part relative to the size of all the data in the table. |
level |
ExternalDictionary
Field |
Description |
name |
string Required field. Name of the external dictionary. |
structure |
Required field. Set of attributes for the external dictionary. |
layout |
Required field. Layout for storing the dictionary in memory. |
fixedLifetime |
int64 Fixed interval between dictionary updates. Includes only one of the fields Setting for the period of time between dictionary updates. |
lifetimeRange |
Range of intervals between dictionary updates for ClickHouse to choose from. Includes only one of the fields Setting for the period of time between dictionary updates. |
httpSource |
HTTP source for the dictionary. Includes only one of the fields Description of the source for the external dictionary. |
mysqlSource |
MySQL source for the dictionary. Includes only one of the fields Description of the source for the external dictionary. |
clickhouseSource |
ClickHouse source for the dictionary. Includes only one of the fields Description of the source for the external dictionary. |
mongodbSource |
MongoDB source for the dictionary. Includes only one of the fields Description of the source for the external dictionary. |
postgresqlSource |
PostgreSQL source for the dictionary. Includes only one of the fields Description of the source for the external dictionary. |
Structure
Field |
Description |
id |
Single numeric key column for the dictionary. |
key |
Composite key for the dictionary, containing of one or more key columns. |
rangeMin |
Field holding the beginning of the range for dictionaries with |
rangeMax |
Field holding the end of the range for dictionaries with |
attributes[] |
Description of the fields available for database queries. |
Id
Numeric key.
Field |
Description |
name |
string Required field. Name of the numeric key. |
Key
Complex key.
Field |
Description |
attributes[] |
Attributes of a complex key. |
Attribute
Field |
Description |
name |
string Required field. Name of the column. |
type |
string Required field. Type of the column. |
nullValue |
string Default value for an element without data (for example, an empty string). |
expression |
string Expression, describing the attribute, if applicable. |
hierarchical |
bool Indication of hierarchy support. |
injective |
bool Indication of injective mapping "id -> attribute". |
Layout
Layout determining how to store the dictionary in memory.
Field |
Description |
type |
enum Type Required field. Layout type for an external dictionary.
|
sizeInCells |
int64 Number of cells in the cache. Rounded up to a power of two. |
maxArraySize |
int64 Maximum dictionary key size. |
Range
Field |
Description |
min |
int64 Minimum dictionary lifetime. |
max |
int64 Maximum dictionary lifetime. |
HttpSource
Field |
Description |
url |
string Required field. URL of the source dictionary available over HTTP. |
format |
string Required field. The data format. Valid values are all formats supported by ClickHouse SQL dialect. |
headers[] |
HTTP headers. |
Header
Field |
Description |
name |
string Required field. |
value |
string Required field. |
MysqlSource
Field |
Description |
db |
string Required field. Name of the MySQL database to connect to. |
table |
string Required field. Name of the database table to use as a ClickHouse dictionary. |
port |
int64 Default port to use when connecting to a replica of the dictionary source. |
user |
string Name of the default user for replicas of the dictionary source. |
password |
string Password of the default user for replicas of the dictionary source. |
replicas[] |
List of MySQL replicas of the database used as dictionary source. |
where |
string Selection criteria for the data in the specified MySQL table. |
invalidateQuery |
string Query for checking the dictionary status, to pull only updated data. |
closeConnection |
Should the connection be closed after each request. |
shareConnection |
Should a connection be shared for some requests. |
Replica
Field |
Description |
host |
string Required field. MySQL host of the replica. |
priority |
int64 Required field. The priority of the replica that ClickHouse takes into account when connecting. |
port |
int64 Port to use when connecting to the replica. |
user |
string Name of the MySQL database user. |
password |
string Password of the MySQL database user. |
ClickhouseSource
Field |
Description |
db |
string Required field. Name of the ClickHouse database. |
table |
string Required field. Name of the table in the specified database to be used as the dictionary source. |
host |
string ClickHouse host of the specified database. |
port |
int64 Port to use when connecting to the host. |
user |
string Required field. Name of the ClickHouse database user. |
password |
string Password of the ClickHouse database user. |
where |
string Selection criteria for the data in the specified ClickHouse table. |
secure |
Use ssl for connection. |
MongodbSource
Field |
Description |
db |
string Required field. Name of the MongoDB database. |
collection |
string Required field. Name of the collection in the specified database to be used as the dictionary source. |
host |
string MongoDB host of the specified database. |
port |
int64 Port to use when connecting to the host. |
user |
string Required field. Name of the MongoDB database user. |
password |
string Password of the MongoDB database user. |
options |
string |
PostgresqlSource
Field |
Description |
db |
string Required field. Name of the PostrgreSQL database. |
table |
string Required field. Name of the table in the specified database to be used as the dictionary source. |
hosts[] |
string Name of the PostrgreSQL host |
port |
int64 Port to use when connecting to the host. |
user |
string Required field. Name of the PostrgreSQL database user. |
password |
string Password of the PostrgreSQL database user. |
invalidateQuery |
string Query for checking the dictionary status, to pull only updated data. |
sslMode |
enum SslMode Mode of SSL TCP/IP connection to the PostgreSQL host.
|
GraphiteRollup
Rollup settings for the GraphiteMergeTree table engine.
Field |
Description |
name |
string Required field. Name for the specified combination of settings for Graphite rollup. |
patterns[] |
Pattern to use for the rollup. |
pathColumnName |
string The name of the column storing the metric name (Graphite sensor). |
timeColumnName |
string The name of the column storing the time of measuring the metric. |
valueColumnName |
string The name of the column storing the value of the metric at the time set in time_column_name. |
versionColumnName |
string The name of the column storing the version of the metric. |
Pattern
Field |
Description |
regexp |
string Pattern for metric names. |
function |
string Required field. Name of the aggregating function to apply to data of the age specified in |
retention[] |
Age of data to use for thinning. |
Retention
Field |
Description |
age |
int64 Minimum age of the data in seconds. |
precision |
int64 Precision of determining the age of the data, in seconds. |
Kafka
Field |
Description |
securityProtocol |
enum SecurityProtocol
|
saslMechanism |
enum SaslMechanism
|
saslUsername |
string |
saslPassword |
string |
enableSslCertificateVerification |
|
maxPollIntervalMs |
|
sessionTimeoutMs |
|
debug |
enum Debug
|
autoOffsetReset |
enum AutoOffsetReset
|
KafkaTopic
Field |
Description |
name |
string Required field. |
settings |
Required field. |
Rabbitmq
Field |
Description |
username |
string RabbitMQ |
password |
string RabbitMQ |
vhost |
string RabbitMQ |
QueryMaskingRule
Field |
Description |
name |
string Name for the rule. |
regexp |
string Required field. RE2 compatible regular expression. |
replace |
string Substitution string for sensitive data. |
QueryCache
Field |
Description |
maxSizeInBytes |
The maximum cache size in bytes. |
maxEntries |
The maximum number of SELECT query results stored in the cache. |
maxEntrySizeInBytes |
The maximum size in bytes SELECT query results may have to be saved in the cache. |
maxEntrySizeInRows |
The maximum number of rows SELECT query results may have to be saved in the cache. |
Resources
Field |
Description |
resourcePresetId |
string ID of the preset for computational resources available to a host (CPU, memory etc.). |
diskSize |
int64 Volume of the storage available to a host, in bytes. |
diskTypeId |
string Type of the storage environment for the host.
|
Zookeeper
Field |
Description |
resources |
Resources allocated to ZooKeeper hosts. If not set, minimal available resources will be used. |
Access
Field |
Description |
dataLens |
bool Allow to export data from the cluster to DataLens. |
webSql |
bool Allow SQL queries to the cluster databases from the management console. See SQL queries in the management console for more details. |
metrika |
bool Allow to import data from Yandex Metrica and AppMetrica to the cluster. See AppMetrica documentation |
serverless |
bool Allow access to cluster for Serverless. |
dataTransfer |
bool Allow access for DataTransfer |
yandexQuery |
bool Allow access for Query |
CloudStorage
Field |
Description |
enabled |
bool Whether to use Object Storage for storing ClickHouse data. |
moveFactor |
|
dataCacheEnabled |
|
dataCacheMaxSize |
|
preferNotToMerge |
DatabaseSpec
Field |
Description |
name |
string Required field. Name of the ClickHouse database. 1-63 characters long. |
UserSpec
Field |
Description |
name |
string Required field. Name of the ClickHouse user. |
password |
string Required field. Password of the ClickHouse user. |
permissions[] |
Set of permissions to grant to the user. If not set, it's granted permissions to access all databases. |
settings |
|
quotas[] |
Set of quotas assigned to the user. |
Permission
Field |
Description |
databaseName |
string Name of the database that the permission grants access to. |
UserSettings
ClickHouse user settings. Supported settings are a limited subset of all settings
described in ClickHouse documentation
Field |
Description |
readonly |
Restricts permissions for non-DDL queries. To restrict permissions for DDL queries, use
See in-depth description in ClickHouse documentation |
allowDdl |
Determines whether DDL queries are allowed (e.g., CREATE, ALTER, RENAME, etc). Default value: true. See in-depth description in ClickHouse documentation |
allowIntrospectionFunctions |
Enables introspections functions More info see in ClickHouse documentation |
connectTimeout |
Connection timeout in milliseconds. Value must be greater than 0 (default: 10000, 10 seconds). |
connectTimeoutWithFailover |
The timeout in milliseconds for connecting to a remote server for a Distributed table engine. Applies only if the cluster uses sharding and replication. If unsuccessful, several attempts are made to connect to various replicas. Default value: 50. More info see in ClickHouse documentation |
receiveTimeout |
Receive timeout in milliseconds. Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). |
sendTimeout |
Send timeout in milliseconds. Value must be greater than 0 (default: 300000, 300 seconds or 5 minutes). |
timeoutBeforeCheckingExecutionSpeed |
Timeout (in seconds) between checks of execution speed. It is checked that execution speed is not less that specified in Default value: 10. |
insertQuorum |
Enables or disables write quorum for ClickHouse cluster. When used, write quorum guarantees that ClickHouse has written data to the quorum of insert_quorum replicas with no errors until the You can use See in-depth description in ClickHouse documentation |
insertQuorumTimeout |
Quorum write timeout in milliseconds. If the write quorum is enabled in the cluster, this timeout expires and some data is not written to the Minimum value: 1000, 1 second (default: 60000, 1 minute). |
insertQuorumParallel |
See in-depth description in ClickHouse documentation |
insertNullAsDefault |
Enables the insertion of default values instead of NULL into columns with not nullable data type. Default value: true. More info see in ClickHouse documentation |
selectSequentialConsistency |
Determines the behavior of SELECT queries from the replicated table: if enabled, ClickHouse will terminate a query with error message in case the replica does not have a chunk written with the quorum and will not read the parts that have not yet been written with the quorum. Default value: false (sequential consistency is disabled). |
deduplicateBlocksInDependentMaterializedViews |
See in-depth description in ClickHouse documentation |
replicationAlterPartitionsSync |
Wait mode for asynchronous actions in ALTER queries on replicated tables:
See in-depth description in ClickHouse documentation |
maxReplicaDelayForDistributedQueries |
Max replica delay in milliseconds. If a replica lags more than the set value, this replica is not used and becomes a stale one. Minimum value: 1000, 1 second (default: 300000, 300 seconds or 5 minutes). See in-depth description in ClickHouse documentation |
fallbackToStaleReplicasForDistributedQueries |
Enables or disables query forcing to a stale replica in case the actual data is unavailable. Default value: true (query forcing is enabled). See in-depth description in ClickHouse documentation |
distributedProductMode |
enum DistributedProductMode Determine the behavior of distributed subqueries. See in-depth description in ClickHouse documentation
|
distributedAggregationMemoryEfficient |
Enables of disables memory saving mode when doing distributed aggregation. When ClickHouse works with a distributed query, external aggregation is done on remote servers. Default value: false (memory saving mode is disabled). See in-depth description in ClickHouse documentation |
distributedDdlTaskTimeout |
Timeout for DDL queries, in milliseconds. |
skipUnavailableShards |
Enables or disables silent skipping of unavailable shards. A shard is considered unavailable if all its replicas are also unavailable. Default value: false (silent skipping is disabled). See in-depth description in ClickHouse documentation |
compileExpressions |
Enables or disables expression compilation. Use this setting in combination with Default value: false (expression compilation is disabled). |
minCountToCompileExpression |
How many identical expressions ClickHouse has to encounter before they are compiled. Minimum value: 0 (default: 3). For the 0 value compilation is synchronous: a query waits for expression compilation process to complete prior to continuing execution. For all other values, compilation is asynchronous: the compilation process executes in a separate thread. |
maxBlockSize |
The maximum block size for reading. Data in ClickHouse is organized and processed by blocks (block is a set of columns' parts). This setting is a recommendation for size of block (in a count of rows) that should be loaded from tables. Value must be greater than 0 (default: 65536). See in-depth description in ClickHouse documentation |
minInsertBlockSizeRows |
Limits the minimum number of rows in a block to be inserted in a table by INSERT query. Minimal value: 0, block squashing is disabled (default: 1048576). |
minInsertBlockSizeBytes |
Limits the minimum number of bytes in a block to be inserted in a table by INSERT query. Minimal value: 0, block squashing is disabled (default: 268435456, 256 MB). |
maxInsertBlockSize |
Allows to form blocks of the specified size (in bytes) when inserting data in a table. Value must be greater than 0 (default: 1048576). See in-depth description in ClickHouse documentation |
minBytesToUseDirectIo |
Limits the minimum number of bytes to enable unbuffered direct reads from disk (Direct I/O). By default, ClickHouse does not read data directly from disk, but relies on the filesystem and its cache instead. If the total amount of the data to read is greater than the value of this setting, then ClickHouse will fetch this data directly from the disk. Minimal value and default value: 0, Direct I/O is disabled. |
useUncompressedCache |
Determines whether to use the cache of uncompressed blocks, or not. This setting has effect only for tables of the MergeTree family. Default value: false (uncompressed cache is disabled). See in-depth description in ClickHouse documentation |
mergeTreeMaxRowsToUseCache |
Limits the maximum size in rows of the request that can use the cache of uncompressed data. The cache is not used for requests larger Use this setting in combination with Value must be greater than 0 (default: 128x8192). |
mergeTreeMaxBytesToUseCache |
Limits the maximum size in bytes of the request that can use the cache of uncompressed data. The cache is not used for requests larger Use this setting in combination with Value must be greater than 0 (default: 192x10x1024x1024). |
mergeTreeMinRowsForConcurrentRead |
Limits the minimum number of rows to be read from a file to enable concurrent read. This setting has effect only for tables of the MergeTree family. Value must be greater than 0 (default: 20x8192). |
mergeTreeMinBytesForConcurrentRead |
Limits the number of bytes to be read from a file to enable concurrent read. This setting has effect only for tables of the MergeTree family. Value must be greater than 0 (default: 24x10x1024x1024). |
maxBytesBeforeExternalGroupBy |
Sets the threshold of RAM consumption (in bytes) after that the temporary data, collected during the GROUP BY operation, should be flushed to disk to limit the RAM comsumption. By default, aggregation is done by employing hash table that resides in RAM. Minimal value and default value: 0, GROUP BY in the external memory is disabled. When using aggregation in external memory, it is recommended to set the value of this setting twice as low as the See in-depth description in ClickHouse documentation See also: the |
maxBytesBeforeExternalSort |
This setting is equivalent of the |
groupByTwoLevelThreshold |
Sets the threshold of the number of keys, after that the two-level aggregation should be used. Minimal value: 0, threshold is not set (default: 10000). |
groupByTwoLevelThresholdBytes |
Sets the threshold of the number of bytes, after that the two-level aggregation should be used. Minimal value: 0, threshold is not set (default: 100000000). |
priority |
Sets the priority of a query.
This setting should be set up for each query individually. If ClickHouse is working with the high-priority queries, and a low-priority query enters, then the low-priority query is paused until higher-priority queries are completed. Minimal value and default value: 0, priority is not used. |
maxThreads |
Limits the maximum number of threads to process the request (setting does not take threads that read data from remote servers into account). This setting applies to threads that perform the same stages of the query processing pipeline in parallel. Minimal value and default value: 0 (the thread number is calculated automatically based on the number of physical CPU cores, no HyperThreading cores are taken into account). See in-depth description in ClickHouse documentation |
maxMemoryUsage |
Limits the maximum memory usage (in bytes) for processing of a single user's query on a single server. This limitation is enforced for any user's single query on a single server. Minimal value: 0, no limitation is set. If you use See in-depth description in ClickHouse documentation |
maxMemoryUsageForUser |
Limits the maximum memory usage (in bytes) for processing of user's queries on a single server. This limitation is enforced for all queries that belong to one user and run simultaneously on a single server. Minimal value and default value: 0, no limitation is set. |
maxNetworkBandwidth |
The maximum speed of data exchange over the network in bytes per second for a query. Minimal value and default value: 0, no limitation is set. |
maxNetworkBandwidthForUser |
The maximum speed of data exchange over the network in bytes per second for all concurrently running user queries. Minimal value and default value: 0, no limitation is set. |
maxPartitionsPerInsertBlock |
See in-depth description in ClickHouse documentation |
maxConcurrentQueriesForUser |
The maximum number of concurrent requests per user. |
forceIndexByDate |
If enabled, query is not executed if the ClickHouse can't use index by date. Default value: false (setting is disabled, query executes even if ClickHouse can't use index by date). See in-depth description in ClickHouse documentation |
forcePrimaryKey |
If enabled, query is not executed if the ClickHouse can't use index by primary key. Default value: false (setting is disabled, query executes even if ClickHouse can't use index by primary key). See in-depth description in ClickHouse documentation |
maxRowsToRead |
Limits the maximum number of rows that can be read from a table when running a query. Minimal value and default value: 0, no limitation is set. See in-depth description in ClickHouse documentation |
maxBytesToRead |
Limits the maximum number of bytes (uncompressed data) that can be read from a table when running a query. Minimal value and default value: 0, no limitation is set. |
readOverflowMode |
enum OverflowMode Determines the behavior on exceeding limits
|
maxRowsToGroupBy |
Limits the maximum number of unique keys received from aggregation function. Minimal value and default value: 0, no limitation is set. |
groupByOverflowMode |
enum GroupByOverflowMode Determines the behavior on exceeding limits
|
maxRowsToSort |
Limits the maximum number of rows that can be read from a table for sorting. Minimal value and default value: 0, no limitation is set. |
maxBytesToSort |
Limits the maximum number of bytes (uncompressed data) that can be read from a table for sorting. Minimal value and default value: 0, no limitation is set. |
sortOverflowMode |
enum OverflowMode Determines the behavior on exceeding limits
|
maxResultRows |
Limits the number of rows in the result. Minimal value and default value: 0, no limitation is set. |
maxResultBytes |
Limits the number of bytes in the result. Minimal value and default value: 0, no limitation is set. |
resultOverflowMode |
enum OverflowMode Determines the behavior on exceeding limits
|
maxRowsInDistinct |
Limits the maximum number of different rows when using DISTINCT. Minimal value and default value: 0, no limitation is set. |
maxBytesInDistinct |
Limits the maximum size of a hash table in bytes (uncompressed data) when using DISTINCT. |
distinctOverflowMode |
enum OverflowMode Determines the behavior on exceeding limits
|
maxRowsToTransfer |
Limits the maximum number of rows that can be passed to a remote server or saved in a temporary table when using GLOBAL IN. Minimal value and default value: 0, no limitation is set. |
maxBytesToTransfer |
Limits the maximum number of bytes (uncompressed data) that can be passed to a remote server or saved in a temporary Minimal value and default value: 0, no limitation is set. |
transferOverflowMode |
enum OverflowMode Determines the behavior on exceeding limits
|
maxExecutionTime |
Limits the maximum query execution time in milliseconds. Minimal value and default value: 0, no limitation is set. |
timeoutOverflowMode |
enum OverflowMode Determines the behavior on exceeding limits
|
maxRowsInSet |
Limit on the number of rows in the set resulting from the execution of the IN section. |
maxBytesInSet |
Limit on the number of bytes in the set resulting from the execution of the IN section. |
setOverflowMode |
enum OverflowMode Determine the behavior on exceeding max_rows_in_set or max_bytes_in_set limit.
|
maxRowsInJoin |
Limit on maximum size of the hash table for JOIN, in rows. |
maxBytesInJoin |
Limit on maximum size of the hash table for JOIN, in bytes. |
joinOverflowMode |
enum OverflowMode Determine the behavior on exceeding max_rows_in_join or max_bytes_in_join limit.
|
joinAlgorithm[] |
enum JoinAlgorithm See in-depth description in ClickHouse documentation
|
anyJoinDistinctRightTableKeys |
See in-depth description in ClickHouse documentation |
maxColumnsToRead |
Limits the maximum number of columns that can be read from a table in a single query. Minimal value and default value: 0, no limitation is set. |
maxTemporaryColumns |
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, including constant columns. Minimal value and default value: 0, no limitation is set. |
maxTemporaryNonConstColumns |
Limits the maximum number of temporary columns that must be kept in RAM at the same time when running a query, excluding constant columns. Minimal value and default value: 0, no limitation is set. |
maxQuerySize |
Limits the size of the part of a query that can be transferred to RAM for parsing with the SQL parser, in bytes. Value must be greater than 0 (default: 262144). See in-depth description in ClickHouse documentation |
maxAstDepth |
Limits the maximum depth of query syntax tree. Executing a big and complex query may result in building a syntax tree of enormous depth. For example, the **SELECT *** query may result in more complex and deeper syntax tree, compared to the SELECT ... WHERE ... query, containing constraints and conditions, in the most cases. Value must be greater than 0 (default: 1000). See in-depth description in ClickHouse documentation |
maxAstElements |
Limits the maximum size of query syntax tree in number of nodes. Executing a big and complex query may result in building a syntax tree of enormous size. Value must be greater than 0 (default: 50000). See in-depth description in ClickHouse documentation |
maxExpandedAstElements |
Limits the maximum size of query syntax tree in number of nodes after expansion of aliases and the asterisk values. Executing a big and complex query may result in building a syntax tree of enormous size. Value must be greater than 0 (default: 500000). |
minExecutionSpeed |
Minimal execution speed in rows per second. |
minExecutionSpeedBytes |
Minimal execution speed in bytes per second. |
countDistinctImplementation |
enum CountDistinctImplementation Aggregate function to use for implementation of count(DISTINCT ...).
|
inputFormatValuesInterpretExpressions |
Enables or disables SQL parser if the fast stream parser cannot parse the data. Enable this setting, if the data that you want to insert into a table contains SQL expressions. For example, the stream parser is unable to parse a value that contains now() expression; therefore an INSERT query for this value will fail and no data will be inserted into a table. This setting has effect only if you use Values Default value: true (SQL parser is enabled). See in-depth description in ClickHouse documentation |
inputFormatDefaultsForOmittedFields |
Enables or disables replacing omitted input values with default values of the respective columns when performing INSERT queries. Default value: true (replacing is enabled). |
inputFormatNullAsDefault |
See in-depth description in ClickHouse documentation |
dateTimeInputFormat |
enum DateTimeInputFormat See in-depth description in ClickHouse documentation
|
inputFormatWithNamesUseHeader |
See in-depth description in ClickHouse documentation |
outputFormatJsonQuote_64bitIntegers |
Enables quoting of 64-bit integers in JSON output format. If this setting is enabled, then 64-bit integers (UInt64 and Int64) will be quoted when written to JSON output in order to maintain compatibility with the most of the JavaScript engines. Default value: false (quoting 64-bit integers is disabled). |
outputFormatJsonQuoteDenormals |
Enables special floating-point values (+nan, -nan, +inf and -inf) in JSON output format. Default value: false (special values do not present in output). |
dateTimeOutputFormat |
enum DateTimeOutputFormat See in-depth description in ClickHouse documentation
|
lowCardinalityAllowInNativeFormat |
Determines whether to use LowCardinality type in Native format.
LowCardinality columns (aka sparse columns) store data in more effective way, compared to regular columns, by using hash tables. If you use a third-party ClickHouse client that can't work with LowCardinality columns, then this client will not be able to correctly interpret the result of the query that asks for data stored in LowCardinality column. Official ClickHouse client works with LowCardinality columns out-of-the-box. Default value: true (LowCardinality columns are used in Native format). |
allowSuspiciousLowCardinalityTypes |
Allows specifying LowCardinality modifier for types of small fixed size (8 or less) in CREATE TABLE statements. Enabling this may increase merge times and memory consumption. More info see in ClickHouse documentation |
emptyResultForAggregationByEmptySet |
Enables returning of empty result when aggregating without keys (with GROUP BY operation absent) on empty set (e.g., SELECT count(*) FROM table WHERE 0).
|
httpConnectionTimeout |
HTTP connection timeout, in milliseconds. Value must be greater than 0 (default: 1000, 1 second). |
httpReceiveTimeout |
HTTP receive timeout, in milliseconds. Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes). |
httpSendTimeout |
HTTP send timeout, in milliseconds. Value must be greater than 0 (default: 1800000, 1800 seconds, 30 minutes). |
enableHttpCompression |
Enables or disables data compression in HTTP responses. By default, ClickHouse stores data compressed. When executing a query, its result is uncompressed. Enable this setting and add the Accept-Encoding: ClickHouse support the following compression methods: gzip, br and deflate. Default value: false (compression is disabled). See in-depth description in ClickHouse documentation |
sendProgressInHttpHeaders |
Enables progress notifications using X-ClickHouse-Progress HTTP header. Default value: false (notifications disabled). |
httpHeadersProgressInterval |
Minimum interval between progress notifications with X-ClickHouse-Progress HTTP header, in milliseconds. Value must be greater than 0 (default: 100). |
addHttpCorsHeader |
Adds CORS header in HTTP responses. Default value: false (header is not added). |
cancelHttpReadonlyQueriesOnClientClose |
Cancels HTTP read-only queries (e.g. SELECT) when a client closes the connection without waiting for the response. Default value: false. More info see in ClickHouse documentation |
maxHttpGetRedirects |
Limits the maximum number of HTTP GET redirect hops for URL-engine If the parameter is set to 0 (default), no hops is allowed. More info see in ClickHouse documentation |
joinedSubqueryRequiresAlias |
|
joinUseNulls |
|
transformNullIn |
|
quotaMode |
enum QuotaMode Quota accounting mode. Possible values: QUOTA_MODE_DEFAULT, QUOTA_MODE_KEYED and QUOTA_MODE_KEYED_BY_IP.
|
flattenNested |
Sets the data format of a nested More info see in ClickHouse documentation |
formatRegexp |
string Regular expression (for Regexp format) |
formatRegexpEscapingRule |
enum FormatRegexpEscapingRule See in-depth description in ClickHouse documentation
|
formatRegexpSkipUnmatched |
See in-depth description in ClickHouse documentation |
asyncInsert |
Enables asynchronous inserts. Disabled by default. More info see in ClickHouse documentation |
asyncInsertThreads |
The maximum number of threads for background data parsing and insertion. If the parameter is set to 0, asynchronous insertions are disabled. Default value: 16. More info see in ClickHouse documentation |
waitForAsyncInsert |
Enables waiting for processing of asynchronous insertion. If enabled, server returns OK only after the data is inserted. More info see in ClickHouse documentation |
waitForAsyncInsertTimeout |
The timeout (in seconds) for waiting for processing of asynchronous insertion. Default value: 120. More info see in ClickHouse documentation |
asyncInsertMaxDataSize |
The maximum size of the unparsed data in bytes collected per query before being inserted. If the parameter is set to 0, asynchronous insertions are disabled. Default value: 100000. More info see in ClickHouse documentation |
asyncInsertBusyTimeout |
The maximum timeout in milliseconds since the first INSERT query before inserting collected data. If the parameter is set to 0, the timeout is disabled. Default value: 200. More info see in ClickHouse documentation |
asyncInsertStaleTimeout |
The maximum timeout in milliseconds since the last INSERT query before dumping collected data. If enabled, the settings prolongs the More info see in ClickHouse documentation |
memoryProfilerStep |
Memory profiler step (in bytes). If the next query step requires more memory than this parameter specifies, the memory profiler collects the allocating stack trace. Values lower than a few megabytes slow down query processing. Default value: 4194304 (4 MB). Zero means disabled memory profiler. |
memoryProfilerSampleProbability |
Collect random allocations and deallocations and write them into system.trace_log with 'MemorySample' trace_type. The probability is for every alloc/free regardless to the size of the allocation. Possible values: from 0 to 1. Default: 0. |
maxFinalThreads |
Sets the maximum number of parallel threads for the SELECT query data read phase with the FINAL modifier. |
inputFormatParallelParsing |
Enables or disables order-preserving parallel parsing of data formats. Supported only for TSV |
inputFormatImportNestedJson |
Enables or disables the insertion of JSON data with nested objects. |
localFilesystemReadMethod |
enum LocalFilesystemReadMethod Method of reading data from local filesystem, one of: read, pread, mmap, io_uring, pread_threadpool. The 'io_uring' method is experimental and does not work for Log, TinyLog, StripeLog, File, Set and Join, and other tables with append-able files in presence of concurrent reads and writes.
|
maxReadBufferSize |
The maximum size of the buffer to read from the filesystem. |
insertKeeperMaxRetries |
The setting sets the maximum number of retries for ClickHouse Keeper (or ZooKeeper) requests during insert into replicated MergeTree. Only Keeper requests which failed due to network error, Keeper session timeout, or request timeout are considered for retries. |
maxTemporaryDataOnDiskSizeForUser |
The maximum amount of data consumed by temporary files on disk in bytes for all concurrently running user queries. Zero means unlimited. |
maxTemporaryDataOnDiskSizeForQuery |
The maximum amount of data consumed by temporary files on disk in bytes for all concurrently running queries. Zero means unlimited. |
maxParserDepth |
Limits maximum recursion depth in the recursive descent parser. Allows controlling the stack size. |
remoteFilesystemReadMethod |
enum RemoteFilesystemReadMethod Method of reading data from remote filesystem, one of: read, threadpool.
|
memoryOvercommitRatioDenominator |
It represents soft memory limit in case when hard limit is reached on user level. This value is used to compute overcommit ratio for the query. Zero means skip the query. |
memoryOvercommitRatioDenominatorForUser |
It represents soft memory limit in case when hard limit is reached on global level. This value is used to compute overcommit ratio for the query. Zero means skip the query. |
memoryUsageOvercommitMaxWaitMicroseconds |
Maximum time thread will wait for memory to be freed in the case of memory overcommit on a user level. If the timeout is reached and memory is not freed, an exception is thrown. |
logQueryThreads |
Setting up query threads logging. Query threads log into the system.query_thread_log |
maxInsertThreads |
The maximum number of threads to execute the INSERT SELECT query. |
useHedgedRequests |
Enables hedged requests logic for remote queries. It allows to establish many connections with different replicas for query. New connection is enabled in case existent connection(s) with replica(s) were not established within hedged_connection_timeout or no data was received within receive_data_timeout. Query uses the first connection which send non empty progress packet (or data packet, if allow_changing_replica_until_first_data_packet); other connections are cancelled. Queries with max_parallel_replicas > 1 are supported. |
idleConnectionTimeout |
Timeout to close idle TCP connections after specified number of milliseconds. |
hedgedConnectionTimeoutMs |
Connection timeout for establishing connection with replica for Hedged requests. |
loadBalancing |
enum LoadBalancing Specifies the algorithm of replicas selection that is used for distributed query processing, one of: random, nearest_hostname, in_order, first_or_random, round_robin.
|
preferLocalhostReplica |
Enables/disables preferable using the localhost replica when processing distributed queries. |
compile |
The setting is deprecated and has no effect. |
minCountToCompile |
The setting is deprecated and has no effect. |
UserQuota
ClickHouse quota representation. Each quota associated with an user and limits it resource usage for an interval.
See in-depth description ClickHouse documentation
Field |
Description |
intervalDuration |
Duration of interval for quota in milliseconds. |
queries |
The total number of queries. |
errors |
The number of queries that threw exception. |
resultRows |
The total number of rows given as the result.. |
readRows |
The total number of source rows read from tables for running the query, on all remote servers. |
executionTime |
The total query execution time, in milliseconds (wall time). |
HostSpec
Field |
Description |
zoneId |
string ID of the availability zone where the host resides. |
type |
enum Type Required field. Type of the host to be deployed.
|
subnetId |
string ID of the subnet that the host should belong to. This subnet should be a part |
assignPublicIp |
bool Whether the host should get a public IP address on creation. After a host has been created, this setting cannot be changed. To remove an assigned public IP, or to assign Possible values:
|
shardName |
string Name of the shard that the host is assigned to. |
MaintenanceWindow
A maintenance window settings.
Field |
Description |
anytime |
Maintenance operation can be scheduled anytime. Includes only one of the fields The maintenance policy in effect. |
weeklyMaintenanceWindow |
Maintenance operation can be scheduled on a weekly basis. Includes only one of the fields The maintenance policy in effect. |
AnytimeMaintenanceWindow
Field |
Description |
Empty |
WeeklyMaintenanceWindow
Weelky maintenance window settings.
Field |
Description |
day |
enum WeekDay Day of the week (in
|
hour |
int64 Hour of the day in UTC (in |
operation.Operation
{
"id": "string",
"description": "string",
"createdAt": "google.protobuf.Timestamp",
"createdBy": "string",
"modifiedAt": "google.protobuf.Timestamp",
"done": "bool",
"metadata": {
"clusterId": "string"
},
// Includes only one of the fields `error`, `response`
"error": "google.rpc.Status",
"response": {
"id": "string",
"folderId": "string",
"createdAt": "google.protobuf.Timestamp",
"name": "string",
"description": "string",
"labels": "string",
"environment": "Environment",
"monitoring": [
{
"name": "string",
"description": "string",
"link": "string"
}
],
"config": {
"version": "string",
"clickhouse": {
"config": {
"effectiveConfig": {
"logLevel": "LogLevel",
"mergeTree": {
"replicatedDeduplicationWindow": "google.protobuf.Int64Value",
"replicatedDeduplicationWindowSeconds": "google.protobuf.Int64Value",
"partsToDelayInsert": "google.protobuf.Int64Value",
"partsToThrowInsert": "google.protobuf.Int64Value",
"inactivePartsToDelayInsert": "google.protobuf.Int64Value",
"inactivePartsToThrowInsert": "google.protobuf.Int64Value",
"maxReplicatedMergesInQueue": "google.protobuf.Int64Value",
"numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge": "google.protobuf.Int64Value",
"maxBytesToMergeAtMinSpaceInPool": "google.protobuf.Int64Value",
"maxBytesToMergeAtMaxSpaceInPool": "google.protobuf.Int64Value",
"minBytesForWidePart": "google.protobuf.Int64Value",
"minRowsForWidePart": "google.protobuf.Int64Value",
"ttlOnlyDropParts": "google.protobuf.BoolValue",
"allowRemoteFsZeroCopyReplication": "google.protobuf.BoolValue",
"mergeWithTtlTimeout": "google.protobuf.Int64Value",
"mergeWithRecompressionTtlTimeout": "google.protobuf.Int64Value",
"maxPartsInTotal": "google.protobuf.Int64Value",
"maxNumberOfMergesWithTtlInPool": "google.protobuf.Int64Value",
"cleanupDelayPeriod": "google.protobuf.Int64Value",
"numberOfFreeEntriesInPoolToExecuteMutation": "google.protobuf.Int64Value",
"maxAvgPartSizeForTooManyParts": "google.protobuf.Int64Value",
"minAgeToForceMergeSeconds": "google.protobuf.Int64Value",
"minAgeToForceMergeOnPartitionOnly": "google.protobuf.BoolValue",
"mergeSelectingSleepMs": "google.protobuf.Int64Value",
"mergeMaxBlockSize": "google.protobuf.Int64Value",
"checkSampleColumnIsCorrect": "google.protobuf.BoolValue",
"maxMergeSelectingSleepMs": "google.protobuf.Int64Value",
"maxCleanupDelayPeriod": "google.protobuf.Int64Value"
},
"compression": [
{
"method": "Method",
"minPartSize": "int64",
"minPartSizeRatio": "double",
"level": "google.protobuf.Int64Value"
}
],
"dictionaries": [
{
"name": "string",
"structure": {
"id": {
"name": "string"
},
"key": {
"attributes": [
{
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
}
]
},
"rangeMin": {
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
},
"rangeMax": {
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
},
"attributes": [
{
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
}
]
},
"layout": {
"type": "Type",
"sizeInCells": "int64",
"maxArraySize": "int64"
},
// Includes only one of the fields `fixedLifetime`, `lifetimeRange`
"fixedLifetime": "int64",
"lifetimeRange": {
"min": "int64",
"max": "int64"
},
// end of the list of possible fields
// Includes only one of the fields `httpSource`, `mysqlSource`, `clickhouseSource`, `mongodbSource`, `postgresqlSource`
"httpSource": {
"url": "string",
"format": "string",
"headers": [
{
"name": "string",
"value": "string"
}
]
},
"mysqlSource": {
"db": "string",
"table": "string",
"port": "int64",
"user": "string",
"password": "string",
"replicas": [
{
"host": "string",
"priority": "int64",
"port": "int64",
"user": "string",
"password": "string"
}
],
"where": "string",
"invalidateQuery": "string",
"closeConnection": "google.protobuf.BoolValue",
"shareConnection": "google.protobuf.BoolValue"
},
"clickhouseSource": {
"db": "string",
"table": "string",
"host": "string",
"port": "int64",
"user": "string",
"password": "string",
"where": "string",
"secure": "google.protobuf.BoolValue"
},
"mongodbSource": {
"db": "string",
"collection": "string",
"host": "string",
"port": "int64",
"user": "string",
"password": "string",
"options": "string"
},
"postgresqlSource": {
"db": "string",
"table": "string",
"hosts": [
"string"
],
"port": "int64",
"user": "string",
"password": "string",
"invalidateQuery": "string",
"sslMode": "SslMode"
}
// end of the list of possible fields
}
],
"graphiteRollup": [
{
"name": "string",
"patterns": [
{
"regexp": "string",
"function": "string",
"retention": [
{
"age": "int64",
"precision": "int64"
}
]
}
],
"pathColumnName": "string",
"timeColumnName": "string",
"valueColumnName": "string",
"versionColumnName": "string"
}
],
"kafka": {
"securityProtocol": "SecurityProtocol",
"saslMechanism": "SaslMechanism",
"saslUsername": "string",
"saslPassword": "string",
"enableSslCertificateVerification": "google.protobuf.BoolValue",
"maxPollIntervalMs": "google.protobuf.Int64Value",
"sessionTimeoutMs": "google.protobuf.Int64Value",
"debug": "Debug",
"autoOffsetReset": "AutoOffsetReset"
},
"kafkaTopics": [
{
"name": "string",
"settings": {
"securityProtocol": "SecurityProtocol",
"saslMechanism": "SaslMechanism",
"saslUsername": "string",
"saslPassword": "string",
"enableSslCertificateVerification": "google.protobuf.BoolValue",
"maxPollIntervalMs": "google.protobuf.Int64Value",
"sessionTimeoutMs": "google.protobuf.Int64Value",
"debug": "Debug",
"autoOffsetReset": "AutoOffsetReset"
}
}
],
"rabbitmq": {
"username": "string",
"password": "string",
"vhost": "string"
},
"maxConnections": "google.protobuf.Int64Value",
"maxConcurrentQueries": "google.protobuf.Int64Value",
"keepAliveTimeout": "google.protobuf.Int64Value",
"uncompressedCacheSize": "google.protobuf.Int64Value",
"markCacheSize": "google.protobuf.Int64Value",
"maxTableSizeToDrop": "google.protobuf.Int64Value",
"maxPartitionSizeToDrop": "google.protobuf.Int64Value",
"builtinDictionariesReloadInterval": "google.protobuf.Int64Value",
"timezone": "string",
"geobaseEnabled": "google.protobuf.BoolValue",
"geobaseUri": "string",
"queryLogRetentionSize": "google.protobuf.Int64Value",
"queryLogRetentionTime": "google.protobuf.Int64Value",
"queryThreadLogEnabled": "google.protobuf.BoolValue",
"queryThreadLogRetentionSize": "google.protobuf.Int64Value",
"queryThreadLogRetentionTime": "google.protobuf.Int64Value",
"partLogRetentionSize": "google.protobuf.Int64Value",
"partLogRetentionTime": "google.protobuf.Int64Value",
"metricLogEnabled": "google.protobuf.BoolValue",
"metricLogRetentionSize": "google.protobuf.Int64Value",
"metricLogRetentionTime": "google.protobuf.Int64Value",
"traceLogEnabled": "google.protobuf.BoolValue",
"traceLogRetentionSize": "google.protobuf.Int64Value",
"traceLogRetentionTime": "google.protobuf.Int64Value",
"textLogEnabled": "google.protobuf.BoolValue",
"textLogRetentionSize": "google.protobuf.Int64Value",
"textLogRetentionTime": "google.protobuf.Int64Value",
"textLogLevel": "LogLevel",
"opentelemetrySpanLogEnabled": "google.protobuf.BoolValue",
"opentelemetrySpanLogRetentionSize": "google.protobuf.Int64Value",
"opentelemetrySpanLogRetentionTime": "google.protobuf.Int64Value",
"queryViewsLogEnabled": "google.protobuf.BoolValue",
"queryViewsLogRetentionSize": "google.protobuf.Int64Value",
"queryViewsLogRetentionTime": "google.protobuf.Int64Value",
"asynchronousMetricLogEnabled": "google.protobuf.BoolValue",
"asynchronousMetricLogRetentionSize": "google.protobuf.Int64Value",
"asynchronousMetricLogRetentionTime": "google.protobuf.Int64Value",
"sessionLogEnabled": "google.protobuf.BoolValue",
"sessionLogRetentionSize": "google.protobuf.Int64Value",
"sessionLogRetentionTime": "google.protobuf.Int64Value",
"zookeeperLogEnabled": "google.protobuf.BoolValue",
"zookeeperLogRetentionSize": "google.protobuf.Int64Value",
"zookeeperLogRetentionTime": "google.protobuf.Int64Value",
"asynchronousInsertLogEnabled": "google.protobuf.BoolValue",
"asynchronousInsertLogRetentionSize": "google.protobuf.Int64Value",
"asynchronousInsertLogRetentionTime": "google.protobuf.Int64Value",
"backgroundPoolSize": "google.protobuf.Int64Value",
"backgroundMergesMutationsConcurrencyRatio": "google.protobuf.Int64Value",
"backgroundSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundFetchesPoolSize": "google.protobuf.Int64Value",
"backgroundMovePoolSize": "google.protobuf.Int64Value",
"backgroundDistributedSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundBufferFlushSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundMessageBrokerSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundCommonPoolSize": "google.protobuf.Int64Value",
"defaultDatabase": "google.protobuf.StringValue",
"totalMemoryProfilerStep": "google.protobuf.Int64Value",
"totalMemoryTrackerSampleProbability": "google.protobuf.DoubleValue",
"queryMaskingRules": [
{
"name": "string",
"regexp": "string",
"replace": "string"
}
],
"dictionariesLazyLoad": "google.protobuf.BoolValue",
"queryCache": {
"maxSizeInBytes": "google.protobuf.Int64Value",
"maxEntries": "google.protobuf.Int64Value",
"maxEntrySizeInBytes": "google.protobuf.Int64Value",
"maxEntrySizeInRows": "google.protobuf.Int64Value"
}
},
"userConfig": {
"logLevel": "LogLevel",
"mergeTree": {
"replicatedDeduplicationWindow": "google.protobuf.Int64Value",
"replicatedDeduplicationWindowSeconds": "google.protobuf.Int64Value",
"partsToDelayInsert": "google.protobuf.Int64Value",
"partsToThrowInsert": "google.protobuf.Int64Value",
"inactivePartsToDelayInsert": "google.protobuf.Int64Value",
"inactivePartsToThrowInsert": "google.protobuf.Int64Value",
"maxReplicatedMergesInQueue": "google.protobuf.Int64Value",
"numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge": "google.protobuf.Int64Value",
"maxBytesToMergeAtMinSpaceInPool": "google.protobuf.Int64Value",
"maxBytesToMergeAtMaxSpaceInPool": "google.protobuf.Int64Value",
"minBytesForWidePart": "google.protobuf.Int64Value",
"minRowsForWidePart": "google.protobuf.Int64Value",
"ttlOnlyDropParts": "google.protobuf.BoolValue",
"allowRemoteFsZeroCopyReplication": "google.protobuf.BoolValue",
"mergeWithTtlTimeout": "google.protobuf.Int64Value",
"mergeWithRecompressionTtlTimeout": "google.protobuf.Int64Value",
"maxPartsInTotal": "google.protobuf.Int64Value",
"maxNumberOfMergesWithTtlInPool": "google.protobuf.Int64Value",
"cleanupDelayPeriod": "google.protobuf.Int64Value",
"numberOfFreeEntriesInPoolToExecuteMutation": "google.protobuf.Int64Value",
"maxAvgPartSizeForTooManyParts": "google.protobuf.Int64Value",
"minAgeToForceMergeSeconds": "google.protobuf.Int64Value",
"minAgeToForceMergeOnPartitionOnly": "google.protobuf.BoolValue",
"mergeSelectingSleepMs": "google.protobuf.Int64Value",
"mergeMaxBlockSize": "google.protobuf.Int64Value",
"checkSampleColumnIsCorrect": "google.protobuf.BoolValue",
"maxMergeSelectingSleepMs": "google.protobuf.Int64Value",
"maxCleanupDelayPeriod": "google.protobuf.Int64Value"
},
"compression": [
{
"method": "Method",
"minPartSize": "int64",
"minPartSizeRatio": "double",
"level": "google.protobuf.Int64Value"
}
],
"dictionaries": [
{
"name": "string",
"structure": {
"id": {
"name": "string"
},
"key": {
"attributes": [
{
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
}
]
},
"rangeMin": {
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
},
"rangeMax": {
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
},
"attributes": [
{
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
}
]
},
"layout": {
"type": "Type",
"sizeInCells": "int64",
"maxArraySize": "int64"
},
// Includes only one of the fields `fixedLifetime`, `lifetimeRange`
"fixedLifetime": "int64",
"lifetimeRange": {
"min": "int64",
"max": "int64"
},
// end of the list of possible fields
// Includes only one of the fields `httpSource`, `mysqlSource`, `clickhouseSource`, `mongodbSource`, `postgresqlSource`
"httpSource": {
"url": "string",
"format": "string",
"headers": [
{
"name": "string",
"value": "string"
}
]
},
"mysqlSource": {
"db": "string",
"table": "string",
"port": "int64",
"user": "string",
"password": "string",
"replicas": [
{
"host": "string",
"priority": "int64",
"port": "int64",
"user": "string",
"password": "string"
}
],
"where": "string",
"invalidateQuery": "string",
"closeConnection": "google.protobuf.BoolValue",
"shareConnection": "google.protobuf.BoolValue"
},
"clickhouseSource": {
"db": "string",
"table": "string",
"host": "string",
"port": "int64",
"user": "string",
"password": "string",
"where": "string",
"secure": "google.protobuf.BoolValue"
},
"mongodbSource": {
"db": "string",
"collection": "string",
"host": "string",
"port": "int64",
"user": "string",
"password": "string",
"options": "string"
},
"postgresqlSource": {
"db": "string",
"table": "string",
"hosts": [
"string"
],
"port": "int64",
"user": "string",
"password": "string",
"invalidateQuery": "string",
"sslMode": "SslMode"
}
// end of the list of possible fields
}
],
"graphiteRollup": [
{
"name": "string",
"patterns": [
{
"regexp": "string",
"function": "string",
"retention": [
{
"age": "int64",
"precision": "int64"
}
]
}
],
"pathColumnName": "string",
"timeColumnName": "string",
"valueColumnName": "string",
"versionColumnName": "string"
}
],
"kafka": {
"securityProtocol": "SecurityProtocol",
"saslMechanism": "SaslMechanism",
"saslUsername": "string",
"saslPassword": "string",
"enableSslCertificateVerification": "google.protobuf.BoolValue",
"maxPollIntervalMs": "google.protobuf.Int64Value",
"sessionTimeoutMs": "google.protobuf.Int64Value",
"debug": "Debug",
"autoOffsetReset": "AutoOffsetReset"
},
"kafkaTopics": [
{
"name": "string",
"settings": {
"securityProtocol": "SecurityProtocol",
"saslMechanism": "SaslMechanism",
"saslUsername": "string",
"saslPassword": "string",
"enableSslCertificateVerification": "google.protobuf.BoolValue",
"maxPollIntervalMs": "google.protobuf.Int64Value",
"sessionTimeoutMs": "google.protobuf.Int64Value",
"debug": "Debug",
"autoOffsetReset": "AutoOffsetReset"
}
}
],
"rabbitmq": {
"username": "string",
"password": "string",
"vhost": "string"
},
"maxConnections": "google.protobuf.Int64Value",
"maxConcurrentQueries": "google.protobuf.Int64Value",
"keepAliveTimeout": "google.protobuf.Int64Value",
"uncompressedCacheSize": "google.protobuf.Int64Value",
"markCacheSize": "google.protobuf.Int64Value",
"maxTableSizeToDrop": "google.protobuf.Int64Value",
"maxPartitionSizeToDrop": "google.protobuf.Int64Value",
"builtinDictionariesReloadInterval": "google.protobuf.Int64Value",
"timezone": "string",
"geobaseEnabled": "google.protobuf.BoolValue",
"geobaseUri": "string",
"queryLogRetentionSize": "google.protobuf.Int64Value",
"queryLogRetentionTime": "google.protobuf.Int64Value",
"queryThreadLogEnabled": "google.protobuf.BoolValue",
"queryThreadLogRetentionSize": "google.protobuf.Int64Value",
"queryThreadLogRetentionTime": "google.protobuf.Int64Value",
"partLogRetentionSize": "google.protobuf.Int64Value",
"partLogRetentionTime": "google.protobuf.Int64Value",
"metricLogEnabled": "google.protobuf.BoolValue",
"metricLogRetentionSize": "google.protobuf.Int64Value",
"metricLogRetentionTime": "google.protobuf.Int64Value",
"traceLogEnabled": "google.protobuf.BoolValue",
"traceLogRetentionSize": "google.protobuf.Int64Value",
"traceLogRetentionTime": "google.protobuf.Int64Value",
"textLogEnabled": "google.protobuf.BoolValue",
"textLogRetentionSize": "google.protobuf.Int64Value",
"textLogRetentionTime": "google.protobuf.Int64Value",
"textLogLevel": "LogLevel",
"opentelemetrySpanLogEnabled": "google.protobuf.BoolValue",
"opentelemetrySpanLogRetentionSize": "google.protobuf.Int64Value",
"opentelemetrySpanLogRetentionTime": "google.protobuf.Int64Value",
"queryViewsLogEnabled": "google.protobuf.BoolValue",
"queryViewsLogRetentionSize": "google.protobuf.Int64Value",
"queryViewsLogRetentionTime": "google.protobuf.Int64Value",
"asynchronousMetricLogEnabled": "google.protobuf.BoolValue",
"asynchronousMetricLogRetentionSize": "google.protobuf.Int64Value",
"asynchronousMetricLogRetentionTime": "google.protobuf.Int64Value",
"sessionLogEnabled": "google.protobuf.BoolValue",
"sessionLogRetentionSize": "google.protobuf.Int64Value",
"sessionLogRetentionTime": "google.protobuf.Int64Value",
"zookeeperLogEnabled": "google.protobuf.BoolValue",
"zookeeperLogRetentionSize": "google.protobuf.Int64Value",
"zookeeperLogRetentionTime": "google.protobuf.Int64Value",
"asynchronousInsertLogEnabled": "google.protobuf.BoolValue",
"asynchronousInsertLogRetentionSize": "google.protobuf.Int64Value",
"asynchronousInsertLogRetentionTime": "google.protobuf.Int64Value",
"backgroundPoolSize": "google.protobuf.Int64Value",
"backgroundMergesMutationsConcurrencyRatio": "google.protobuf.Int64Value",
"backgroundSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundFetchesPoolSize": "google.protobuf.Int64Value",
"backgroundMovePoolSize": "google.protobuf.Int64Value",
"backgroundDistributedSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundBufferFlushSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundMessageBrokerSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundCommonPoolSize": "google.protobuf.Int64Value",
"defaultDatabase": "google.protobuf.StringValue",
"totalMemoryProfilerStep": "google.protobuf.Int64Value",
"totalMemoryTrackerSampleProbability": "google.protobuf.DoubleValue",
"queryMaskingRules": [
{
"name": "string",
"regexp": "string",
"replace": "string"
}
],
"dictionariesLazyLoad": "google.protobuf.BoolValue",
"queryCache": {
"maxSizeInBytes": "google.protobuf.Int64Value",
"maxEntries": "google.protobuf.Int64Value",
"maxEntrySizeInBytes": "google.protobuf.Int64Value",
"maxEntrySizeInRows": "google.protobuf.Int64Value"
}
},
"defaultConfig": {
"logLevel": "LogLevel",
"mergeTree": {
"replicatedDeduplicationWindow": "google.protobuf.Int64Value",
"replicatedDeduplicationWindowSeconds": "google.protobuf.Int64Value",
"partsToDelayInsert": "google.protobuf.Int64Value",
"partsToThrowInsert": "google.protobuf.Int64Value",
"inactivePartsToDelayInsert": "google.protobuf.Int64Value",
"inactivePartsToThrowInsert": "google.protobuf.Int64Value",
"maxReplicatedMergesInQueue": "google.protobuf.Int64Value",
"numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge": "google.protobuf.Int64Value",
"maxBytesToMergeAtMinSpaceInPool": "google.protobuf.Int64Value",
"maxBytesToMergeAtMaxSpaceInPool": "google.protobuf.Int64Value",
"minBytesForWidePart": "google.protobuf.Int64Value",
"minRowsForWidePart": "google.protobuf.Int64Value",
"ttlOnlyDropParts": "google.protobuf.BoolValue",
"allowRemoteFsZeroCopyReplication": "google.protobuf.BoolValue",
"mergeWithTtlTimeout": "google.protobuf.Int64Value",
"mergeWithRecompressionTtlTimeout": "google.protobuf.Int64Value",
"maxPartsInTotal": "google.protobuf.Int64Value",
"maxNumberOfMergesWithTtlInPool": "google.protobuf.Int64Value",
"cleanupDelayPeriod": "google.protobuf.Int64Value",
"numberOfFreeEntriesInPoolToExecuteMutation": "google.protobuf.Int64Value",
"maxAvgPartSizeForTooManyParts": "google.protobuf.Int64Value",
"minAgeToForceMergeSeconds": "google.protobuf.Int64Value",
"minAgeToForceMergeOnPartitionOnly": "google.protobuf.BoolValue",
"mergeSelectingSleepMs": "google.protobuf.Int64Value",
"mergeMaxBlockSize": "google.protobuf.Int64Value",
"checkSampleColumnIsCorrect": "google.protobuf.BoolValue",
"maxMergeSelectingSleepMs": "google.protobuf.Int64Value",
"maxCleanupDelayPeriod": "google.protobuf.Int64Value"
},
"compression": [
{
"method": "Method",
"minPartSize": "int64",
"minPartSizeRatio": "double",
"level": "google.protobuf.Int64Value"
}
],
"dictionaries": [
{
"name": "string",
"structure": {
"id": {
"name": "string"
},
"key": {
"attributes": [
{
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
}
]
},
"rangeMin": {
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
},
"rangeMax": {
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
},
"attributes": [
{
"name": "string",
"type": "string",
"nullValue": "string",
"expression": "string",
"hierarchical": "bool",
"injective": "bool"
}
]
},
"layout": {
"type": "Type",
"sizeInCells": "int64",
"maxArraySize": "int64"
},
// Includes only one of the fields `fixedLifetime`, `lifetimeRange`
"fixedLifetime": "int64",
"lifetimeRange": {
"min": "int64",
"max": "int64"
},
// end of the list of possible fields
// Includes only one of the fields `httpSource`, `mysqlSource`, `clickhouseSource`, `mongodbSource`, `postgresqlSource`
"httpSource": {
"url": "string",
"format": "string",
"headers": [
{
"name": "string",
"value": "string"
}
]
},
"mysqlSource": {
"db": "string",
"table": "string",
"port": "int64",
"user": "string",
"password": "string",
"replicas": [
{
"host": "string",
"priority": "int64",
"port": "int64",
"user": "string",
"password": "string"
}
],
"where": "string",
"invalidateQuery": "string",
"closeConnection": "google.protobuf.BoolValue",
"shareConnection": "google.protobuf.BoolValue"
},
"clickhouseSource": {
"db": "string",
"table": "string",
"host": "string",
"port": "int64",
"user": "string",
"password": "string",
"where": "string",
"secure": "google.protobuf.BoolValue"
},
"mongodbSource": {
"db": "string",
"collection": "string",
"host": "string",
"port": "int64",
"user": "string",
"password": "string",
"options": "string"
},
"postgresqlSource": {
"db": "string",
"table": "string",
"hosts": [
"string"
],
"port": "int64",
"user": "string",
"password": "string",
"invalidateQuery": "string",
"sslMode": "SslMode"
}
// end of the list of possible fields
}
],
"graphiteRollup": [
{
"name": "string",
"patterns": [
{
"regexp": "string",
"function": "string",
"retention": [
{
"age": "int64",
"precision": "int64"
}
]
}
],
"pathColumnName": "string",
"timeColumnName": "string",
"valueColumnName": "string",
"versionColumnName": "string"
}
],
"kafka": {
"securityProtocol": "SecurityProtocol",
"saslMechanism": "SaslMechanism",
"saslUsername": "string",
"saslPassword": "string",
"enableSslCertificateVerification": "google.protobuf.BoolValue",
"maxPollIntervalMs": "google.protobuf.Int64Value",
"sessionTimeoutMs": "google.protobuf.Int64Value",
"debug": "Debug",
"autoOffsetReset": "AutoOffsetReset"
},
"kafkaTopics": [
{
"name": "string",
"settings": {
"securityProtocol": "SecurityProtocol",
"saslMechanism": "SaslMechanism",
"saslUsername": "string",
"saslPassword": "string",
"enableSslCertificateVerification": "google.protobuf.BoolValue",
"maxPollIntervalMs": "google.protobuf.Int64Value",
"sessionTimeoutMs": "google.protobuf.Int64Value",
"debug": "Debug",
"autoOffsetReset": "AutoOffsetReset"
}
}
],
"rabbitmq": {
"username": "string",
"password": "string",
"vhost": "string"
},
"maxConnections": "google.protobuf.Int64Value",
"maxConcurrentQueries": "google.protobuf.Int64Value",
"keepAliveTimeout": "google.protobuf.Int64Value",
"uncompressedCacheSize": "google.protobuf.Int64Value",
"markCacheSize": "google.protobuf.Int64Value",
"maxTableSizeToDrop": "google.protobuf.Int64Value",
"maxPartitionSizeToDrop": "google.protobuf.Int64Value",
"builtinDictionariesReloadInterval": "google.protobuf.Int64Value",
"timezone": "string",
"geobaseEnabled": "google.protobuf.BoolValue",
"geobaseUri": "string",
"queryLogRetentionSize": "google.protobuf.Int64Value",
"queryLogRetentionTime": "google.protobuf.Int64Value",
"queryThreadLogEnabled": "google.protobuf.BoolValue",
"queryThreadLogRetentionSize": "google.protobuf.Int64Value",
"queryThreadLogRetentionTime": "google.protobuf.Int64Value",
"partLogRetentionSize": "google.protobuf.Int64Value",
"partLogRetentionTime": "google.protobuf.Int64Value",
"metricLogEnabled": "google.protobuf.BoolValue",
"metricLogRetentionSize": "google.protobuf.Int64Value",
"metricLogRetentionTime": "google.protobuf.Int64Value",
"traceLogEnabled": "google.protobuf.BoolValue",
"traceLogRetentionSize": "google.protobuf.Int64Value",
"traceLogRetentionTime": "google.protobuf.Int64Value",
"textLogEnabled": "google.protobuf.BoolValue",
"textLogRetentionSize": "google.protobuf.Int64Value",
"textLogRetentionTime": "google.protobuf.Int64Value",
"textLogLevel": "LogLevel",
"opentelemetrySpanLogEnabled": "google.protobuf.BoolValue",
"opentelemetrySpanLogRetentionSize": "google.protobuf.Int64Value",
"opentelemetrySpanLogRetentionTime": "google.protobuf.Int64Value",
"queryViewsLogEnabled": "google.protobuf.BoolValue",
"queryViewsLogRetentionSize": "google.protobuf.Int64Value",
"queryViewsLogRetentionTime": "google.protobuf.Int64Value",
"asynchronousMetricLogEnabled": "google.protobuf.BoolValue",
"asynchronousMetricLogRetentionSize": "google.protobuf.Int64Value",
"asynchronousMetricLogRetentionTime": "google.protobuf.Int64Value",
"sessionLogEnabled": "google.protobuf.BoolValue",
"sessionLogRetentionSize": "google.protobuf.Int64Value",
"sessionLogRetentionTime": "google.protobuf.Int64Value",
"zookeeperLogEnabled": "google.protobuf.BoolValue",
"zookeeperLogRetentionSize": "google.protobuf.Int64Value",
"zookeeperLogRetentionTime": "google.protobuf.Int64Value",
"asynchronousInsertLogEnabled": "google.protobuf.BoolValue",
"asynchronousInsertLogRetentionSize": "google.protobuf.Int64Value",
"asynchronousInsertLogRetentionTime": "google.protobuf.Int64Value",
"backgroundPoolSize": "google.protobuf.Int64Value",
"backgroundMergesMutationsConcurrencyRatio": "google.protobuf.Int64Value",
"backgroundSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundFetchesPoolSize": "google.protobuf.Int64Value",
"backgroundMovePoolSize": "google.protobuf.Int64Value",
"backgroundDistributedSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundBufferFlushSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundMessageBrokerSchedulePoolSize": "google.protobuf.Int64Value",
"backgroundCommonPoolSize": "google.protobuf.Int64Value",
"defaultDatabase": "google.protobuf.StringValue",
"totalMemoryProfilerStep": "google.protobuf.Int64Value",
"totalMemoryTrackerSampleProbability": "google.protobuf.DoubleValue",
"queryMaskingRules": [
{
"name": "string",
"regexp": "string",
"replace": "string"
}
],
"dictionariesLazyLoad": "google.protobuf.BoolValue",
"queryCache": {
"maxSizeInBytes": "google.protobuf.Int64Value",
"maxEntries": "google.protobuf.Int64Value",
"maxEntrySizeInBytes": "google.protobuf.Int64Value",
"maxEntrySizeInRows": "google.protobuf.Int64Value"
}
}
},
"resources": {
"resourcePresetId": "string",
"diskSize": "int64",
"diskTypeId": "string"
}
},
"zookeeper": {
"resources": {
"resourcePresetId": "string",
"diskSize": "int64",
"diskTypeId": "string"
}
},
"backupWindowStart": "google.type.TimeOfDay",
"access": {
"dataLens": "bool",
"webSql": "bool",
"metrika": "bool",
"serverless": "bool",
"dataTransfer": "bool",
"yandexQuery": "bool"
},
"cloudStorage": {
"enabled": "bool",
"moveFactor": "google.protobuf.DoubleValue",
"dataCacheEnabled": "google.protobuf.BoolValue",
"dataCacheMaxSize": "google.protobuf.Int64Value",
"preferNotToMerge": "google.protobuf.BoolValue"
},
"sqlDatabaseManagement": "google.protobuf.BoolValue",
"sqlUserManagement": "google.protobuf.BoolValue",
"embeddedKeeper": "google.protobuf.BoolValue",
"backupRetainPeriodDays": "google.protobuf.Int64Value"
},
"networkId": "string",
"health": "Health",
"status": "Status",
"serviceAccountId": "string",
"maintenanceWindow": {
// Includes only one of the fields `anytime`, `weeklyMaintenanceWindow`
"anytime": "AnytimeMaintenanceWindow",
"weeklyMaintenanceWindow": {
"day": "WeekDay",
"hour": "int64"
}
// end of the list of possible fields
},
"plannedOperation": {
"info": "string",
"delayedUntil": "google.protobuf.Timestamp"
},
"securityGroupIds": [
"string"
],
"deletionProtection": "bool"
}
// end of the list of possible fields
}
An Operation resource. For more information, see Operation.
Field |
Description |
id |
string ID of the operation. |
description |
string Description of the operation. 0-256 characters long. |
createdAt |
Creation timestamp. |
createdBy |
string ID of the user or service account who initiated the operation. |
modifiedAt |
The time when the Operation resource was last modified. |
done |
bool If the value is |
metadata |
Service-specific metadata associated with the operation. |
error |
The error result of the operation in case of failure or cancellation. Includes only one of the fields The operation result. |
response |
The normal response of the operation in case of success. Includes only one of the fields The operation result. |
CreateClusterMetadata
Field |
Description |
clusterId |
string ID of the ClickHouse cluster that is being created. |
Cluster
A ClickHouse Cluster resource. For more information, see the
Cluster section in the Developer's Guide.
Field |
Description |
id |
string ID of the ClickHouse cluster. |
folderId |
string ID of the folder that the ClickHouse cluster belongs to. |
createdAt |
Creation timestamp in RFC3339 |
name |
string Name of the ClickHouse cluster. |
description |
string Description of the ClickHouse cluster. 0-256 characters long. |
labels |
string Custom labels for the ClickHouse cluster as |
environment |
enum Environment Deployment environment of the ClickHouse cluster.
|
monitoring[] |
Description of monitoring systems relevant to the ClickHouse cluster. |
config |
Configuration of the ClickHouse cluster. |
networkId |
string ID of the network that the cluster belongs to. |
health |
enum Health Aggregated cluster health.
|
status |
enum Status Current state of the cluster.
|
serviceAccountId |
string ID of the service account used for access to Object Storage. |
maintenanceWindow |
Maintenance window for the cluster. |
plannedOperation |
Planned maintenance operation to be started for the cluster within the nearest |
securityGroupIds[] |
string User security groups |
deletionProtection |
bool Deletion Protection inhibits deletion of the cluster |
Monitoring
Monitoring system metadata.
Field |
Description |
name |
string Name of the monitoring system. |
description |
string Description of the monitoring system. |
link |
string Link to the monitoring system charts for the ClickHouse cluster. |
ClusterConfig
Field |
Description |
version |
string Version of the ClickHouse server software. |
clickhouse |
Configuration and resource allocation for ClickHouse hosts. |
zookeeper |
Configuration and resource allocation for ZooKeeper hosts. |
backupWindowStart |
Time to start the daily backup, in the UTC timezone. |
access |
Access policy for external services. |
cloudStorage |
|
sqlDatabaseManagement |
Whether database management through SQL commands is enabled. |
sqlUserManagement |
Whether user management through SQL commands is enabled. |
embeddedKeeper |
Whether cluster should use embedded Keeper instead of Zookeeper. |
backupRetainPeriodDays |
Retain period of automatically created backup in days |
Clickhouse
Field |
Description |
config |
Configuration settings of a ClickHouse server. |
resources |
Resources allocated to ClickHouse hosts. |
ClickhouseConfigSet
Field |
Description |
effectiveConfig |
Required field. Effective settings for a ClickHouse cluster (a combination of settings defined |
userConfig |
User-defined settings for a ClickHouse cluster. |
defaultConfig |
Default configuration for a ClickHouse cluster. |
ClickhouseConfig
ClickHouse configuration options. Detailed description for each set of options
is available in ClickHouse documentation
Any options not listed here are not supported.
Field |
Description |
logLevel |
enum LogLevel Logging level for the ClickHouse cluster. Possible values: TRACE, DEBUG, INFORMATION, WARNING, ERROR.
|
mergeTree |
Settings for the MergeTree engine. |
compression[] |
Compression settings for the ClickHouse cluster. |
dictionaries[] |
Configuration of external dictionaries to be used by the ClickHouse cluster. |
graphiteRollup[] |
Settings for thinning Graphite data. |
kafka |
|
kafkaTopics[] |
|
rabbitmq |
|
maxConnections |
Maximum number of inbound connections. |
maxConcurrentQueries |
Maximum number of simultaneously processed requests. |
keepAliveTimeout |
Number of milliseconds that ClickHouse waits for incoming requests before closing the connection. |
uncompressedCacheSize |
Cache size (in bytes) for uncompressed data used by MergeTree tables. |
markCacheSize |
Approximate size (in bytes) of the cache of "marks" used by MergeTree tables. |
maxTableSizeToDrop |
Maximum size of the table that can be deleted using a DROP query. |
maxPartitionSizeToDrop |
Maximum size of the partition that can be deleted using a DROP query. |
builtinDictionariesReloadInterval |
The setting is deprecated and has no effect. |
timezone |
string The server's time zone to be used in DateTime fields conversions. Specified as an IANA identifier. |
geobaseEnabled |
Enable or disable geobase. |
geobaseUri |
string Address of the archive with the user geobase in Object Storage. |
queryLogRetentionSize |
The maximum size that query_log can grow to before old data will be removed. If set to 0, automatic removal of |
queryLogRetentionTime |
The maximum time that query_log records will be retained before removal. If set to 0, automatic removal of |
queryThreadLogEnabled |
Whether query_thread_log system table is enabled. |
queryThreadLogRetentionSize |
The maximum size that query_thread_log can grow to before old data will be removed. If set to 0, automatic removal of |
queryThreadLogRetentionTime |
The maximum time that query_thread_log records will be retained before removal. If set to 0, automatic removal of |
partLogRetentionSize |
The maximum size that part_log can grow to before old data will be removed. If set to 0, automatic removal of |
partLogRetentionTime |
The maximum time that part_log records will be retained before removal. If set to 0, automatic removal of |
metricLogEnabled |
Whether metric_log system table is enabled. |
metricLogRetentionSize |
The maximum size that metric_log can grow to before old data will be removed. If set to 0, automatic removal of |
metricLogRetentionTime |
The maximum time that metric_log records will be retained before removal. If set to 0, automatic removal of |
traceLogEnabled |
Whether trace_log system table is enabled. |
traceLogRetentionSize |
The maximum size that trace_log can grow to before old data will be removed. If set to 0, automatic removal of |
traceLogRetentionTime |
The maximum time that trace_log records will be retained before removal. If set to 0, automatic removal of |
textLogEnabled |
Whether text_log system table is enabled. |
textLogRetentionSize |
The maximum size that text_log can grow to before old data will be removed. If set to 0, automatic removal of |
textLogRetentionTime |
The maximum time that text_log records will be retained before removal. If set to 0, automatic removal of |
textLogLevel |
enum LogLevel Logging level for text_log system table. Possible values: TRACE, DEBUG, INFORMATION, WARNING, ERROR.
|
opentelemetrySpanLogEnabled |
Enable or disable opentelemetry_span_log system table. Default value: false. |
opentelemetrySpanLogRetentionSize |
The maximum size that opentelemetry_span_log can grow to before old data will be removed. If set to 0 (default), |
opentelemetrySpanLogRetentionTime |
The maximum time that opentelemetry_span_log records will be retained before removal. If set to 0, |
queryViewsLogEnabled |
Enable or disable query_views_log system table. Default value: false. |
queryViewsLogRetentionSize |
The maximum size that query_views_log can grow to before old data will be removed. If set to 0 (default), |
queryViewsLogRetentionTime |
The maximum time that query_views_log records will be retained before removal. If set to 0, |
asynchronousMetricLogEnabled |
Enable or disable asynchronous_metric_log system table. Default value: false. |
asynchronousMetricLogRetentionSize |
The maximum size that asynchronous_metric_log can grow to before old data will be removed. If set to 0 (default), |
asynchronousMetricLogRetentionTime |
The maximum time that asynchronous_metric_log records will be retained before removal. If set to 0, |
sessionLogEnabled |
Enable or disable session_log system table. Default value: false. |
sessionLogRetentionSize |
The maximum size that session_log can grow to before old data will be removed. If set to 0 (default), |
sessionLogRetentionTime |
The maximum time that session_log records will be retained before removal. If set to 0, |
zookeeperLogEnabled |
Enable or disable zookeeper_log system table. Default value: false. |
zookeeperLogRetentionSize |
The maximum size that zookeeper_log can grow to before old data will be removed. If set to 0 (default), |
zookeeperLogRetentionTime |
The maximum time that zookeeper_log records will be retained before removal. If set to 0, |
asynchronousInsertLogEnabled |
Enable or disable asynchronous_insert_log system table. Default value: false. |
asynchronousInsertLogRetentionSize |
The maximum size that asynchronous_insert_log can grow to before old data will be removed. If set to 0 (default), |
asynchronousInsertLogRetentionTime |
The maximum time that asynchronous_insert_log records will be retained before removal. If set to 0, |
backgroundPoolSize |
|
backgroundMergesMutationsConcurrencyRatio |
Sets a ratio between the number of threads and the number of background merges and mutations that can be executed concurrently. For example, if the ratio equals to 2 and background_pool_size is set to 16 then ClickHouse can execute 32 background merges concurrently. This is possible, because background operations could be suspended and postponed. This is needed to give small merges more execution priority. You can only increase this ratio at runtime. To lower it you have to restart the server. The same as for background_pool_size setting background_merges_mutations_concurrency_ratio could be applied from the default profile for backward compatibility. |
backgroundSchedulePoolSize |
|
backgroundFetchesPoolSize |
Sets the number of threads performing background fetches for tables with ReplicatedMergeTree engines. Default value: 8. More info see in ClickHouse documentation |
backgroundMovePoolSize |
|
backgroundDistributedSchedulePoolSize |
|
backgroundBufferFlushSchedulePoolSize |
|
backgroundMessageBrokerSchedulePoolSize |
|
backgroundCommonPoolSize |
The maximum number of threads that will be used for performing a variety of operations (mostly garbage collection) for *MergeTree-engine tables in a background. |
defaultDatabase |
The default database. To get a list of cluster databases, see Yandex Managed ClickHouse documentation. |
totalMemoryProfilerStep |
Sets the memory size (in bytes) for a stack trace at every peak allocation step. Default value: 4194304. More info see in ClickHouse documentation |
totalMemoryTrackerSampleProbability |
|
queryMaskingRules[] |
Regexp-based rules, which will be applied to queries as well as all log messages before storing them in server logs, system.query_log, system.text_log, system.processes tables, and in logs sent to the client. That allows preventing sensitive data leakage from SQL queries (like names, emails, personal identifiers or credit card numbers) to logs. |
dictionariesLazyLoad |
Lazy loading of dictionaries. |
queryCache |
Query cache |
MergeTree
Options specific to the MergeTree table engine.
Field |
Description |
replicatedDeduplicationWindow |
Number of blocks of hashes to keep in ZooKeeper. |
replicatedDeduplicationWindowSeconds |
Period of time to keep blocks of hashes for. |
partsToDelayInsert |
If table contains at least that many active parts in single partition, artificially slow down insert into table. |
partsToThrowInsert |
If more than this number active parts in single partition, throw 'Too many parts ...' exception. |
inactivePartsToDelayInsert |
|
inactivePartsToThrowInsert |
|
maxReplicatedMergesInQueue |
How many tasks of merging and mutating parts are allowed simultaneously in ReplicatedMergeTree queue. |
numberOfFreeEntriesInPoolToLowerMaxSizeOfMerge |
If there is less than specified number of free entries in background pool (or replicated queue), start to lower |
maxBytesToMergeAtMinSpaceInPool |
Maximum in total size of parts to merge, when there are minimum free threads in background pool (or entries |
maxBytesToMergeAtMaxSpaceInPool |
|
minBytesForWidePart |
Minimum number of bytes in a data part that can be stored in Wide format. More info see in ClickHouse documentation |
minRowsForWidePart |
Minimum number of rows in a data part that can be stored in Wide format. More info see in ClickHouse documentation |
ttlOnlyDropParts |
Enables or disables complete dropping of data parts where all rows are expired in MergeTree tables. More info see in ClickHouse documentation |
allowRemoteFsZeroCopyReplication |
|
mergeWithTtlTimeout |
|
mergeWithRecompressionTtlTimeout |
|
maxPartsInTotal |
|
maxNumberOfMergesWithTtlInPool |
|
cleanupDelayPeriod |
|
numberOfFreeEntriesInPoolToExecuteMutation |
|
maxAvgPartSizeForTooManyParts |
The 'too many parts' check according to 'parts_to_delay_insert' and 'parts_to_throw_insert' will be active only if the average part size (in the relevant partition) is not larger than the specified threshold. If it is larger than the specified threshold, the INSERTs will be neither delayed or rejected. This allows to have hundreds of terabytes in a single table on a single server if the parts are successfully merged to larger parts. This does not affect the thresholds on inactive parts or total parts. |
minAgeToForceMergeSeconds |
Merge parts if every part in the range is older than the value of min_age_to_force_merge_seconds. |
minAgeToForceMergeOnPartitionOnly |
Whether min_age_to_force_merge_seconds should be applied only on the entire partition and not on subset. |
mergeSelectingSleepMs |
Sleep time for merge selecting when no part is selected. A lower setting triggers selecting tasks in background_schedule_pool frequently, which results in a large number of requests to ClickHouse Keeper in large-scale clusters. |
mergeMaxBlockSize |
The number of rows that are read from the merged parts into memory. |
checkSampleColumnIsCorrect |
Enables the check at table creation, that the data type of a column for sampling or sampling expression is correct. The data type must be one of unsigned integer types |
maxMergeSelectingSleepMs |
Maximum sleep time for merge selecting, a lower setting will trigger selecting tasks in background_schedule_pool frequently which result in large amount of requests to zookeeper in large-scale clusters. |
maxCleanupDelayPeriod |
Maximum period to clean old queue logs, blocks hashes and parts. |
Compression
Field |
Description |
method |
enum Method Compression method to use for the specified combination of
|
minPartSize |
int64 Minimum size of a part of a table. |
minPartSizeRatio |
double Minimum ratio of a part relative to the size of all the data in the table. |
level |
ExternalDictionary
Field |
Description |
name |
string Required field. Name of the external dictionary. |
structure |
Required field. Set of attributes for the external dictionary. |
layout |
Required field. Layout for storing the dictionary in memory. |
fixedLifetime |
int64 Fixed interval between dictionary updates. Includes only one of the fields Setting for the period of time between dictionary updates. |
lifetimeRange |
Range of intervals between dictionary updates for ClickHouse to choose from. Includes only one of the fields Setting for the period of time between dictionary updates. |
httpSource |
HTTP source for the dictionary. Includes only one of the fields Description of the source for the external dictionary. |
mysqlSource |
MySQL source for the dictionary. Includes only one of the fields Description of the source for the external dictionary. |
clickhouseSource |
ClickHouse source for the dictionary. Includes only one of the fields Description of the source for the external dictionary. |
mongodbSource |
MongoDB source for the dictionary. Includes only one of the fields Description of the source for the external dictionary. |
postgresqlSource |
PostgreSQL source for the dictionary. Includes only one of the fields Description of the source for the external dictionary. |
Structure
Field |
Description |
id |
Single numeric key column for the dictionary. |
key |
Composite key for the dictionary, containing of one or more key columns. |
rangeMin |
Field holding the beginning of the range for dictionaries with |
rangeMax |
Field holding the end of the range for dictionaries with |
attributes[] |
Description of the fields available for database queries. |
Id
Numeric key.
Field |
Description |
name |
string Required field. Name of the numeric key. |
Key
Complex key.
Field |
Description |
attributes[] |
Attributes of a complex key. |
Attribute
Field |
Description |
name |
string Required field. Name of the column. |
type |
string Required field. Type of the column. |
nullValue |
string Default value for an element without data (for example, an empty string). |
expression |
string Expression, describing the attribute, if applicable. |
hierarchical |
bool Indication of hierarchy support. |
injective |
bool Indication of injective mapping "id -> attribute". |
Layout
Layout determining how to store the dictionary in memory.
Field |
Description |
type |
enum Type Required field. Layout type for an external dictionary.
|
sizeInCells |
int64 Number of cells in the cache. Rounded up to a power of two. |
maxArraySize |
int64 Maximum dictionary key size. |
Range
Field |
Description |
min |
int64 Minimum dictionary lifetime. |
max |
int64 Maximum dictionary lifetime. |
HttpSource
Field |
Description |
url |
string Required field. URL of the source dictionary available over HTTP. |
format |
string Required field. The data format. Valid values are all formats supported by ClickHouse SQL dialect. |
headers[] |
HTTP headers. |
Header
Field |
Description |
name |
string Required field. |
value |
string Required field. |
MysqlSource
Field |
Description |
db |
string Required field. Name of the MySQL database to connect to. |
table |
string Required field. Name of the database table to use as a ClickHouse dictionary. |
port |
int64 Default port to use when connecting to a replica of the dictionary source. |
user |
string Name of the default user for replicas of the dictionary source. |
password |
string Password of the default user for replicas of the dictionary source. |
replicas[] |
List of MySQL replicas of the database used as dictionary source. |
where |
string Selection criteria for the data in the specified MySQL table. |
invalidateQuery |
string Query for checking the dictionary status, to pull only updated data. |
closeConnection |
Should the connection be closed after each request. |
shareConnection |
Should a connection be shared for some requests. |
Replica
Field |
Description |
host |
string Required field. MySQL host of the replica. |
priority |
int64 Required field. The priority of the replica that ClickHouse takes into account when connecting. |
port |
int64 Port to use when connecting to the replica. |
user |
string Name of the MySQL database user. |
password |
string Password of the MySQL database user. |
ClickhouseSource
Field |
Description |
db |
string Required field. Name of the ClickHouse database. |
table |
string Required field. Name of the table in the specified database to be used as the dictionary source. |
host |
string ClickHouse host of the specified database. |
port |
int64 Port to use when connecting to the host. |
user |
string Required field. Name of the ClickHouse database user. |
password |
string Password of the ClickHouse database user. |
where |
string Selection criteria for the data in the specified ClickHouse table. |
secure |
Use ssl for connection. |
MongodbSource
Field |
Description |
db |
string Required field. Name of the MongoDB database. |
collection |
string Required field. Name of the collection in the specified database to be used as the dictionary source. |
host |
string MongoDB host of the specified database. |
port |
int64 Port to use when connecting to the host. |
user |
string Required field. Name of the MongoDB database user. |
password |
string Password of the MongoDB database user. |
options |
string |
PostgresqlSource
Field |
Description |
db |
string Required field. Name of the PostrgreSQL database. |
table |
string Required field. Name of the table in the specified database to be used as the dictionary source. |
hosts[] |
string Name of the PostrgreSQL host |
port |
int64 Port to use when connecting to the host. |
user |
string Required field. Name of the PostrgreSQL database user. |
password |
string Password of the PostrgreSQL database user. |
invalidateQuery |
string Query for checking the dictionary status, to pull only updated data. |
sslMode |
enum SslMode Mode of SSL TCP/IP connection to the PostgreSQL host.
|
GraphiteRollup
Rollup settings for the GraphiteMergeTree table engine.
Field |
Description |
name |
string Required field. Name for the specified combination of settings for Graphite rollup. |
patterns[] |
Pattern to use for the rollup. |
pathColumnName |
string The name of the column storing the metric name (Graphite sensor). |
timeColumnName |
string The name of the column storing the time of measuring the metric. |
valueColumnName |
string The name of the column storing the value of the metric at the time set in time_column_name. |
versionColumnName |
string The name of the column storing the version of the metric. |
Pattern
Field |
Description |
regexp |
string Pattern for metric names. |
function |
string Required field. Name of the aggregating function to apply to data of the age specified in |
retention[] |
Age of data to use for thinning. |
Retention
Field |
Description |
age |
int64 Minimum age of the data in seconds. |
precision |
int64 Precision of determining the age of the data, in seconds. |
Kafka
Field |
Description |
securityProtocol |
enum SecurityProtocol
|
saslMechanism |
enum SaslMechanism
|
saslUsername |
string |
saslPassword |
string |
enableSslCertificateVerification |
|
maxPollIntervalMs |
|
sessionTimeoutMs |
|
debug |
enum Debug
|
autoOffsetReset |
enum AutoOffsetReset
|
KafkaTopic
Field |
Description |
name |
string Required field. |
settings |
Required field. |
Rabbitmq
Field |
Description |
username |
string RabbitMQ |
password |
string RabbitMQ |
vhost |
string RabbitMQ |
QueryMaskingRule
Field |
Description |
name |
string Name for the rule. |
regexp |
string Required field. RE2 compatible regular expression. |
replace |
string Substitution string for sensitive data. |
QueryCache
Field |
Description |
maxSizeInBytes |
The maximum cache size in bytes. |
maxEntries |
The maximum number of SELECT query results stored in the cache. |
maxEntrySizeInBytes |
The maximum size in bytes SELECT query results may have to be saved in the cache. |
maxEntrySizeInRows |
The maximum number of rows SELECT query results may have to be saved in the cache. |
Resources
Field |
Description |
resourcePresetId |
string ID of the preset for computational resources available to a host (CPU, memory etc.). |
diskSize |
int64 Volume of the storage available to a host, in bytes. |
diskTypeId |
string Type of the storage environment for the host.
|
Zookeeper
Field |
Description |
resources |
Resources allocated to ZooKeeper hosts. |
Access
Field |
Description |
dataLens |
bool Allow to export data from the cluster to DataLens. |
webSql |
bool Allow SQL queries to the cluster databases from the management console. See SQL queries in the management console for more details. |
metrika |
bool Allow to import data from Yandex Metrica and AppMetrica to the cluster. See AppMetrica documentation |
serverless |
bool Allow access to cluster for Serverless. |
dataTransfer |
bool Allow access for DataTransfer |
yandexQuery |
bool Allow access for Query |
CloudStorage
Field |
Description |
enabled |
bool Whether to use Object Storage for storing ClickHouse data. |
moveFactor |
|
dataCacheEnabled |
|
dataCacheMaxSize |
|
preferNotToMerge |
MaintenanceWindow
A maintenance window settings.
Field |
Description |
anytime |
Maintenance operation can be scheduled anytime. Includes only one of the fields The maintenance policy in effect. |
weeklyMaintenanceWindow |
Maintenance operation can be scheduled on a weekly basis. Includes only one of the fields The maintenance policy in effect. |
AnytimeMaintenanceWindow
Field |
Description |
Empty |
WeeklyMaintenanceWindow
Weelky maintenance window settings.
Field |
Description |
day |
enum WeekDay Day of the week (in
|
hour |
int64 Hour of the day in UTC (in |
MaintenanceOperation
A planned maintenance operation.
Field |
Description |
info |
string Information about this maintenance operation. |
delayedUntil |
Time until which this maintenance operation is delayed. |