S3 Test Run Report

DateJul 10, 2025 23:59
Duration2h 36m
Framework TestFlows 2.0.250110.1002922

Artifacts

Test artifacts can be found at https://altinity-build-artifacts.s3.amazonaws.com/index.html#0/1abef611c9a40dce30ed011ede491e532b2a3e72/regression/aarch64/with_analyzer/zookeeper/without_thread_fuzzer/s3/gcs/

Attributes

projectAltinity/ClickHouse
project.id159717931
packagehttps://s3.amazonaws.com/altinity-build-artifacts/25.3/1abef611c9a40dce30ed011ede491e532b2a3e72/package_aarch64/clickhouse-common-static_25.3.3.20183.altinityantalya_arm64.deb
version25.3.3.20183.altinityantalya
user.nameEnmk
repositoryhttps://github.com/Altinity/clickhouse-regression
commit.hash88c93f843cd48cd9defc6cec6b98d6b98f94adde
job.nameS3 (gcs)
job.retry1
job.urlhttps://github.com/Altinity/ClickHouse/actions/runs/16207084940
archaarch64
localTrue
clickhouse_versionNone
clickhouse_pathhttps://s3.amazonaws.com/altinity-build-artifacts/25.3/1abef611c9a40dce30ed011ede491e532b2a3e72/package_aarch64/clickhouse-common-static_25.3.3.20183.altinityantalya_arm64.deb
as_binaryFalse
base_osNone
keeper_pathNone
zookeeper_versionNone
use_keeperFalse
stressFalse
collect_service_logsTrue
thread_fuzzerFalse
with_analyzerTrue
reuse_envFalse
storages['gcs']
minio_uriSecret(name='minio_uri')
minio_root_userSecret(name='minio_root_user')
minio_root_passwordSecret(name='minio_root_password')
aws_s3_bucketSecret(name='aws_s3_bucket')
aws_s3_regionSecret(name='aws_s3_region')
aws_s3_key_idSecret(name='aws_s3_key_id')
aws_s3_access_keySecret(name='aws_s3_access_key')
gcs_uriSecret(name='gcs_uri')
gcs_key_idSecret(name='gcs_key_id')
gcs_key_secretSecret(name='gcs_key_secret')
azure_account_nameSecret(name='azure_account_name')
azure_storage_keySecret(name='azure_storage_key')
azure_containerSecret(name='azure_container')

Summary

88.8%OK
7.4%Known
1.1%Fail

Statistics

Units Skip OK Fail Error XFail XError Retried
Modules
1
1
Features
15
2
10
3
Scenarios
198
12
143
1
38
4
Checks
54
54
Examples
54
2
52
Steps
30315
30
30250
14
19
2

Fails

Test NameResultMessage
/s3Fail 2h 36m
AssertionError
Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 799, in 
    regression()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 171, in capture_cluster_args
    return func(self, cluster_args=cluster_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 348, in capture_s3_args
    return func(self, s3_args=s3_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 771, in regression
    Feature(test=gcs_regression)(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 711, in gcs_regression
    Feature(test=load("s3.tests.table_function", "gcs"))(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 869, in gcs
    outline()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 727, in outline
    scenario()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 76, in syntax_s3Cluster
    assert r == expected, error()
           ^^^^^^^^^^^^^
AssertionError: Oops! Assertion failed

The following assertion was not satisfied
  assert r == expected, error()

Assertion values
  assert r == expected, error()
         ^ is ''
  assert r == expected, error()
              ^ is '427'
  assert r == expected, error()
           ^ is = False
    @@ -0,0 +1 @@
    +427
  assert r == expected, error()
  ^ is False

Where
  File '/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py', line 76 in 'syntax_s3Cluster'

68\|              with And(f"I import the data from S3 into the second table {table2_name}"):
69\|                  insert_from_s3_function(filename=f"syntax_{cluster_name}.csv", cluster_name=cluster_name, table_name=table2_name)
70\|  
71\|              with Then(
72\|                  f"""I check that a simple SELECT * query on the second table
73\|                          {table2_name} returns matching data"""
74\|              ):
75\|                  r = self.context.cluster.node("clickhouse1").query(f"SELECT * FROM {table2_name} FORMAT CSV").output.strip()
76\|>                 assert r == expected, error()
77\|  
78\|  
79\|  @TestScenario
/s3/gcsFail 2h 36m
AssertionError
Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 799, in 
    regression()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 171, in capture_cluster_args
    return func(self, cluster_args=cluster_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 348, in capture_s3_args
    return func(self, s3_args=s3_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 771, in regression
    Feature(test=gcs_regression)(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 711, in gcs_regression
    Feature(test=load("s3.tests.table_function", "gcs"))(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 869, in gcs
    outline()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 727, in outline
    scenario()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 76, in syntax_s3Cluster
    assert r == expected, error()
           ^^^^^^^^^^^^^
AssertionError: Oops! Assertion failed

The following assertion was not satisfied
  assert r == expected, error()

Assertion values
  assert r == expected, error()
         ^ is ''
  assert r == expected, error()
              ^ is '427'
  assert r == expected, error()
           ^ is = False
    @@ -0,0 +1 @@
    +427
  assert r == expected, error()
  ^ is False

Where
  File '/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py', line 76 in 'syntax_s3Cluster'

68\|              with And(f"I import the data from S3 into the second table {table2_name}"):
69\|                  insert_from_s3_function(filename=f"syntax_{cluster_name}.csv", cluster_name=cluster_name, table_name=table2_name)
70\|  
71\|              with Then(
72\|                  f"""I check that a simple SELECT * query on the second table
73\|                          {table2_name} returns matching data"""
74\|              ):
75\|                  r = self.context.cluster.node("clickhouse1").query(f"SELECT * FROM {table2_name} FORMAT CSV").output.strip()
76\|>                 assert r == expected, error()
77\|  
78\|  
79\|  @TestScenario
/s3/gcs/part 1Fail 1h 11m
AssertionError
Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 799, in 
    regression()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 171, in capture_cluster_args
    return func(self, cluster_args=cluster_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 348, in capture_s3_args
    return func(self, s3_args=s3_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 771, in regression
    Feature(test=gcs_regression)(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 711, in gcs_regression
    Feature(test=load("s3.tests.table_function", "gcs"))(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 869, in gcs
    outline()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 727, in outline
    scenario()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 76, in syntax_s3Cluster
    assert r == expected, error()
           ^^^^^^^^^^^^^
AssertionError: Oops! Assertion failed

The following assertion was not satisfied
  assert r == expected, error()

Assertion values
  assert r == expected, error()
         ^ is ''
  assert r == expected, error()
              ^ is '427'
  assert r == expected, error()
           ^ is = False
    @@ -0,0 +1 @@
    +427
  assert r == expected, error()
  ^ is False

Where
  File '/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py', line 76 in 'syntax_s3Cluster'

68\|              with And(f"I import the data from S3 into the second table {table2_name}"):
69\|                  insert_from_s3_function(filename=f"syntax_{cluster_name}.csv", cluster_name=cluster_name, table_name=table2_name)
70\|  
71\|              with Then(
72\|                  f"""I check that a simple SELECT * query on the second table
73\|                          {table2_name} returns matching data"""
74\|              ):
75\|                  r = self.context.cluster.node("clickhouse1").query(f"SELECT * FROM {table2_name} FORMAT CSV").output.strip()
76\|>                 assert r == expected, error()
77\|  
78\|  
79\|  @TestScenario
/s3/gcs/part 1/table functionFail 8m 13s
AssertionError
Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 799, in 
    regression()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 171, in capture_cluster_args
    return func(self, cluster_args=cluster_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 348, in capture_s3_args
    return func(self, s3_args=s3_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 771, in regression
    Feature(test=gcs_regression)(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 711, in gcs_regression
    Feature(test=load("s3.tests.table_function", "gcs"))(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 869, in gcs
    outline()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 727, in outline
    scenario()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 76, in syntax_s3Cluster
    assert r == expected, error()
           ^^^^^^^^^^^^^
AssertionError: Oops! Assertion failed

The following assertion was not satisfied
  assert r == expected, error()

Assertion values
  assert r == expected, error()
         ^ is ''
  assert r == expected, error()
              ^ is '427'
  assert r == expected, error()
           ^ is = False
    @@ -0,0 +1 @@
    +427
  assert r == expected, error()
  ^ is False

Where
  File '/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py', line 76 in 'syntax_s3Cluster'

68\|              with And(f"I import the data from S3 into the second table {table2_name}"):
69\|                  insert_from_s3_function(filename=f"syntax_{cluster_name}.csv", cluster_name=cluster_name, table_name=table2_name)
70\|  
71\|              with Then(
72\|                  f"""I check that a simple SELECT * query on the second table
73\|                          {table2_name} returns matching data"""
74\|              ):
75\|                  r = self.context.cluster.node("clickhouse1").query(f"SELECT * FROM {table2_name} FORMAT CSV").output.strip()
76\|>                 assert r == expected, error()
77\|  
78\|  
79\|  @TestScenario
/s3/gcs/part 1/table function/syntax s3ClusterFail 9s 644ms
AssertionError
Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 799, in 
    regression()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 171, in capture_cluster_args
    return func(self, cluster_args=cluster_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 348, in capture_s3_args
    return func(self, s3_args=s3_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 771, in regression
    Feature(test=gcs_regression)(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 711, in gcs_regression
    Feature(test=load("s3.tests.table_function", "gcs"))(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 869, in gcs
    outline()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 727, in outline
    scenario()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py", line 76, in syntax_s3Cluster
    assert r == expected, error()
           ^^^^^^^^^^^^^
AssertionError: Oops! Assertion failed

The following assertion was not satisfied
  assert r == expected, error()

Assertion values
  assert r == expected, error()
         ^ is ''
  assert r == expected, error()
              ^ is '427'
  assert r == expected, error()
           ^ is = False
    @@ -0,0 +1 @@
    +427
  assert r == expected, error()
  ^ is False

Where
  File '/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function.py', line 76 in 'syntax_s3Cluster'

68\|              with And(f"I import the data from S3 into the second table {table2_name}"):
69\|                  insert_from_s3_function(filename=f"syntax_{cluster_name}.csv", cluster_name=cluster_name, table_name=table2_name)
70\|  
71\|              with Then(
72\|                  f"""I check that a simple SELECT * query on the second table
73\|                          {table2_name} returns matching data"""
74\|              ):
75\|                  r = self.context.cluster.node("clickhouse1").query(f"SELECT * FROM {table2_name} FORMAT CSV").output.strip()
76\|>                 assert r == expected, error()
77\|  
78\|  
79\|  @TestScenario

Known Fails

Test NameResultMessage
/s3/gcs/part 1/invalid table function/invalid pathXError 30s 48ms
https://github.com/ClickHouse/ClickHouse/issues/59084
ExpectTimeoutError
Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/venv/lib/python3.12/site-packages/testflows/uexpect/uexpect.py", line 223, in read
    d = self.queue.get(timeout=timeleft)
        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/lib/python3.12/queue.py", line 179, in get
    raise Empty
_queue.Empty

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/venv/lib/python3.12/site-packages/testflows/uexpect/uexpect.py", line 187, in expect
    data = self.read(timeout=min(timeleft, 0.1), raise_exception=True)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/venv/lib/python3.12/site-packages/testflows/uexpect/uexpect.py", line 235, in read
    raise TimeoutError(timeout)
testflows.uexpect.uexpect.TimeoutError: Timeout 0.038s

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 799, in 
    regression()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 171, in capture_cluster_args
    return func(self, cluster_args=cluster_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 348, in capture_s3_args
    return func(self, s3_args=s3_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 771, in regression
    Feature(test=gcs_regression)(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 714, in gcs_regression
    Feature(test=load("s3.tests.table_function_invalid", "gcs"))(uri=uri)
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function_invalid.py", line 411, in gcs
    outline(uri=uri)
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function_invalid.py", line 395, in outline
    scenario()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function_invalid.py", line 114, in invalid_path
    insert_to_s3_function_invalid(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/table_function_invalid.py", line 34, in insert_to_s3_function_invalid
    node.query(query, message=message, exitcode=exitcode, timeout=timeout)
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/cluster.py", line 1140, in query
    r = self.cluster.bash(self.name)(command, *args, **kwargs)
        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
testflows.uexpect.uexpect.ExpectTimeoutError: Timeout 30.000s for '(bash# )\|(\n)'
/s3/gcs/part 1/disk/cacheXFail 4ms
Under development for 22.8 and newer.
None
/s3/gcs/part 1/disk/cache defaultXFail 1ms
Under development for 22.8 and newer.
None
/s3/gcs/part 1/disk/cache pathXFail 1ms
Under development for 22.8 and newer.
None
/s3/gcs/part 1/disk/low cardinality offsetXFail 51s 986ms
https://github.com/ClickHouse/ClickHouse/pull/44875
AssertionError
Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 799, in 
    regression()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 171, in capture_cluster_args
    return func(self, cluster_args=cluster_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 348, in capture_s3_args
    return func(self, s3_args=s3_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 771, in regression
    Feature(test=gcs_regression)(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 715, in gcs_regression
    Feature(test=load("s3.tests.disk", "gcs"))(uri=uri, bucket_prefix=bucket_prefix)
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/disk.py", line 2560, in gcs
    disk_tests(uri=uri, bucket_prefix=bucket_prefix)
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/disk.py", line 2538, in disk_tests
    scenario()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/disk.py", line 2500, in low_cardinality_offset
    assert output == "23999\n", error()
           ^^^^^^^^^^^^^^^^^^^
AssertionError: Oops! Assertion failed

The following assertion was not satisfied
  assert output == "23999\n", error()

Assertion values
  assert output == "23999\n", error()
         ^ is '23999'
  assert output == "23999\n", error()
                ^ is = False

  assert output == "23999\n", error()
  ^ is False

Where
  File '/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/disk.py', line 2500 in 'low_cardinality_offset'

2492\|                      "1",
2493\|                  ),
2494\|                  (
2495\|                      "merge_tree_min_bytes_for_concurrent_read_for_remote_filesystem",
2496\|                      "1",
2497\|                  ),
2498\|              ],
2499\|          ).output
2500\|>         assert output == "23999\n", error()
2501\|  
2502\|  
2503\|  @TestFeature
/s3/gcs/part 1/disk/no restartXFail 1ms
https://github.com/ClickHouse/ClickHouse/issues/58924
None
/s3/gcs/part 1/invalid disk/cache path conflictXFail 1ms
Under development for 22.8 and newer.
None
/s3/gcs/part 2/combinatoric table/engine=VersionedCollapsingMergeTree,replicated=True,n_cols=2000,n_tables=3,part_type=wideXFail 4m 26s
Needs investigation, rows not appearing
AssertionError
Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 799, in 
    regression()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 171, in capture_cluster_args
    return func(self, cluster_args=cluster_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 348, in capture_s3_args
    return func(self, s3_args=s3_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 771, in regression
    Feature(test=gcs_regression)(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 721, in gcs_regression
    Feature(test=load("s3.tests.combinatoric_table", "feature"))(uri=uri)
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/combinatoric_table.py", line 182, in feature
    Scenario(title, test=check_table_combination)(**table_config)
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/combinatoric_table.py", line 149, in check_table_combination
    retry(assert_row_count, timeout=60, delay=5)(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/common.py", line 1742, in assert_row_count
    assert rows == actual_count, error()
           ^^^^^^^^^^^^^^^^^^^^
AssertionError: Oops! Assertion failed

The following assertion was not satisfied
  assert rows == actual_count, error()

Assertion values
  assert rows == actual_count, error()
         ^ is 1500
  assert rows == actual_count, error()
                 ^ is 0
  assert rows == actual_count, error()
              ^ is = False
  assert rows == actual_count, error()
  ^ is False

Where
  File '/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/common.py', line 1742 in 'assert_row_count'

1734\|  
1735\|  @TestStep(Then)
1736\|  def assert_row_count(self, node, table_name: str, rows: int = 1000000):
1737\|      """Assert that the number of rows in a table is as expected."""
1738\|      if node is None:
1739\|          node = current().context.node
1740\|  
1741\|      actual_count = get_row_count(node=node, table_name=table_name)
1742\|>     assert rows == actual_count, error()
1743\|  
1744\|  
1745\|  @TestStep(Then)
/s3/gcs/part 2/combinatoric table/engine=AggregatingMergeTree,replicated=True,n_cols=2000,n_tables=3,part_type=compactXError 8m 2s
Times out, needs investigation
ExpectTimeoutError
Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/venv/lib/python3.12/site-packages/testflows/uexpect/uexpect.py", line 223, in read
    d = self.queue.get(timeout=timeleft)
        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/lib/python3.12/queue.py", line 179, in get
    raise Empty
_queue.Empty

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/venv/lib/python3.12/site-packages/testflows/uexpect/uexpect.py", line 187, in expect
    data = self.read(timeout=min(timeleft, 0.1), raise_exception=True)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/venv/lib/python3.12/site-packages/testflows/uexpect/uexpect.py", line 235, in read
    raise TimeoutError(timeout)
testflows.uexpect.uexpect.TimeoutError: Timeout 0.087s

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 799, in 
    regression()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 171, in capture_cluster_args
    return func(self, cluster_args=cluster_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 348, in capture_s3_args
    return func(self, s3_args=s3_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 771, in regression
    Feature(test=gcs_regression)(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 721, in gcs_regression
    Feature(test=load("s3.tests.combinatoric_table", "feature"))(uri=uri)
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/combinatoric_table.py", line 182, in feature
    Scenario(title, test=check_table_combination)(**table_config)
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/combinatoric_table.py", line 117, in check_table_combination
    table = create_test_table(
            ^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/combinatoric_table.py", line 81, in create_test_table
    yield create_table(
          ^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/tables.py", line 482, in create_table
    node.query(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/cluster.py", line 1116, in query
    r = self.cluster.bash(None)(command, *args, **kwargs)
        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
testflows.uexpect.uexpect.ExpectTimeoutError: Timeout 300.000s for '(bash# )\|(\n)'
/s3/gcs/part 2/combinatoric table/engine=SummingMergeTree,replicated=True,n_cols=2000,n_tables=3,part_type=compactXError 8m 2s
Times out, needs investigation
ExpectTimeoutError
Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/venv/lib/python3.12/site-packages/testflows/uexpect/uexpect.py", line 223, in read
    d = self.queue.get(timeout=timeleft)
        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/lib/python3.12/queue.py", line 179, in get
    raise Empty
_queue.Empty

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/venv/lib/python3.12/site-packages/testflows/uexpect/uexpect.py", line 187, in expect
    data = self.read(timeout=min(timeleft, 0.1), raise_exception=True)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/venv/lib/python3.12/site-packages/testflows/uexpect/uexpect.py", line 235, in read
    raise TimeoutError(timeout)
testflows.uexpect.uexpect.TimeoutError: Timeout 0.086s

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 799, in 
    regression()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 171, in capture_cluster_args
    return func(self, cluster_args=cluster_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 348, in capture_s3_args
    return func(self, s3_args=s3_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 771, in regression
    Feature(test=gcs_regression)(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 721, in gcs_regression
    Feature(test=load("s3.tests.combinatoric_table", "feature"))(uri=uri)
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/combinatoric_table.py", line 182, in feature
    Scenario(title, test=check_table_combination)(**table_config)
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/combinatoric_table.py", line 117, in check_table_combination
    table = create_test_table(
            ^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/combinatoric_table.py", line 81, in create_test_table
    yield create_table(
          ^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/tables.py", line 482, in create_table
    node.query(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/cluster.py", line 1116, in query
    r = self.cluster.bash(None)(command, *args, **kwargs)
        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
testflows.uexpect.uexpect.ExpectTimeoutError: Timeout 300.000s for '(bash# )\|(\n)'
/s3/gcs/part 2/combinatoric table/engine=MergeTree,replicated=True,n_cols=10,n_tables=3,part_type=unspecifiedXError 5m 15s
Times out, needs investigation
ExpectTimeoutError
Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/venv/lib/python3.12/site-packages/testflows/uexpect/uexpect.py", line 223, in read
    d = self.queue.get(timeout=timeleft)
        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/usr/lib/python3.12/queue.py", line 179, in get
    raise Empty
_queue.Empty

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/venv/lib/python3.12/site-packages/testflows/uexpect/uexpect.py", line 187, in expect
    data = self.read(timeout=min(timeleft, 0.1), raise_exception=True)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/venv/lib/python3.12/site-packages/testflows/uexpect/uexpect.py", line 235, in read
    raise TimeoutError(timeout)
testflows.uexpect.uexpect.TimeoutError: Timeout 0.098s

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 799, in 
    regression()
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 171, in capture_cluster_args
    return func(self, cluster_args=cluster_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/argparser.py", line 348, in capture_s3_args
    return func(self, s3_args=s3_args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 771, in regression
    Feature(test=gcs_regression)(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/regression.py", line 721, in gcs_regression
    Feature(test=load("s3.tests.combinatoric_table", "feature"))(uri=uri)
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/combinatoric_table.py", line 182, in feature
    Scenario(title, test=check_table_combination)(**table_config)
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/combinatoric_table.py", line 117, in check_table_combination
    table = create_test_table(
            ^^^^^^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../s3/tests/combinatoric_table.py", line 81, in create_test_table
    yield create_table(
          ^^^^^^^^^^^^^
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/tables.py", line 482, in create_table
    node.query(
  File "/home/ubuntu/_work/ClickHouse/ClickHouse/s3/../helpers/cluster.py", line 1140, in query
    r = self.cluster.bash(self.name)(command, *args, **kwargs)
        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
testflows.uexpect.uexpect.ExpectTimeoutError: Timeout 300.000s for '(bash# )\|(\n)'

Results

Test Name Result Duration
/s3 Fail 2h 36m
/s3/gcs Fail 2h 36m
/s3/gcs/part 1 Fail 1h 11m
/s3/gcs/part 1/sanity OK 1m 13s
/s3/gcs/part 1/sanity/sanity OK 16s 916ms
/s3/gcs/part 1/table function Fail 8m 13s
/s3/gcs/part 1/table function/auto OK 1m 35s
/s3/gcs/part 1/table function/compression OK 1m 34s
/s3/gcs/part 1/table function/credentials OK 920ms
/s3/gcs/part 1/table function/credentials s3Cluster OK 9s 623ms
/s3/gcs/part 1/table function/data format OK 1m 55s
/s3/gcs/part 1/table function/measure file size Skip 1ms
/s3/gcs/part 1/table function/measure file size s3Cluster Skip 912us
/s3/gcs/part 1/table function/multipart OK 8s 42ms
/s3/gcs/part 1/table function/multiple columns OK 1s 371ms
/s3/gcs/part 1/table function/partition OK 1s 309ms
/s3/gcs/part 1/table function/partition s3Cluster OK 8s 493ms
/s3/gcs/part 1/table function/remote host filter OK 51s 291ms
/s3/gcs/part 1/table function/syntax OK 1s 63ms
/s3/gcs/part 1/table function/syntax s3Cluster Fail 9s 644ms
/s3/gcs/part 1/table function/wildcard OK 1m 34s
/s3/gcs/part 1/invalid table function OK 35s 83ms
/s3/gcs/part 1/invalid table function/empty path OK 276ms
/s3/gcs/part 1/invalid table function/empty structure OK 275ms
/s3/gcs/part 1/invalid table function/invalid bucket OK 279ms
/s3/gcs/part 1/invalid table function/invalid compression OK 279ms
/s3/gcs/part 1/invalid table function/invalid credentials OK 490ms
/s3/gcs/part 1/invalid table function/invalid format OK 544ms
/s3/gcs/part 1/invalid table function/invalid path XError 30s 48ms
/s3/gcs/part 1/invalid table function/invalid region Skip 7ms
/s3/gcs/part 1/invalid table function/invalid structure OK 1s 765ms
/s3/gcs/part 1/invalid table function/invalid wildcard OK 1s 91ms
/s3/gcs/part 1/disk OK 41m 9s
/s3/gcs/part 1/disk/access OK 52s 249ms
/s3/gcs/part 1/disk/access skip check OK 55s 442ms
/s3/gcs/part 1/disk/add storage OK 1m 44s
/s3/gcs/part 1/disk/alter move OK 1m 2s
/s3/gcs/part 1/disk/alter on cluster modify ttl OK 1m 56s
/s3/gcs/part 1/disk/cache XFail 4ms
/s3/gcs/part 1/disk/cache default XFail 1ms
/s3/gcs/part 1/disk/cache path XFail 1ms
/s3/gcs/part 1/disk/compact parts OK 52s 288ms
/s3/gcs/part 1/disk/config over restart OK 1m 20s
/s3/gcs/part 1/disk/default move factor OK 1m 28s
/s3/gcs/part 1/disk/delete OK 3m 30s
/s3/gcs/part 1/disk/download appropriate disk OK 1m 28s
/s3/gcs/part 1/disk/drop sync OK 1m 0s
/s3/gcs/part 1/disk/environment credentials Skip 2ms
/s3/gcs/part 1/disk/exports OK 54s 273ms
/s3/gcs/part 1/disk/generic url Skip 2ms
/s3/gcs/part 1/disk/imports OK 55s 789ms
/s3/gcs/part 1/disk/log OK 3m 14s
/s3/gcs/part 1/disk/low cardinality offset XFail 51s 986ms
/s3/gcs/part 1/disk/max single part upload size syntax OK 54s 205ms
/s3/gcs/part 1/disk/mergetree OK 4m 10s
/s3/gcs/part 1/disk/mergetree collapsing OK 1m 2s
/s3/gcs/part 1/disk/mergetree versionedcollapsing OK 1m 2s
/s3/gcs/part 1/disk/metadata OK 1m 2s
/s3/gcs/part 1/disk/min bytes for seek syntax OK 54s 732ms
/s3/gcs/part 1/disk/multiple storage OK 1m 3s
/s3/gcs/part 1/disk/multiple storage query OK 1m 5s
/s3/gcs/part 1/disk/no restart XFail 1ms
/s3/gcs/part 1/disk/perform ttl move on insert OK 1m 45s
/s3/gcs/part 1/disk/perform ttl move on insert default OK 1m 2s
/s3/gcs/part 1/disk/performance ttl move OK 1m 16s
/s3/gcs/part 1/disk/remote host filter OK 1m 44s
/s3/gcs/part 1/disk/specific url Skip 985us
/s3/gcs/part 1/disk/syntax OK 1m 2s
/s3/gcs/part 1/disk/wide parts OK 53s 968ms
/s3/gcs/part 1/invalid disk OK 3m 47s
/s3/gcs/part 1/invalid disk/access default OK 10s 757ms
/s3/gcs/part 1/invalid disk/access failed OK 10s 777ms
/s3/gcs/part 1/invalid disk/access failed skip check OK 52s 186ms
/s3/gcs/part 1/invalid disk/cache path conflict XFail 1ms
/s3/gcs/part 1/invalid disk/empty endpoint OK 9s 832ms
/s3/gcs/part 1/invalid disk/invalid endpoint OK 2m 3s
/s3/gcs/part 1/invalid disk/invalid type OK 20s 497ms
/s3/gcs/part 1/alter OK 16m 57s
/s3/gcs/part 1/alter/normal OK 4m 15s
/s3/gcs/part 1/alter/normal/attach from OK 13s 753ms
/s3/gcs/part 1/alter/normal/columns OK 19s 140ms
/s3/gcs/part 1/alter/normal/detach OK 27s 74ms
/s3/gcs/part 1/alter/normal/drop OK 53s 201ms
/s3/gcs/part 1/alter/normal/fetch OK 33s 709ms
/s3/gcs/part 1/alter/normal/freeze OK 23s 100ms
/s3/gcs/part 1/alter/normal/index OK 11s 5ms
/s3/gcs/part 1/alter/normal/move to table OK 19s 392ms
/s3/gcs/part 1/alter/normal/order by OK 7s 668ms
/s3/gcs/part 1/alter/normal/projection OK 12s 889ms
/s3/gcs/part 1/alter/normal/replace OK 18s 50ms
/s3/gcs/part 1/alter/normal/sample by OK 7s 657ms
/s3/gcs/part 1/alter/normal/update delete OK 9s 24ms
/s3/gcs/part 1/alter/encrypted OK 5m 36s
/s3/gcs/part 1/alter/encrypted/attach from OK 16s 631ms
/s3/gcs/part 1/alter/encrypted/columns OK 20s 162ms
/s3/gcs/part 1/alter/encrypted/detach OK 37s 410ms
/s3/gcs/part 1/alter/encrypted/drop OK 1m 13s
/s3/gcs/part 1/alter/encrypted/fetch OK 39s 332ms
/s3/gcs/part 1/alter/encrypted/freeze OK 31s 709ms
/s3/gcs/part 1/alter/encrypted/index OK 16s 296ms
/s3/gcs/part 1/alter/encrypted/move to table OK 30s 830ms
/s3/gcs/part 1/alter/encrypted/order by OK 8s 791ms
/s3/gcs/part 1/alter/encrypted/projection OK 18s 51ms
/s3/gcs/part 1/alter/encrypted/replace OK 20s 478ms
/s3/gcs/part 1/alter/encrypted/sample by OK 8s 388ms
/s3/gcs/part 1/alter/encrypted/update delete OK 14s 648ms
/s3/gcs/part 1/alter/zero copy OK 2m 13s
/s3/gcs/part 1/alter/zero copy/attach from OK 13s 440ms
/s3/gcs/part 1/alter/zero copy/columns OK 13s 535ms
/s3/gcs/part 1/alter/zero copy/detach Skip 1ms
/s3/gcs/part 1/alter/zero copy/drop Skip 1ms
/s3/gcs/part 1/alter/zero copy/fetch Skip 1ms
/s3/gcs/part 1/alter/zero copy/freeze OK 23s 308ms
/s3/gcs/part 1/alter/zero copy/index OK 10s 765ms
/s3/gcs/part 1/alter/zero copy/move to table OK 18s 18ms
/s3/gcs/part 1/alter/zero copy/order by OK 7s 597ms
/s3/gcs/part 1/alter/zero copy/projection OK 12s 547ms
/s3/gcs/part 1/alter/zero copy/replace OK 18s 57ms
/s3/gcs/part 1/alter/zero copy/sample by OK 7s 521ms
/s3/gcs/part 1/alter/zero copy/update delete OK 8s 810ms
/s3/gcs/part 1/alter/zero copy encrypted OK 3m 4s
/s3/gcs/part 1/alter/zero copy encrypted/attach from OK 17s 894ms
/s3/gcs/part 1/alter/zero copy encrypted/columns OK 19s 618ms
/s3/gcs/part 1/alter/zero copy encrypted/detach Skip 1ms
/s3/gcs/part 1/alter/zero copy encrypted/drop Skip 1ms
/s3/gcs/part 1/alter/zero copy encrypted/fetch Skip 906us
/s3/gcs/part 1/alter/zero copy encrypted/freeze OK 29s 575ms
/s3/gcs/part 1/alter/zero copy encrypted/index OK 16s 124ms
/s3/gcs/part 1/alter/zero copy encrypted/move to table OK 29s 787ms
/s3/gcs/part 1/alter/zero copy encrypted/order by OK 8s 550ms
/s3/gcs/part 1/alter/zero copy encrypted/projection OK 18s 263ms
/s3/gcs/part 1/alter/zero copy encrypted/replace OK 20s 238ms
/s3/gcs/part 1/alter/zero copy encrypted/sample by OK 8s 467ms
/s3/gcs/part 1/alter/zero copy encrypted/update delete OK 15s 681ms
/s3/gcs/part 2 OK 1h 21m
/s3/gcs/part 2/combinatoric table OK 39m 30s
/s3/gcs/part 2/combinatoric table/engine=MergeTree,replicated=True,n_cols=2000,n_tables=3,part_type=compact OK 44s 269ms
/s3/gcs/part 2/combinatoric table/engine=MergeTree,replicated=False,n_cols=500,n_tables=1,part_type=wide OK 5m 8s
/s3/gcs/part 2/combinatoric table/engine=ReplacingMergeTree,replicated=True,n_cols=500,n_tables=3,part_type=unspecified OK 44s 546ms
/s3/gcs/part 2/combinatoric table/engine=ReplacingMergeTree,replicated=False,n_cols=2000,n_tables=1,part_type=compact OK 8s 57ms
/s3/gcs/part 2/combinatoric table/engine=CollapsingMergeTree,replicated=True,n_cols=10,n_tables=1,part_type=wide OK 30s 909ms
/s3/gcs/part 2/combinatoric table/engine=CollapsingMergeTree,replicated=False,n_cols=2000,n_tables=3,part_type=unspecified OK 22s 439ms
/s3/gcs/part 2/combinatoric table/engine=VersionedCollapsingMergeTree,replicated=True,n_cols=2000,n_tables=3,part_type=wide XFail 4m 26s
/s3/gcs/part 2/combinatoric table/engine=VersionedCollapsingMergeTree,replicated=False,n_cols=10,n_tables=3,part_type=compact OK 11s 344ms
/s3/gcs/part 2/combinatoric table/engine=AggregatingMergeTree,replicated=True,n_cols=2000,n_tables=3,part_type=compact XError 8m 2s
/s3/gcs/part 2/combinatoric table/engine=AggregatingMergeTree,replicated=False,n_cols=500,n_tables=1,part_type=unspecified OK 6s 877ms
/s3/gcs/part 2/combinatoric table/engine=SummingMergeTree,replicated=True,n_cols=2000,n_tables=3,part_type=compact XError 8m 2s
/s3/gcs/part 2/combinatoric table/engine=SummingMergeTree,replicated=False,n_cols=500,n_tables=1,part_type=compact OK 6s 968ms
/s3/gcs/part 2/combinatoric table/engine=MergeTree,replicated=True,n_cols=10,n_tables=3,part_type=unspecified XError 5m 15s
/s3/gcs/part 2/combinatoric table/engine=ReplacingMergeTree,replicated=True,n_cols=10,n_tables=3,part_type=wide OK 1m 12s
/s3/gcs/part 2/combinatoric table/engine=CollapsingMergeTree,replicated=True,n_cols=500,n_tables=3,part_type=compact OK 42s 84ms
/s3/gcs/part 2/combinatoric table/engine=VersionedCollapsingMergeTree,replicated=True,n_cols=500,n_tables=1,part_type=unspecified OK 14s 445ms
/s3/gcs/part 2/combinatoric table/engine=AggregatingMergeTree,replicated=True,n_cols=10,n_tables=3,part_type=wide OK 1m 12s
/s3/gcs/part 2/combinatoric table/engine=SummingMergeTree,replicated=True,n_cols=10,n_tables=3,part_type=wide OK 1m 14s
/s3/gcs/part 2/combinatoric table/engine=SummingMergeTree,replicated=True,n_cols=10,n_tables=1,part_type=unspecified OK 11s 424ms
/s3/gcs/part 2/zero copy replication Skip 1ms
/s3/gcs/part 2/backup OK 3m 30s
/s3/gcs/part 2/backup/local and s3 disk OK 1m 8s
/s3/gcs/part 2/backup/local and s3 volumes OK 1m 6s
/s3/gcs/part 2/backup/s3 disk OK 1m 14s
/s3/gcs/part 2/orphans Skip 1ms
/s3/gcs/part 2/settings OK 29m 26s
/s3/gcs/part 2/settings/setting combinations OK 28m 34s
/s3/gcs/part 2/table function performance OK 8m 48s
/s3/gcs/part 2/table function performance/setup OK 833ms
/s3/gcs/part 2/table function performance/wildcard OK 8m 47s

Generated by TestFlows Open-Source Test Framework v2.0.250110.1002922