ClickHouse Tests in TestFlows
This directory contains integration tests written using TestFlows that involves several ClickHouse instances, custom configs, ZooKeeper, etc.
- 1 Supported environment
- 2 Prerequisites
- 3 Running CI/CD
- 4 Running CI/CD with CI/CD trigger
- 5 Running tests locally
- 5.1 Output Verbosity
- 5.2 Running Only Selected Tests
- 5.3 How To Debug Why Test Failed
- 5.4 Running S3 Suites
- 5.4.1 Minio
- 5.4.2 AWS S3 Storage
- 5.4.3 GCS (Google Cloud Storage)
- 5.5 Running Tiered Storage Suites
- 5.5.1 Normal
- 5.5.2 Running on Minio
- 5.5.3 Running on AWS S3 Storage
- 5.5.4 Running on GCS Storage
- 5.5.5 Pausing In Tests
- 6 Running GitHub Actions
- Ubuntu 22.04
- Python 3 >= 3.8
- TestFlows >= 1.9.51
- Docker Compose == 1.29.2
To install all necessary Python modules (including TestFlows and Docker Compose), execute the following command:
pip3 install -r pip_requirements.txtSpecify suite variable to select running only specific suites
| Variable | ||
|---|---|---|
suite |
window_functions |
Window Functions suite |
suite |
aes_encryption |
AES Encryption Functions suite |
suite |
clickhouse_keeper |
ClickHouse Keeper suite |
suite |
datetime64_extended_range |
Extended DateTime64 suite |
suite |
disk_level_encryption |
Disk Level Encryption |
suite |
example |
Example suite |
suite |
extended_precision_data_types |
Extended Precision Data Types suite |
suite |
kafka |
Kafka suite |
suite |
kerberos |
Kerberos suite |
suite |
ldap |
LDAP suite |
suite |
lightweight_delete |
Lightweight Delete suite |
suite |
map_type |
Map Data Type suite |
suite |
parquet |
Parquet Data Type suite |
suite |
parquet_minio |
Parquet MinIO suite |
suite |
parquet_s3 |
Parquet AWS S3 suite |
suite |
part_moves_between_shards |
Part Moves Between Shards suite |
suite |
s3 |
All S3 (MinIO, AWS, GCS) suites |
suite |
s3_aws |
S3 AWS suite |
suite |
s3_gcs |
S3 GCS suite |
suite |
ssl_server |
SSL Server suite |
suite |
tiered_storage |
All Tiered Storage (Local Disk, MinIO, AWS, GCS) suites |
suite |
tiered_storage_aws |
Tiered Storage AWS suite |
suite |
tiered_storage_gcs |
Tiered Storage GCS suite |
suite |
window_functions |
Window Functions suite |
suite |
benchmark |
S3 Benchmak suite |
When running the CI/CD pipeline, provide the following variables:
Example values using altinity/clickhouse-server:21.8.15.15.altinitystable
| Variables | ||
|---|---|---|
Variable |
package |
docker://altinity/clickhouse-server |
Variable |
version |
21.8.15.15.altinitystable |
When running the CI/CD pipeline, provide the following variables:
| Variables | ||
|---|---|---|
Variable |
package |
deb://builds.altinity.cloud/apt-repo/pool/main |
Variable |
version |
The version to use for tests. For example, 21.8.8.1.altinitystable |
Get the link to the deb package: PR -> ClickHouse build check (actions) -> Details -> copy link to a deb package.
Break down the link into CI/CD variables:
Example values using https://s3.amazonaws.com/clickhouse-builds/37882/f74618722585d507cf5fe6d9284cf32028c67716/package_release/clickhouse-client_22.7.1.1738_amd64.deb
| Variables | ||
|---|---|---|
Variable |
package |
deb://s3.amazonaws.com/clickhouse-builds/37882/f74618722585d507cf5fe6d9284cf32028c67716/package_release NOTE: 'deb' instead of 'https' and no '/' on the end. |
Variable |
version |
22.7.1.1738 |
Variable |
package_version_postfix |
By default all (supports older versions), specify amd64 for newer PRs where all packages have amd64 postfix. |
To run the CI/CD pipline you can use cicd-trigger.py
./cicd-trigger.py
with following options
-w --wait Wait for pipeline to finish.
--package Specify docker:// or deb:// package.
--version Specify clickhouse version.
--package-postfix Postfix of the clickhouse-server and clickhouse-client deb package, default: 'amd64'. Choices 'amd64', 'all'.
--suite Select test suite to run, default: 'all'. Choices "all", "aes_encryption", "aggregate_functions", "atomic_insert", "base_58", "clickhouse_keeper", "datetime64_extended_range", "disk_level_encryption", "dns", "example", "extended_precision_data_types", "kafka", "kerberos", "ldap", "lightweight_delete", "map_type", "parquet", "part_moves_between_shards", "rbac", "s3", "s3_aws", "s3_gcs", "selects", "ssl_server", "tiered_storage", "tiered_storage_aws", "tiered_storage_gcs", "window_functions", "benchmark".
--output Tests stdout output style, default: 'classic'. Choices 'nice', 'classic', 'short', etc.
--parallel Enable or disable running tests in parallel, default: 'on'. Choices 'on', 'off'.
--token Personal access token or private token with api access to the gitlab project, default: 'GITLAB_TOKEN' environment variable.
--options Extra options that will be added to test run command.
--arch Architecture to run the tests on, default: 'amd64'. Choices 'amd64', 'arm64'.
--branch Specify which branch to run the tests on, default: 'main'.
--artifacts Specify whether to upload artifacts internally or publically, default: 'internal'. Choices 'internal', 'public'. Bucket for internal upload: 'altinity-internal-test-reports'. Bucket for public upload 'altinity-test-reports'.
--debug Enable script running in debug mode, default: 'False'. Choices 'True', 'False'.
You can run tests locally by passing --local and --clickhouse-binary-path to the top level regression.py or
cd into any sub-folders to run suite specific regression.py.
--localspecifies that regression will be run locally--clickhouse-binary-pathspecifies the path to the ClickHouse binary on the host machine that will be used during the regression run. You can also use docker image that should havedocker://prefix. For example,--clickhouse-binary-path docker://clickhouse/clickhouse-server:22.3.6.5-alpine
Note: you can pass
-hor--helpargument to theregression.pyto see a help message.python3 regression.py -h
Note: make sure that the ClickHouse binary has correct permissions. If you are using
/usr/bin/clickhouseits owner and group is set toroot:rootby default and it needs to be changed toclickhouse:clickhouse. You can change the owner and the group using the following command.sudo chown clickhouse:clickhouse /usr/bin/clickhouse
Using the default ClickHouse installation and its server binary at /usr/bin/clickhouse, you can run
regressions locally using the following command.
python3 regression.py --local --clickhouse-binary-path "/usr/bin/clickhouse"You can control verbosity of the output by specifying the output format with -o or --output option.
See --help for more details.
You can run only the selected tests by passing --only option to the regression.py.
For example,
./regression.py --local --clickhouse-binary-path /usr/bin/clickhouse --only "/clickhouse/rbac/syntax/grant privilege/*"will execute all rbac/syntax/grant privilege tests.
If you want to run only a single test such as the /clickhouse/rbac/syntax/grant privilege/grant privileges/privilege='KILL QUERY', on=('*.*',), allow_introspection=False you can do it as follows
./regression.py --local --clickhouse-binary-path /usr/bin/clickhouse --only "/clickhouse/rbac/syntax/grant privilege/grant privileges/privilege='KILL QUERY', on=('[*].[*]',), allow_introspection=False/*"Note that you need to surround special characters such as
*with square brackets, for example[*].
Note that you need to end the filtering pattern with
/*to run all the steps inside the test.
For more information, please see Filtering section in the TestFlows Handbook.
If TestFlows check does not pass you should look at the end of the test_run.txt.out.log to find the list
of failing tests. For example,
clickhouse_testflows_tests_volume
Start tests
➤ Dec 02,2020 22:22:24 /clickhouse
...
Failing
✘ [ Fail ] /clickhouse/rbac/syntax/grant privilege/grant privileges/privilege='SELECT', on=('db0.table0', 'db0.*', '*.*', 'tb0', '*'), allow_column=True, allow_introspection=False
✘ [ Fail ] /clickhouse/rbac/syntax/grant privilege/grant privileges
✘ [ Fail ] /clickhouse/rbac/syntax/grant privilege
✘ [ Fail ] /clickhouse/rbac/syntax
✘ [ Fail ] /clickhouse/rbac
✘ [ Fail ] /clickhouseIn this case the failing test is
/clickhouse/rbac/syntax/grant privilege/grant privileges/privilege='SELECT', on=('db0.table0', 'db0.*', '*.*', 'tb0', '*'), allow_column=True, allow_introspection=False
while the others
✘ [ Fail ] /clickhouse/rbac/syntax/grant privilege/grant privileges
✘ [ Fail ] /clickhouse/rbac/syntax/grant privilege
✘ [ Fail ] /clickhouse/rbac/syntax
✘ [ Fail ] /clickhouse/rbac
✘ [ Fail ] /clickhouse
failed because the first fail gets "bubble-up" the test execution tree all the way to the top level test which is the
/clickhouse.
You need to download the test.log that contains all raw messages.
Once you know the name of the failing test and you have the test.log that contains all the raw messages
for all the tests, you can use tfs show test messages command.
You get the
tfscommand by installing TestFlows.
For example,
cat test.log | tfs show test messages "/clickhouse/rbac/syntax/grant privilege/grant privileges/privilege='SELECT', on=\('db0.table0', 'db0.\*', '\*.\*', 'tb0', '\*'\), allow_column=True, allow_introspection=False"Note: that characters that are treated as special in extended regular expressions need to be escaped. In this case we have to escape the
*,(, and the)characters in the test name.
You can use the test.log with many of the commands provided by the
tfs utility.
See
tfs --helpfor more information.
For example, you can get a list of failing tests from the test.log using the
tfs show fails command as follows
$ cat test.log | tfs show failsor get the results using the tfs show results command as follows
$ cat test.log | tfs show resultsor you can transform the log to see only the new fails using the
tfs transform fail --new command as follows
$ cat test.log | tfs transform fails --newMinio is the default test suite, but can be specificed using --storage minio.
Examples:
Explicit storage declaration:
$ s3/regression.py --local --clickhouse-binary-path docker://clickhouse/clickhouse-server:22.3.6.5-alpine --storage minioUtilizing default values:
$ s3/regression.py --local --clickhouse-binary-path docker://clickhouse/clickhouse-server:22.3.6.5-alpineYou can also specify the minio uri (--minio-uri), root user (--minio-root-user), and root password (--minio-root-password). However, this is not necessary.
Aws requires a region(--aws-s3-region) and a bucket(--aws-s3-bucket) (the bucket must end with /), in addition to the key id and secret access key.
Aws can be specified using --storage aws_s3.
Env variables:
$ export AWS_ACCESS_KEY_ID=
$ export AWS_SECRET_ACCESS_KEY=
$ export AWS_DEFAULT_REGION=
$ export S3_AMAZON_BUCKET=Examples:
Inline:
$ s3/regression.py --local --clickhouse-binary-path docker://clickhouse/clickhouse-server:22.3.6.5-alpine --aws_s3_key_id [masked] --aws_s3_access_key [masked] --aws-s3-bucket [masked] --aws-s3-region [masked] --storage aws_s3Env:
$ s3/regression.py --local --clickhouse-binary-path docker://clickhouse/clickhouse-server:22.3.6.5-alpine --storage aws_s3GCS requires a gcs uri (--gcs-uri) (the uri must end with /), gcs key id (--gcs-key-id), and gcs key secret(--gcs-key-secret), in addition to the s3 key id and secret access key.
GCS can be specified using --storage gcs.
Env variables:
$ export GCS_URI=
$ export GCS_KEY_ID=
$ export GCS_KEY_SECRET=Examples:
Inline:
$ s3/regression.py --local --clickhouse-binary-path docker://clickhouse/clickhouse-server:22.3.6.5-alpine --gcs-uri [masked] --gcs-key-id [masked] --gcs-key-secret [masked] --storage gcsEnv:
$ s3/regression.py --local --clickhouse-binary-path docker://clickhouse/clickhouse-server:22.3.6.5-alpine --storage gcsNormal tiered storage suite does not require any variables to be provided.
From the regression directory, it can be run with the following command:
$ tiered_storage/regression.py --local --clickhouse-binary-path docker://clickhouse/clickhouse-server:22.3.6.5-alpineMinio tiered storage suite only requires that --with-minio is specified.
It can be run with the following command:
$ tiered_storage/regression.py --local --clickhouse-binary-path docker://clickhouse/clickhouse-server:22.3.6.5-alpine --with-minioAWS S3 tiered storage requires an access key (--aws-s3-access-key), a key id (--aws-s3-key-id), and a uri (--aws-s3-uri). The uri must end with /.
These can be passed as environment variables. AWS S3 must be specified using --with-s3amazon.
Env variables:
$ export AWS_ACCESS_KEY_ID=
$ export AWS_SECRET_ACCESS_KEY=
$ export S3_AMAZON_URI=Examples:
Inline:
$ tiered_storage/regression.py --local --clickhouse-binary-path docker://clickhouse/clickhouse-server:22.3.6.5-alpine --aws-s3-key-id [masked] --aws-s3-access-key [masked] --aws-s3-uri [masked] --with-s3amazonEnv:
$ tiered_storage/regression.py --local --clickhouse-binary-path docker://clickhouse/clickhouse-server:22.3.6.5-alpine --with-s3amazonGCS tiered storage requires a gcs uri (--gcs-uri) (the uri must end with /), gcs key id (--gcs-key-id), and gcs key secret(--gcs-key-secret).
GCS can be specified using --with-s3gcs.
Env variables:
$ export GCS_URI=
$ export GCS_KEY_ID=
$ export GCS_KEY_SECRET=Examples:
Inline:
$ tiered_storage/regression.py --local --clickhouse-binary-path docker://clickhouse/clickhouse-server:22.3.6.5-alpine --gcs-uri [masked] --gcs-key-id [masked] --gcs-key-secret [masked] --with-s3gcsEnv:
$ tiered_storage/regression.py --local --clickhouse-binary-path docker://clickhouse/clickhouse-server:22.3.6.5-alpine --with-s3gcsYou can explicitly specify PAUSE_BEFORE, PAUSE_AFTER, PAUSE_ON_PASS and PAUSE_ON_FAIL flags inside your test program.
For example,
with Test("my test"):
with Step("my step 1", flags=PAUSE_BEFORE):
note("my step 1")
with Step("my step 2", flags=PAUSE_AFTER):
note("my step 2")
with Step("my step 2", flags=PAUSE_ON_PASS):
note("my step 2")
with Step("my step 2", flags=PAUSE_ON_FAIL):
note("my step 2")For decorated tests Flags decorator can be used to set these flags.
@TestScenario
@Flags(PAUSE_BEFORE|PAUSE_AFTER) # pause before and after this test
def my_scenario(self):
passThis can be used for getting access to Docker Compose environment with condition equal to cluster condition on current step by executing standard Docker Compose commands ("ps", "exec" etc.) from "*_env" folder. It allows to make some manual checks/changes on dockers and continue test with new manually set conditions.
To run GitHub actions, navigate to Actions, select Run CI/CD regression tests. Inside Run workflow dropdown menu specify the package, version, suite and upload destination of artifacts.
Package: docker:// or https:// package specifier to use for tests. For example:
- docker://altinity/clickhouse-server
- docker://clickhouse/clickhouse-server
- https://s3.amazonaws.com/altinity-build-artifacts/217/acf34c9fc6932aaf9af69425612070b50529f484/package_release/clickhouse-client_22.8.11.17.altinitystable_amd64.deb
Version: Version of clickhouse to use for tests. The test verifies that node version matches specified version. When package option uses docker:// specifier then the version is the image tag. For example:
- 22.3.9.19-alpine
- 22.3.8.40.altinitystable
- latest
Suite: Which suite to run. Default runs all suites.
Artifacts: Specify whether to upload to internal or public s3 bucket. 'altinity-internal-test-reports' for internal upload, 'altinity-test-reports' for public.