Skip to content

Commit

Permalink
Remove ConfigHdp3 since it's same as the default config
Browse files Browse the repository at this point in the history
  • Loading branch information
nineinchnick authored and wendigo committed Sep 13, 2023
1 parent c36646b commit ceb8e15
Show file tree
Hide file tree
Showing 46 changed files with 92 additions and 159 deletions.
92 changes: 17 additions & 75 deletions .github/workflows/ci.yml
Original file line number Diff line number Diff line change
Expand Up @@ -902,14 +902,15 @@ jobs:
cat <<EOF > .github/test-pt-matrix.yaml
config:
- default
- hdp3
# TODO: config-apache-hive3
suite:
- suite-1
- suite-2
- suite-3
# suite-4 does not exist
- suite-5
- suite-6-non-generic
- suite-7-non-generic
- suite-8-non-generic
- suite-azure
- suite-delta-lake-databricks91
- suite-delta-lake-databricks104
Expand All @@ -920,53 +921,45 @@ jobs:
- suite-clients
- suite-functions
- suite-tpch
- suite-tpcds
- suite-storage-formats-detailed
- suite-parquet
- suite-oauth2
- suite-ldap
- suite-compatibility
- suite-all-connectors-smoke
- suite-delta-lake-oss
- suite-kafka
- suite-cassandra
- suite-clickhouse
- suite-mysql
- suite-iceberg
- suite-hudi
- suite-ignite
exclude:
- config: default
ignore exclusion if: >-
${{ github.event_name != 'pull_request'
|| github.event.pull_request.head.repo.full_name == github.repository
|| contains(github.event.pull_request.labels.*.name, 'tests:all')
|| contains(github.event.pull_request.labels.*.name, 'tests:hive')
}}
- suite: suite-azure
config: default
- suite: suite-azure
ignore exclusion if: >-
${{ env.CI_SKIP_SECRETS_PRESENCE_CHECKS != '' ||
secrets.AZURE_ABFS_CONTAINER != '' ||
secrets.AZURE_ABFS_ACCOUNT != '' ||
secrets.AZURE_ABFS_ACCESSKEY != '' }}
- suite: suite-gcs
config: default
- suite: suite-gcs
ignore exclusion if: >-
${{ env.CI_SKIP_SECRETS_PRESENCE_CHECKS != '' || secrets.GCP_CREDENTIALS_KEY != '' }}
- suite: suite-delta-lake-databricks91
config: hdp3
- suite: suite-delta-lake-databricks91
ignore exclusion if: >-
${{ env.CI_SKIP_SECRETS_PRESENCE_CHECKS != '' || secrets.DATABRICKS_TOKEN != '' }}
- suite: suite-delta-lake-databricks104
config: hdp3
- suite: suite-delta-lake-databricks104
ignore exclusion if: >-
${{ env.CI_SKIP_SECRETS_PRESENCE_CHECKS != '' || secrets.DATABRICKS_TOKEN != '' }}
- suite: suite-delta-lake-databricks113
config: hdp3
- suite: suite-delta-lake-databricks113
ignore exclusion if: >-
${{ env.CI_SKIP_SECRETS_PRESENCE_CHECKS != '' || secrets.DATABRICKS_TOKEN != '' }}
- suite: suite-delta-lake-databricks122
config: hdp3
- suite: suite-delta-lake-databricks122
ignore exclusion if: >-
${{ env.CI_SKIP_SECRETS_PRESENCE_CHECKS != '' || secrets.DATABRICKS_TOKEN != '' }}
- suite: suite-delta-lake-databricks133
config: hdp3
- suite: suite-delta-lake-databricks133
ignore exclusion if: >-
${{ env.CI_SKIP_SECRETS_PRESENCE_CHECKS != '' || secrets.DATABRICKS_TOKEN != '' }}
Expand All @@ -983,60 +976,9 @@ jobs:
# value of the property, and the exclusion will apply normally.
- "false"
include:
# this suite is not meant to be run with different configs
- config: default
suite: suite-6-non-generic
# this suite is not meant to be run with different configs
- config: default
suite: suite-7-non-generic
# this suite is not meant to be run with different configs
- config: default
suite: suite-8-non-generic
# this suite is not meant to be run with different configs
- config: default
suite: suite-tpcds
# this suite is not meant to be run with different configs
- config: default
suite: suite-parquet
# this suite is not meant to be run with different configs
- config: default
suite: suite-oauth2
# this suite is not meant to be run with different configs
- config: default
suite: suite-ldap
# this suite is not meant to be run with different configs
- config: default
suite: suite-compatibility
# this suite is designed specifically for apache-hive3. TODO remove the suite once we can run all regular tests on apache-hive3.
- config: apache-hive3
suite: suite-hms-only
# this suite is not meant to be run with different configs
- config: default
suite: suite-all-connectors-smoke
# this suite is not meant to be run with different configs
- config: default
suite: suite-delta-lake-oss
# this suite is not meant to be run with different configs
- config: default
suite: suite-kafka
# this suite is not meant to be run with different configs
- config: default
suite: suite-cassandra
# this suite is not meant to be run with different configs
- config: default
suite: suite-clickhouse
# this suite is not meant to be run with different configs
- config: default
suite: suite-mysql
# this suite is not meant to be run with different configs
- config: default
suite: suite-iceberg
# this suite is not meant to be run with different configs
- config: default
suite: suite-hudi
# this suite is not meant to be run with different configs
- config: default
suite: suite-ignite
EOF
- name: Build PT matrix (all)
if: |
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -81,6 +81,8 @@ public static DockerContainer createHadoopContainer(DockerFiles dockerFiles, Por
.withCopyFileToContainer(forHostPath(dockerFiles.getDockerFilesHostPath("health-checks/hadoop-health-check.sh")), CONTAINER_HEALTH_D + "hadoop-health-check.sh")
.withCopyFileToContainer(forHostPath(dockerFiles.getDockerFilesHostPath("common/hadoop/hadoop-run.sh")), "/usr/local/hadoop-run.sh")
.withCopyFileToContainer(forHostPath(dockerFiles.getDockerFilesHostPath("common/hadoop/apply-config-overrides.sh")), CONTAINER_HADOOP_INIT_D + "00-apply-config-overrides.sh")
// When hive performs implicit coercion to/from timestamp for ORC files, it depends on timezone of the HiveServer
.withEnv("TZ", "UTC")
.withCommand("/usr/local/hadoop-run.sh")
.withExposedLogPaths("/var/log/hadoop-yarn", "/var/log/hadoop-hdfs", "/var/log/hive", "/var/log/container-health.log")
.withStartupCheckStrategy(new IsRunningStartupCheckStrategy())
Expand Down

This file was deleted.

Original file line number Diff line number Diff line change
Expand Up @@ -87,7 +87,6 @@ public void extendEnvironment(Environment.Builder builder)

String containerGcpCredentialsFile = CONTAINER_TRINO_ETC + "gcp-credentials.json";
builder.configureContainer(HADOOP, container -> {
container.setDockerImageName("ghcr.io/trinodb/testing/hdp3.1-hive:" + hadoopImageVersion);
container.withCopyFileToContainer(
forHostPath(getCoreSiteOverrideXml(containerGcpCredentialsFile)),
"/docker/presto-product-tests/conf/environment/multinode-gcs/core-site-overrides.xml");
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -75,9 +75,6 @@ public EnvSinglenodeHiveHudiRedirections(
@Override
public void extendEnvironment(Environment.Builder builder)
{
// Using hdp3.1 so we are using Hive metastore with version close to versions of hive-*.jars Spark uses
builder.configureContainer(HADOOP, container -> container.setDockerImageName("ghcr.io/trinodb/testing/hdp3.1-hive:" + hadoopImagesVersion));

builder.addConnector("hive", forHostPath(configDir.getPath("hive.properties")));
builder.addConnector("hudi", forHostPath(configDir.getPath("hudi.properties")));

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -80,8 +80,6 @@ public EnvSinglenodeHudi(
@Override
public void extendEnvironment(Environment.Builder builder)
{
// Using hdp3.1 so we are using Hive metastore with version close to versions of hive-*.jars Spark uses
builder.configureContainer(HADOOP, container -> container.setDockerImageName("ghcr.io/trinodb/testing/hdp3.1-hive:" + hadoopImagesVersion));
builder.addConnector(
"hive",
forHostPath(configDir.getPath("hive.properties")),
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -55,7 +55,6 @@ public EnvSinglenodeSparkHiveNoStatsFallback(Standard standard, Hadoop hadoop, D
@Override
public void extendEnvironment(Environment.Builder builder)
{
builder.configureContainer(HADOOP, container -> container.setDockerImageName("ghcr.io/trinodb/testing/hdp3.1-hive:" + hadoopImagesVersion));
builder.addConnector("hive", forHostPath(configDir.getPath("hive.properties")));
builder.addContainer(createSpark()).containerDependsOn("spark", HADOOP);
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,3 +14,5 @@ hive.fs.cache.max-size=10
hive.max-partitions-per-scan=100
hive.max-partitions-for-eager-load=100
hive.hive-views.enabled=true
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -5,3 +5,5 @@ hive.allow-drop-table=true
hive.metastore-cache-ttl=0s
hive.hive-views.enabled=true
hive.timestamp-precision=NANOSECONDS
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -15,3 +15,5 @@ hive.max-partitions-per-scan=100
hive.max-partitions-for-eager-load=100
hive.hive-views.enabled=true
hive.non-managed-table-writes-enabled=true
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -6,3 +6,5 @@ hive.fs.cache.max-size=10
hive.hive-views.enabled=true
hive.hive-views.run-as-invoker=true
hive.security=sql-standard
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC

This file was deleted.

Original file line number Diff line number Diff line change
Expand Up @@ -11,3 +11,5 @@ hive.allow-comment-table=true
hive.allow-drop-table=true
hive.allow-rename-table=true
hive.translate-hive-views=true
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -3,3 +3,5 @@ hive.config.resources=/docker/presto-product-tests/conf/presto/etc/hive-default-
hive.metastore.uri=thrift://hadoop-master:9083
hive.allow-drop-table=true
hive.cache.enabled=true
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -4,3 +4,5 @@ hive.metastore.uri=thrift://hadoop-master:9083
hive.allow-drop-table=true
hive.cache.enabled=true
hive.cache.location=/tmp/cache
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -9,3 +9,5 @@ hive.allow-rename-column=true
hive.allow-comment-table=true
hive.allow-drop-table=true
hive.allow-rename-table=true
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -6,3 +6,5 @@ hive.s3.endpoint=http://minio:9080/
hive.s3.path-style-access=true
hive.s3.ssl.enabled=false
hive.allow-drop-table=true
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,8 @@ hive.metastore-cache-ttl=0s
hive.allow-add-column=true
hive.allow-drop-column=true
hive.allow-rename-column=true
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC

hive.metastore.authentication.type=KERBEROS
hive.metastore.service.principal=hive/[email protected]
Expand Down
Original file line number Diff line number Diff line change
@@ -1,2 +1,4 @@
connector.name=hive-hadoop2
hive.metastore.uri=thrift://hadoop-master:9083
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
@@ -1,2 +1,4 @@
connector.name=hive
hive.metastore.uri=thrift://hadoop-master:9083
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -11,3 +11,5 @@ hive.allow-comment-table=true
hive.allow-comment-column=true
hive.allow-rename-table=true
hive.delta-lake-catalog-name=delta
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -13,3 +13,5 @@ hive.allow-comment-table=true
hive.allow-comment-column=true
hive.allow-rename-table=true
hive.delta-lake-catalog-name=delta
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -15,3 +15,5 @@ hive.fs.cache.max-size=10
hive.max-partitions-per-scan=100
hive.max-partitions-for-eager-load=100
hive.non-managed-table-writes-enabled=true
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -9,3 +9,5 @@ hive.allow-drop-table=true
hive.allow-rename-table=true
hive.hudi-catalog-name=hudi
hive.hive-views.enabled=true
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -9,3 +9,5 @@ hive.allow-drop-table=true
hive.allow-rename-table=true
hive.iceberg-catalog-name=iceberg
hive.hive-views.enabled=true
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -17,3 +17,5 @@ hive.fs.cache.max-size=10
hive.max-partitions-per-scan=100
hive.max-partitions-for-eager-load=100
hive.non-managed-table-writes-enabled=true
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -2,3 +2,5 @@ connector.name=hive
hive.metastore.uri=thrift://hadoop-master:9083
hive.config.resources=/docker/presto-product-tests/conf/presto/etc/hive-default-fs-site.xml
hive.allow-drop-table=true
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC
Original file line number Diff line number Diff line change
Expand Up @@ -15,6 +15,8 @@ hive.metastore-cache-ttl=0s
hive.allow-add-column=true
hive.allow-drop-column=true
hive.allow-rename-column=true
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC

hive.metastore.authentication.type=KERBEROS
hive.metastore.service.principal=hive/[email protected]
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,8 @@ connector.name=hive
hive.metastore.uri=thrift://hadoop-master:9083
hive.config.resources=/docker/presto-product-tests/conf/presto/etc/hive-default-fs-site.xml,/docker/presto-product-tests/conf/presto/etc/hive-data-protection-site.xml
hive.metastore-cache-ttl=0s
hive.parquet.time-zone=UTC
hive.rcfile.time-zone=UTC

hive.metastore.authentication.type=KERBEROS
hive.metastore.service.principal=hive/[email protected]
Expand Down
Loading

0 comments on commit ceb8e15

Please sign in to comment.