From d86d5ab3930b5e730efd0ee64c47c3009ffbade4 Mon Sep 17 00:00:00 2001 From: Piotr Findeisen Date: Wed, 10 Jan 2024 12:30:48 +0100 Subject: [PATCH] Reduce Hive transactional coverage in suites Hive transactional tables are tested as part of the STORAGE_FORMATS group to ensure test coverage of transactional tables with various metastore and HDFS setups (kerberized or not, impersonation or not). Even so, this coverage can be reduced. We do not need to test e.g. various bucketing versions. --- .../hive/TestHiveTransactionalTable.java | 23 +++++++++++++++---- 1 file changed, 19 insertions(+), 4 deletions(-) diff --git a/testing/trino-product-tests/src/main/java/io/trino/tests/product/hive/TestHiveTransactionalTable.java b/testing/trino-product-tests/src/main/java/io/trino/tests/product/hive/TestHiveTransactionalTable.java index 44eeb86f59a4..d94bfb81b3cf 100644 --- a/testing/trino-product-tests/src/main/java/io/trino/tests/product/hive/TestHiveTransactionalTable.java +++ b/testing/trino-product-tests/src/main/java/io/trino/tests/product/hive/TestHiveTransactionalTable.java @@ -57,6 +57,7 @@ import static io.trino.tests.product.TestGroups.HIVE_TRANSACTIONAL; import static io.trino.tests.product.TestGroups.PROFILE_SPECIFIC_TESTS; import static io.trino.tests.product.TestGroups.STORAGE_FORMATS; +import static io.trino.tests.product.hive.BucketingType.BUCKETED_DEFAULT; import static io.trino.tests.product.hive.BucketingType.BUCKETED_V2; import static io.trino.tests.product.hive.BucketingType.NONE; import static io.trino.tests.product.hive.TestHiveTransactionalTable.CompactionMode.MAJOR; @@ -273,7 +274,14 @@ private void testReadInsertOnly(boolean isPartitioned, BucketingType bucketingTy } } - @Test(groups = {STORAGE_FORMATS, HIVE_TRANSACTIONAL, PROFILE_SPECIFIC_TESTS}, dataProvider = "partitioningAndBucketingTypeDataProvider", timeOut = TEST_TIMEOUT) + @Test(groups = {STORAGE_FORMATS, PROFILE_SPECIFIC_TESTS}, timeOut = TEST_TIMEOUT) + @Flaky(issue = ACID_CORRUPTION_DIRECTORY_ISSUE, match = ACID_CORRUPTION_DIRECTORY_RETRY_PATTERN) + public void testReadFullAcidWithOriginalFilesSmoke() + { + testReadFullAcidWithOriginalFiles(true, BUCKETED_DEFAULT); + } + + @Test(groups = {HIVE_TRANSACTIONAL, PROFILE_SPECIFIC_TESTS}, dataProvider = "partitioningAndBucketingTypeDataProvider", timeOut = TEST_TIMEOUT) @Flaky(issue = ACID_CORRUPTION_DIRECTORY_ISSUE, match = ACID_CORRUPTION_DIRECTORY_RETRY_PATTERN) public void testReadFullAcidWithOriginalFiles(boolean isPartitioned, BucketingType bucketingType) { @@ -317,7 +325,7 @@ public void testReadFullAcidWithOriginalFiles(boolean isPartitioned, BucketingTy } } - @Test(groups = {STORAGE_FORMATS, HIVE_TRANSACTIONAL, PROFILE_SPECIFIC_TESTS}, dataProvider = "partitioningAndBucketingTypeDataProvider", timeOut = TEST_TIMEOUT) + @Test(groups = {HIVE_TRANSACTIONAL, PROFILE_SPECIFIC_TESTS}, dataProvider = "partitioningAndBucketingTypeDataProvider", timeOut = TEST_TIMEOUT) @Flaky(issue = RETRYABLE_FAILURES_ISSUES, match = RETRYABLE_FAILURES_MATCH) public void testUpdateFullAcidWithOriginalFilesTrinoInserting(boolean isPartitioned, BucketingType bucketingType) { @@ -367,7 +375,14 @@ public void testUpdateFullAcidWithOriginalFilesTrinoInserting(boolean isPartitio }); } - @Test(groups = {STORAGE_FORMATS, HIVE_TRANSACTIONAL, PROFILE_SPECIFIC_TESTS}, dataProvider = "partitioningAndBucketingTypeDataProvider", timeOut = TEST_TIMEOUT) + @Test(groups = {STORAGE_FORMATS, PROFILE_SPECIFIC_TESTS}, timeOut = TEST_TIMEOUT) + @Flaky(issue = RETRYABLE_FAILURES_ISSUES, match = RETRYABLE_FAILURES_MATCH) + public void testUpdateFullAcidWithOriginalFilesTrinoInsertingAndDeletingSmoke() + { + testUpdateFullAcidWithOriginalFilesTrinoInsertingAndDeleting(true, BUCKETED_DEFAULT); + } + + @Test(groups = {HIVE_TRANSACTIONAL, PROFILE_SPECIFIC_TESTS}, dataProvider = "partitioningAndBucketingTypeDataProvider", timeOut = TEST_TIMEOUT) @Flaky(issue = RETRYABLE_FAILURES_ISSUES, match = RETRYABLE_FAILURES_MATCH) public void testUpdateFullAcidWithOriginalFilesTrinoInsertingAndDeleting(boolean isPartitioned, BucketingType bucketingType) { @@ -419,7 +434,7 @@ String makeValues(int colStart, int colCount, int fcol, boolean isPartitioned, i .collect(Collectors.joining(", ")); } - @Test(groups = {STORAGE_FORMATS, HIVE_TRANSACTIONAL, PROFILE_SPECIFIC_TESTS}, dataProvider = "partitioningAndBucketingTypeDataProvider", timeOut = TEST_TIMEOUT) + @Test(groups = {HIVE_TRANSACTIONAL, PROFILE_SPECIFIC_TESTS}, dataProvider = "partitioningAndBucketingTypeDataProvider", timeOut = TEST_TIMEOUT) @Flaky(issue = ACID_CORRUPTION_DIRECTORY_ISSUE, match = ACID_CORRUPTION_DIRECTORY_RETRY_PATTERN) public void testReadInsertOnlyWithOriginalFiles(boolean isPartitioned, BucketingType bucketingType) {