Skip to content
Merged
Show file tree
Hide file tree
Changes from 3 commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 2 additions & 1 deletion .github/workflows/util/install-spark-resources.sh
Original file line number Diff line number Diff line change
Expand Up @@ -80,7 +80,8 @@ function install_spark() {
mv python ${INSTALL_DIR}/shims/spark"${spark_version_short}""${scala_suffix}"/spark_home
mv bin ${INSTALL_DIR}/shims/spark"${spark_version_short}""${scala_suffix}"/spark_home

tar --strip-components=1 -xf "${local_source}" spark-"${spark_version}"/sql/core/src/test/resources/
tar --strip-components=1 -xf "${local_source}" spark-"${spark_version}"/sql/core/src/test/resources/ \
spark-"${spark_version}"/sql/hive/src/test/resources/
mkdir -p shims/spark"${spark_version_short}${scala_suffix}"/spark_home/
mv sql shims/spark"${spark_version_short}${scala_suffix}"/spark_home/

Expand Down
16 changes: 13 additions & 3 deletions .github/workflows/velox_backend_enhanced.yml
Original file line number Diff line number Diff line change
Expand Up @@ -130,7 +130,7 @@ jobs:
export SPARK_HOME=/opt/shims/spark34/spark_home/
ls -l $SPARK_HOME
$MVN_CMD clean test -Pspark-3.4 -Pjava-17 -Pbackends-velox -Piceberg -Piceberg-test -Pdelta -Phudi \
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTest \
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.spark.tags.SlowHiveTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTest \
-DargLine="-Dspark.test.home=$SPARK_HOME"
- name: Upload test report
if: always()
Expand Down Expand Up @@ -188,7 +188,7 @@ jobs:
export SPARK_HOME=/opt/shims/spark35/spark_home/
ls -l $SPARK_HOME
$MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Phudi \
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTest \
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.spark.tags.SlowHiveTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTest \
-DargLine="-Dspark.test.home=$SPARK_HOME"
- name: Upload test report
if: always()
Expand All @@ -215,6 +215,9 @@ jobs:
spark-test-spark35-slow:
needs: build-native-lib-centos-7
runs-on: ubuntu-22.04
env: #TODO remove after image update
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Could you provide a bit more detail in the PR description on how we should follow up? Thanks.

Copy link
Copy Markdown
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This is an existing pain point in gluten image update:

  • Docker image is pushed to apache/gluten repo when merge a related commit to main branch
  • This PR that needs update the image is still test on existing 'old' image

Thus we need add scripts in test action before the PR is merged, and remove such actions when the PR is merged and image is updated.

I'm considering to let the such PR that would update the image build a snapshot image and test on the snapshot image, in this case we do not need add additional scripts on test action and do not need a follow-up remove operation.

LANG: C.UTF-8
LC_ALL: C.UTF-8
container: apache/gluten:centos-8-jdk8
steps:
- uses: actions/checkout@v4
Expand All @@ -228,6 +231,10 @@ jobs:
with:
name: arrow-jars-enhanced-centos-7-${{github.sha}}
path: /root/.m2/repository/org/apache/arrow/
- name: Prepare Spark Resources for Spark 3.5.5 #TODO remove after image update
run: |
rm -rf /opt/shims/spark35
bash .github/workflows/util/install-spark-resources.sh 3.5
- name: Build and Run unit test for Spark 3.5.5 (slow tests)
run: |
cd $GITHUB_WORKSPACE/
Expand All @@ -238,6 +245,9 @@ jobs:
$MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Phudi -Ppaimon -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/" \
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
$MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Phudi -Ppaimon -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/" \
-DtagsToInclude=org.apache.spark.tags.SlowHiveTest
- name: Upload test report
if: always()
uses: actions/upload-artifact@v4
Expand Down Expand Up @@ -293,7 +303,7 @@ jobs:
java -version
$MVN_CMD clean test -Pspark-4.0 -Pscala-2.13 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta \
-Pspark-ut -DargLine="-Dspark.test.home=/opt/shims/spark40/spark_home/" \
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTest
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.spark.tags.SlowHiveTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTest
- name: Upload test report
if: always()
uses: actions/upload-artifact@v4
Expand Down
65 changes: 58 additions & 7 deletions .github/workflows/velox_backend_x86.yml
Original file line number Diff line number Diff line change
Expand Up @@ -680,6 +680,8 @@ jobs:
spark-test-spark33:
needs: build-native-lib-centos-7
runs-on: ubuntu-22.04
env:
SPARK_TESTING: true
container: apache/gluten:centos-8-jdk8
steps:
- uses: actions/checkout@v4
Expand Down Expand Up @@ -710,7 +712,7 @@ jobs:
java -version
$MVN_CMD clean test -Pspark-3.3 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Phudi -Ppaimon -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark33/spark_home/" \
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.spark.tags.SlowHiveTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest
- name: Upload test report
if: always()
uses: actions/upload-artifact@v4
Expand All @@ -736,6 +738,10 @@ jobs:
spark-test-spark33-slow:
needs: build-native-lib-centos-7
runs-on: ubuntu-22.04
env:
SPARK_TESTING: true
LANG: C.UTF-8 #TODO remove after image update
LC_ALL: C.UTF-8
container: apache/gluten:centos-8-jdk8
steps:
- uses: actions/checkout@v4
Expand All @@ -749,6 +755,10 @@ jobs:
with:
name: arrow-jars-centos-7-${{github.sha}}
path: /root/.m2/repository/org/apache/arrow/
- name: Prepare Spark Resources for Spark 3.3.1 #TODO remove after image update
run: |
rm -rf /opt/shims/spark33
bash .github/workflows/util/install-spark-resources.sh 3.3

- name: Build and Run unit test for Spark 3.3.1 (slow tests)
run: |
Expand All @@ -760,6 +770,9 @@ jobs:
$MVN_CMD clean test -Pspark-3.3 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Phudi -Ppaimon -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark33/spark_home/" \
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
$MVN_CMD clean test -Pspark-3.3 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Phudi -Ppaimon -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark33/spark_home/" \
-DtagsToInclude=org.apache.spark.tags.SlowHiveTest
- name: Upload test report
if: always()
uses: actions/upload-artifact@v4
Expand Down Expand Up @@ -810,7 +823,7 @@ jobs:
export SPARK_HOME=/opt/shims/spark34/spark_home/
ls -l $SPARK_HOME
$MVN_CMD clean test -Pspark-3.4 -Pjava-17 -Pbackends-velox -Piceberg -Piceberg-test -Pdelta -Phudi -Ppaimon -Pspark-ut \
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest \
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.spark.tags.SlowHiveTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest \
-DargLine="-Dspark.test.home=$SPARK_HOME"
- name: Upload test report
if: always()
Expand All @@ -837,6 +850,9 @@ jobs:
spark-test-spark34-slow:
needs: build-native-lib-centos-7
runs-on: ubuntu-22.04
env: #TODO remove after image update
LANG: C.UTF-8
LC_ALL: C.UTF-8
container: apache/gluten:centos-8-jdk8
steps:
- uses: actions/checkout@v4
Expand All @@ -850,6 +866,10 @@ jobs:
with:
name: arrow-jars-centos-7-${{github.sha}}
path: /root/.m2/repository/org/apache/arrow/
- name: Prepare Spark Resources for Spark 3.4.4 #TODO remove after image update
run: |
rm -rf /opt/shims/spark34
bash .github/workflows/util/install-spark-resources.sh 3.4
- name: Build and Run unit test for Spark 3.4.4 (slow tests)
run: |
cd $GITHUB_WORKSPACE/
Expand All @@ -862,6 +882,9 @@ jobs:
$MVN_CMD clean test -Pspark-3.4 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Phudi -Ppaimon -Pspark-ut \
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest \
-DargLine="-Dspark.test.home=$SPARK_HOME"
$MVN_CMD clean test -Pspark-3.4 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Phudi -Ppaimon -Pspark-ut \
-DtagsToInclude=org.apache.spark.tags.SlowHiveTest \
-DargLine="-Dspark.test.home=$SPARK_HOME"
- name: Upload test report
if: always()
uses: actions/upload-artifact@v4
Expand Down Expand Up @@ -913,7 +936,7 @@ jobs:
java -version
$MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Phudi -Ppaimon -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/" \
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.spark.tags.SlowHiveTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest
- name: Upload test report
if: always()
uses: actions/upload-artifact@v4
Expand Down Expand Up @@ -971,7 +994,7 @@ jobs:
java -version
$MVN_CMD clean test -Pspark-3.5 -Pscala-2.13 -Pjava-17 -Pbackends-velox -Piceberg \
-Pdelta -Pspark-ut -DargLine="-Dspark.test.home=/opt/shims/spark35-scala-2.13/spark_home/" \
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.spark.tags.SlowHiveTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest
- name: Upload test report
if: always()
uses: actions/upload-artifact@v4
Expand All @@ -993,6 +1016,8 @@ jobs:
runs-on: ubuntu-22.04
env:
SPARK_TESTING: true
LANG: C.UTF-8 #TODO remove after image update
LC_ALL: C.UTF-8
container: apache/gluten:centos-8-jdk8
steps:
- uses: actions/checkout@v4
Expand All @@ -1006,6 +1031,10 @@ jobs:
with:
name: arrow-jars-centos-7-${{github.sha}}
path: /root/.m2/repository/org/apache/arrow/
- name: Prepare Spark Resources for Spark 3.5.5 #TODO remove after image update
run: |
rm -rf /opt/shims/spark35
bash .github/workflows/util/install-spark-resources.sh 3.5
- name: Build and Run unit test for Spark 3.5.5 (slow tests)
run: |
cd $GITHUB_WORKSPACE/
Expand All @@ -1016,6 +1045,9 @@ jobs:
$MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Phudi -Ppaimon -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/" \
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
$MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Phudi -Ppaimon -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/" \
-DtagsToInclude=org.apache.spark.tags.SlowHiveTest
- name: Upload test report
if: always()
uses: actions/upload-artifact@v4
Expand Down Expand Up @@ -1067,7 +1099,7 @@ jobs:
java -version
$MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Ppaimon -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/ -Dspark.gluten.sql.columnar.forceShuffledHashJoin=false" \
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.spark.tags.SlowHiveTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest
- name: Upload test report
uses: actions/upload-artifact@v4
with:
Expand All @@ -1088,6 +1120,8 @@ jobs:
runs-on: ubuntu-22.04
env:
SPARK_TESTING: true
LANG: C.UTF-8 #TODO remove after image update
LC_ALL: C.UTF-8
container: apache/gluten:centos-8-jdk8
steps:
- uses: actions/checkout@v4
Expand All @@ -1101,6 +1135,10 @@ jobs:
with:
name: arrow-jars-centos-7-${{github.sha}}
path: /root/.m2/repository/org/apache/arrow/
- name: Prepare Spark Resources for Spark 3.5.5 #TODO remove after image update
run: |
rm -rf /opt/shims/spark35
bash .github/workflows/util/install-spark-resources.sh 3.5
- name: Build and Run unit test for Spark 3.5.5 (slow tests)
run: |
cd $GITHUB_WORKSPACE/
Expand All @@ -1111,6 +1149,9 @@ jobs:
$MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Ppaimon -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/ -Dspark.gluten.sql.columnar.forceShuffledHashJoin=false" \
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
$MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -Ppaimon -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/ -Dspark.gluten.sql.columnar.forceShuffledHashJoin=false" \
-DtagsToInclude=org.apache.spark.tags.SlowHiveTest
- name: Upload test report
uses: actions/upload-artifact@v4
with:
Expand Down Expand Up @@ -1281,7 +1322,7 @@ jobs:
java -version
$MVN_CMD clean test -Pspark-4.0 -Pscala-2.13 -Pjava-17 -Pbackends-velox -Pdelta -Ppaimon \
-Pspark-ut -DargLine="-Dspark.test.home=/opt/shims/spark40/spark_home/" \
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.spark.tags.SlowHiveTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest
- name: Upload test report
if: always()
uses: actions/upload-artifact@v4
Expand All @@ -1303,6 +1344,8 @@ jobs:
runs-on: ubuntu-22.04
env:
SPARK_TESTING: true
LANG: C.UTF-8 #TODO remove after image update
LC_ALL: C.UTF-8
container: apache/gluten:centos-8-jdk17
steps:
- uses: actions/checkout@v4
Expand Down Expand Up @@ -1330,6 +1373,9 @@ jobs:
$MVN_CMD clean test -Pspark-4.0 -Pscala-2.13 -Pjava-17 -Pbackends-velox -Pdelta -Ppaimon -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark40/spark_home/" \
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
$MVN_CMD clean test -Pspark-4.0 -Pscala-2.13 -Pjava-17 -Pbackends-velox -Pdelta -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark40/spark_home/" \
-DtagsToInclude=org.apache.spark.tags.SlowHiveTest
- name: Upload test report
if: always()
uses: actions/upload-artifact@v4
Expand Down Expand Up @@ -1387,7 +1433,7 @@ jobs:
java -version
$MVN_CMD clean test -Pspark-4.1 -Pscala-2.13 -Pjava-17 -Pbackends-velox \
-Pspark-ut -DargLine="-Dspark.test.home=/opt/shims/spark41/spark_home/" \
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.spark.tags.SlowHiveTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.EnhancedFeaturesTest,org.apache.gluten.tags.SkipTest
- name: Upload test report
if: always()
uses: actions/upload-artifact@v4
Expand All @@ -1409,6 +1455,8 @@ jobs:
runs-on: ubuntu-22.04
env:
SPARK_TESTING: true
LANG: C.UTF-8 #TODO remove after image update
LC_ALL: C.UTF-8
container: apache/gluten:centos-9-jdk17
steps:
- uses: actions/checkout@v4
Expand Down Expand Up @@ -1436,6 +1484,9 @@ jobs:
$MVN_CMD clean test -Pspark-4.1 -Pscala-2.13 -Pjava-17 -Pbackends-velox -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark41/spark_home/" \
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
$MVN_CMD clean test -Pspark-4.1 -Pscala-2.13 -Pjava-17 -Pbackends-velox -Pspark-ut \
-DargLine="-Dspark.test.home=/opt/shims/spark41/spark_home/" \
-DtagsToInclude=org.apache.spark.tags.SlowHiveTest
- name: Upload test report
if: always()
uses: actions/upload-artifact@v4
Expand Down
2 changes: 2 additions & 0 deletions dev/docker/Dockerfile.centos8-dynamic-build
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,8 @@ ARG JAVA_VERSION=1.8.0
ENV JAVA_HOME=/usr/lib/jvm/java-${JAVA_VERSION}-openjdk
ENV PATH=$JAVA_HOME/bin:$PATH

ENV LANG=C.UTF-8
ENV LC_ALL=C.UTF-8

RUN set -ex; \
sed -i -e "s|mirrorlist=|#mirrorlist=|g" /etc/yum.repos.d/CentOS-* || true; \
Expand Down
2 changes: 2 additions & 0 deletions dev/docker/Dockerfile.centos9-dynamic-build
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,8 @@ ARG JAVA_VERSION=1.8.0
ENV JAVA_HOME=/usr/lib/jvm/java-${JAVA_VERSION}-openjdk
ENV PATH=$JAVA_HOME/bin:$PATH

ENV LANG=C.UTF-8
ENV LC_ALL=C.UTF-8

RUN set -ex; \
yum update -y && yum install -y epel-release sudo dnf && yum install -y ccache; \
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ import org.apache.spark.sql.execution.joins.{GlutenBroadcastJoinSuite, GlutenExi
import org.apache.spark.sql.execution.python._
import org.apache.spark.sql.extension.{GlutenCollapseProjectExecTransformerSuite, GlutenSessionExtensionSuite}
import org.apache.spark.sql.gluten.GlutenFallbackSuite
import org.apache.spark.sql.hive.execution.GlutenHiveSQLQuerySuite
import org.apache.spark.sql.hive.execution._
import org.apache.spark.sql.sources._

// Some settings' line length exceeds 100
Expand Down Expand Up @@ -829,7 +829,7 @@ class VeloxTestSettings extends BackendTestSettings {
// enableSuite[GlutenSimpleShowCreateTableSuite]
enableSuite[GlutenFileSourceSQLInsertTestSuite]
enableSuite[GlutenDSV2SQLInsertTestSuite]
enableSuite[GlutenSQLQuerySuite]
enableSuite[org.apache.spark.sql.GlutenSQLQuerySuite]
// Decimal precision exceeds.
.exclude("should be able to resolve a persistent view")
// Unstable. Needs to be fixed.
Expand Down Expand Up @@ -867,7 +867,38 @@ class VeloxTestSettings extends BackendTestSettings {
.exclude("cases when literal is max")
enableSuite[GlutenXPathFunctionsSuite]
enableSuite[GlutenFallbackSuite]
enableSuite[GlutenHashAggregationQuerySuite]
// TODO: fix on https://github.com/apache/gluten/issues/11919
.exclude("udaf with all data types")
enableSuite[GlutenHashAggregationQueryWithControlledFallbackSuite]
// TODO: fix on https://github.com/apache/gluten/issues/11919
.exclude("udaf with all data types")
enableSuite[GlutenHiveCommandSuite]
enableSuite[GlutenHiveDDLSuite]
enableSuite[GlutenHiveExplainSuite]
.exclude("explain output of physical plan should contain proper codegen stage ID")
.exclude("EXPLAIN CODEGEN command")
enableSuite[GlutenHivePlanTest]
enableSuite[GlutenHiveQuerySuite]
enableSuite[GlutenHiveResolutionSuite]
enableSuite[GlutenHiveSQLQuerySuite]
enableSuite[GlutenHiveSQLViewSuite]
enableSuite[GlutenHiveScriptTransformationSuite]
enableSuite[GlutenHiveSerDeReadWriteSuite]
enableSuite[GlutenHiveSerDeSuite]
enableSuite[GlutenHiveTableScanSuite]
enableSuite[GlutenHiveTypeCoercionSuite]
enableSuite[GlutenHiveUDAFSuite]
enableSuite[GlutenHiveUDFSuite]
enableSuite[GlutenObjectHashAggregateSuite]
enableSuite[GlutenPruneHiveTablePartitionsSuite]
enableSuite[GlutenPruningSuite]
enableSuite[GlutenSQLMetricsSuite]
enableSuite[org.apache.spark.sql.hive.execution.GlutenSQLQuerySuite]
enableSuite[GlutenHashUDAQuerySuite]
enableSuite[GlutenHashUDAQueryWithControlledFallbackSuite]
enableSuite[GlutenSQLQuerySuiteAE]
enableSuite[GlutenWindowQuerySuite]
enableSuite[GlutenImplicitsTest]
enableSuite[GlutenCollapseProjectExecTransformerSuite]
enableSuite[GlutenSparkSessionExtensionSuite]
Expand Down
Loading
Loading