@@ -616,7 +616,7 @@ jobs:
616
616
with :
617
617
name : arrow-jars-centos-7-${{github.sha}}
618
618
path : /root/.m2/repository/org/apache/arrow/
619
- - name : Prepare
619
+ - name : Prepare Python3.9 and PySpark 3.2.2
620
620
run : |
621
621
dnf module -y install python39 && \
622
622
alternatives --set python3 /usr/bin/python3.9 && \
@@ -687,7 +687,7 @@ jobs:
687
687
with :
688
688
name : arrow-jars-centos-7-${{github.sha}}
689
689
path : /root/.m2/repository/org/apache/arrow/
690
- - name : Prepare
690
+ - name : Prepare Python3.9 and PySpark 3.3.1
691
691
run : |
692
692
dnf module -y install python39 && \
693
693
alternatives --set python3 /usr/bin/python3.9 && \
@@ -823,7 +823,7 @@ jobs:
823
823
ls -l /opt/shims/spark34/spark_home/
824
824
$MVN_CMD clean test -Pspark-3.4 -Pjava-8 -Pbackends-velox -Pceleborn -Pdelta -Phudi -Pspark-ut \
825
825
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags \
826
- -DargLine="-Dspark.test.home=/opt/shims/spark34/spark_home/"
826
+ -DargLine="-Dspark.test.home=/opt/shims/spark34/spark_home/"
827
827
- name : Upload test report
828
828
if : always()
829
829
uses : actions/upload-artifact@v4
@@ -892,15 +892,15 @@ jobs:
892
892
ls -l /opt/shims/spark34/spark_home/
893
893
$MVN_CMD clean test -Pspark-3.4 -Pjava-8 -Pbackends-velox -Pceleborn -Pdelta -Pspark-ut -Phudi \
894
894
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest \
895
- -DargLine="-Dspark.test.home=/opt/shims/spark34/spark_home/"
895
+ -DargLine="-Dspark.test.home=/opt/shims/spark34/spark_home/"
896
896
- name : Upload test report
897
897
if : always()
898
898
uses : actions/upload-artifact@v4
899
899
with :
900
900
name : test-report-spark34-slow-jdk8
901
901
path : ' **/surefire-reports/TEST-*.xml'
902
902
903
- run-spark-test-spark35 :
903
+ run-spark-test-spark35-jdk8 :
904
904
needs : build-native-lib-centos-7
905
905
runs-on : ubuntu-20.04
906
906
container : apache/gluten:centos-8
@@ -916,34 +916,34 @@ jobs:
916
916
with :
917
917
name : arrow-jars-centos-7-${{github.sha}}
918
918
path : /root/.m2/repository/org/apache/arrow/
919
- - name : Prepare
919
+ - name : Prepare Python3.9 and PySpark 3.5.5
920
920
run : |
921
921
dnf module -y install python39 && \
922
922
alternatives --set python3 /usr/bin/python3.9 && \
923
923
pip3 install setuptools && \
924
- pip3 install pyspark==3.5.2 cython && \
924
+ pip3 install pyspark==3.5.5 cython && \
925
925
pip3 install pandas pyarrow
926
- - name : Build and Run unit test for Spark 3.5.2 (other tests)
926
+ - name : Build and Run unit test for Spark 3.5.5 (other tests)
927
927
run : |
928
928
cd $GITHUB_WORKSPACE/
929
929
export SPARK_SCALA_VERSION=2.12
930
- $MVN_CMD clean test -Pspark-3.5 -Pbackends-velox -Pceleborn -Piceberg - Pdelta -Phudi -Pspark-ut \
930
+ $MVN_CMD clean test -Pspark-3.5 -Pbackends-velox -Pceleborn -Pdelta -Phudi -Pspark-ut \
931
931
-DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/" \
932
932
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags
933
933
- name : Upload test report
934
934
if : always()
935
935
uses : actions/upload-artifact@v4
936
936
with :
937
- name : test-report-spark35
937
+ name : test-report-spark35-jdk8
938
938
path : ' **/surefire-reports/TEST-*.xml'
939
939
- name : Upload golden files
940
940
if : failure()
941
941
uses : actions/upload-artifact@v4
942
942
with :
943
- name : golden-files-spark35
943
+ name : golden-files-spark35-jdk8
944
944
path : /tmp/tpch-approved-plan/**
945
945
946
- run-spark-test-spark35-jdk17 :
946
+ run-spark-test-spark35 :
947
947
needs : build-native-lib-centos-7
948
948
runs-on : ubuntu-20.04
949
949
container : apache/gluten:centos-8-jdk17
@@ -978,7 +978,7 @@ jobs:
978
978
if : always()
979
979
uses : actions/upload-artifact@v4
980
980
with :
981
- name : test-report-spark35-jdk17
981
+ name : test-report-spark35
982
982
path : ' **/surefire-reports/TEST-*.xml'
983
983
- name : Upload golden files
984
984
if : failure()
@@ -990,7 +990,7 @@ jobs:
990
990
run-spark-test-spark35-scala213 :
991
991
needs : build-native-lib-centos-7
992
992
runs-on : ubuntu-20.04
993
- container : apache/gluten:centos-8
993
+ container : apache/gluten:centos-8-jdk17
994
994
steps :
995
995
- uses : actions/checkout@v2
996
996
- name : Download All Artifacts
@@ -1003,19 +1003,20 @@ jobs:
1003
1003
with :
1004
1004
name : arrow-jars-centos-7-${{github.sha}}
1005
1005
path : /root/.m2/repository/org/apache/arrow/
1006
- - name : Prepare
1006
+ - name : Prepare Python3.9 and PySpark 3.5.5
1007
1007
run : |
1008
1008
dnf module -y install python39 && \
1009
1009
alternatives --set python3 /usr/bin/python3.9 && \
1010
1010
pip3 install setuptools && \
1011
- pip3 install pyspark==3.5.2 cython && \
1011
+ pip3 install pyspark==3.5.5 cython && \
1012
1012
pip3 install pandas pyarrow
1013
- - name : Build and Run unit test for Spark 3.5.2 with scala-2.13 (other tests)
1013
+ - name : Build and Run unit test for Spark 3.5.5 with scala-2.13 (other tests)
1014
1014
run : |
1015
1015
cd $GITHUB_WORKSPACE/
1016
1016
export SPARK_SCALA_VERSION=2.13
1017
- $MVN_CMD clean test -Pspark-3.5 -Pscala-2.13 -Pbackends-velox -Pceleborn -Piceberg \
1018
- -Pdelta -Pspark-ut -DargLine="-Dspark.test.home=/opt/shims/spark35-scala2.13/spark_home/" \
1017
+ export JAVA_HOME=/usr/lib/jvm/java-17-openjdk
1018
+ $MVN_CMD clean test -Pspark-3.5 -Pscala-2.13 -Pjava-17 -Pbackends-velox -Pceleborn -Piceberg \
1019
+ -Pdelta -Pspark-ut -DargLine="-Dspark.test.home=/opt/shims/spark35-scala2.13/spark_home/ ${EXTRA_FLAGS}" \
1019
1020
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags
1020
1021
- name : Upload test report
1021
1022
if : always()
@@ -1024,7 +1025,7 @@ jobs:
1024
1025
name : test-report-spark35-scala213
1025
1026
path : ' **/surefire-reports/TEST-*.xml'
1026
1027
1027
- run-spark-test-spark35-slow :
1028
+ run-spark-test-spark35-slow-jdk8 :
1028
1029
needs : build-native-lib-centos-7
1029
1030
runs-on : ubuntu-20.04
1030
1031
container : apache/gluten:centos-8
@@ -1040,12 +1041,50 @@ jobs:
1040
1041
with :
1041
1042
name : arrow-jars-centos-7-${{github.sha}}
1042
1043
path : /root/.m2/repository/org/apache/arrow/
1043
- - name : Build and Run unit test for Spark 3.5.2 (slow tests)
1044
+ - name : Prepare Spark Resources for Spark 3.5.5
1045
+ run : |
1046
+ rm -rf /opt/shims/spark35
1047
+ bash .github/workflows/util/install_spark_resources.sh 3.5
1048
+ - name : Build and Run unit test for Spark 3.5.5 (slow tests)
1044
1049
run : |
1045
1050
cd $GITHUB_WORKSPACE/
1046
- $MVN_CMD clean test -Pspark-3.5 -Pbackends-velox -Pceleborn -Piceberg - Pdelta -Phudi -Pspark-ut \
1051
+ $MVN_CMD clean test -Pspark-3.5 -Pbackends-velox -Pceleborn -Pdelta -Phudi -Pspark-ut \
1047
1052
-DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/" \
1048
1053
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
1054
+ - name : Upload test report
1055
+ if : always()
1056
+ uses : actions/upload-artifact@v4
1057
+ with :
1058
+ name : test-report-spark35-slow-jdk8
1059
+ path : ' **/surefire-reports/TEST-*.xml'
1060
+
1061
+ run-spark-test-spark35-slow :
1062
+ needs : build-native-lib-centos-7
1063
+ runs-on : ubuntu-20.04
1064
+ container : apache/gluten:centos-8-jdk17
1065
+ steps :
1066
+ - uses : actions/checkout@v2
1067
+ - name : Download All Artifacts
1068
+ uses : actions/download-artifact@v4
1069
+ with :
1070
+ name : velox-native-lib-centos-7-${{github.sha}}
1071
+ path : ./cpp/build/releases
1072
+ - name : Download Arrow Jars
1073
+ uses : actions/download-artifact@v4
1074
+ with :
1075
+ name : arrow-jars-centos-7-${{github.sha}}
1076
+ path : /root/.m2/repository/org/apache/arrow/
1077
+ - name : Prepare Spark Resources for Spark 3.5.5
1078
+ run : |
1079
+ rm -rf /opt/shims/spark35
1080
+ bash .github/workflows/util/install_spark_resources.sh 3.5
1081
+ - name : Build and Run unit test for Spark 3.5.5 (slow tests)
1082
+ run : |
1083
+ cd $GITHUB_WORKSPACE/
1084
+ export JAVA_HOME=/usr/lib/jvm/java-17-openjdk
1085
+ $MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Pceleborn -Piceberg -Pdelta -Phudi -Pspark-ut \
1086
+ -DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/ ${EXTRA_FLAGS}" \
1087
+ -DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
1049
1088
- name : Upload test report
1050
1089
if : always()
1051
1090
uses : actions/upload-artifact@v4
@@ -1056,7 +1095,7 @@ jobs:
1056
1095
run-spark-test-spark35-ras :
1057
1096
needs : build-native-lib-centos-7
1058
1097
runs-on : ubuntu-20.04
1059
- container : apache/gluten:centos-8
1098
+ container : apache/gluten:centos-8-jdk17
1060
1099
steps :
1061
1100
- uses : actions/checkout@v2
1062
1101
- name : Download All Artifacts
@@ -1069,19 +1108,20 @@ jobs:
1069
1108
with :
1070
1109
name : arrow-jars-centos-7-${{github.sha}}
1071
1110
path : /root/.m2/repository/org/apache/arrow/
1072
- - name : Prepare
1111
+ - name : Prepare Python3.9 and PySpark 3.5.5
1073
1112
run : |
1074
1113
dnf module -y install python39 && \
1075
1114
alternatives --set python3 /usr/bin/python3.9 && \
1076
1115
pip3 install setuptools && \
1077
- pip3 install pyspark==3.5.2 cython && \
1116
+ pip3 install pyspark==3.5.5 cython && \
1078
1117
pip3 install pandas pyarrow
1079
- - name : Build and Run unit test for Spark 3.5.2 (other tests)
1118
+ - name : Build and Run unit test for Spark 3.5.5 (other tests)
1080
1119
run : |
1081
1120
cd $GITHUB_WORKSPACE/
1082
1121
export SPARK_SCALA_VERSION=2.12
1083
- $MVN_CMD clean test -Pspark-3.5 -Pbackends-velox -Pceleborn -Piceberg -Pdelta -Pspark-ut \
1084
- -DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/ -Dspark.gluten.ras.enabled=true" \
1122
+ export JAVA_HOME=/usr/lib/jvm/java-17-openjdk
1123
+ $MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Pceleborn -Piceberg -Pdelta -Pspark-ut \
1124
+ -DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/ -Dspark.gluten.ras.enabled=true ${EXTRA_FLAGS}" \
1085
1125
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags
1086
1126
- name : Upload test report
1087
1127
uses : actions/upload-artifact@v4
@@ -1092,7 +1132,7 @@ jobs:
1092
1132
run-spark-test-spark35-slow-ras :
1093
1133
needs : build-native-lib-centos-7
1094
1134
runs-on : ubuntu-20.04
1095
- container : apache/gluten:centos-8
1135
+ container : apache/gluten:centos-8-jdk17
1096
1136
steps :
1097
1137
- uses : actions/checkout@v2
1098
1138
- name : Download All Artifacts
@@ -1105,11 +1145,16 @@ jobs:
1105
1145
with :
1106
1146
name : arrow-jars-centos-7-${{github.sha}}
1107
1147
path : /root/.m2/repository/org/apache/arrow/
1108
- - name : Build and Run unit test for Spark 3.5.2 (slow tests)
1148
+ - name : Prepare Spark Resources for Spark 3.5.5
1149
+ run : |
1150
+ rm -rf /opt/shims/spark35
1151
+ bash .github/workflows/util/install_spark_resources.sh 3.5
1152
+ - name : Build and Run unit test for Spark 3.5.5 (slow tests)
1109
1153
run : |
1110
1154
cd $GITHUB_WORKSPACE/
1111
- $MVN_CMD clean test -Pspark-3.5 -Pbackends-velox -Pceleborn -Piceberg -Pdelta -Pspark-ut \
1112
- -DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/ -Dspark.gluten.ras.enabled=true" \
1155
+ export JAVA_HOME=/usr/lib/jvm/java-17-openjdk
1156
+ $MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Pceleborn -Piceberg -Pdelta -Pspark-ut \
1157
+ -DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/ -Dspark.gluten.ras.enabled=true ${EXTRA_FLAGS}" \
1113
1158
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
1114
1159
- name : Upload test report
1115
1160
uses : actions/upload-artifact@v4
@@ -1120,7 +1165,7 @@ jobs:
1120
1165
run-spark-test-spark35-smj :
1121
1166
needs : build-native-lib-centos-7
1122
1167
runs-on : ubuntu-20.04
1123
- container : apache/gluten:centos-8
1168
+ container : apache/gluten:centos-8-jdk17
1124
1169
steps :
1125
1170
- uses : actions/checkout@v2
1126
1171
- name : Download All Artifacts
@@ -1133,19 +1178,20 @@ jobs:
1133
1178
with :
1134
1179
name : arrow-jars-centos-7-${{github.sha}}
1135
1180
path : /root/.m2/repository/org/apache/arrow/
1136
- - name : Prepare
1181
+ - name : Prepare Python3.9 and PySpark 3.5.5
1137
1182
run : |
1138
1183
dnf module -y install python39 && \
1139
1184
alternatives --set python3 /usr/bin/python3.9 && \
1140
1185
pip3 install setuptools && \
1141
- pip3 install pyspark==3.5.2 cython && \
1186
+ pip3 install pyspark==3.5.5 cython && \
1142
1187
pip3 install pandas pyarrow
1143
- - name : Build and Run unit test for Spark 3.5.2 (other tests)
1188
+ - name : Build and Run unit test for Spark 3.5.5 (other tests)
1144
1189
run : |
1145
1190
cd $GITHUB_WORKSPACE/
1146
1191
export SPARK_SCALA_VERSION=2.12
1147
- $MVN_CMD clean test -Pspark-3.5 -Pbackends-velox -Pceleborn -Piceberg -Pdelta -Pspark-ut \
1148
- -DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/ -Dspark.gluten.sql.columnar.forceShuffledHashJoin=false" \
1192
+ export JAVA_HOME=/usr/lib/jvm/java-17-openjdk
1193
+ $MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Pceleborn -Piceberg -Pdelta -Pspark-ut \
1194
+ -DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/ -Dspark.gluten.sql.columnar.forceShuffledHashJoin=false ${EXTRA_FLAGS}" \
1149
1195
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags
1150
1196
- name : Upload test report
1151
1197
uses : actions/upload-artifact@v4
@@ -1156,7 +1202,7 @@ jobs:
1156
1202
run-spark-test-spark35-slow-smj :
1157
1203
needs : build-native-lib-centos-7
1158
1204
runs-on : ubuntu-20.04
1159
- container : apache/gluten:centos-8
1205
+ container : apache/gluten:centos-8-jdk17
1160
1206
steps :
1161
1207
- uses : actions/checkout@v2
1162
1208
- name : Download All Artifacts
@@ -1169,11 +1215,16 @@ jobs:
1169
1215
with :
1170
1216
name : arrow-jars-centos-7-${{github.sha}}
1171
1217
path : /root/.m2/repository/org/apache/arrow/
1172
- - name : Build and Run unit test for Spark 3.5.2 (slow tests)
1218
+ - name : Prepare Spark Resources for Spark 3.5.5
1219
+ run : |
1220
+ rm -rf /opt/shims/spark35
1221
+ bash .github/workflows/util/install_spark_resources.sh 3.5
1222
+ - name : Build and Run unit test for Spark 3.5.5 (slow tests)
1173
1223
run : |
1174
1224
cd $GITHUB_WORKSPACE/
1175
- $MVN_CMD clean test -Pspark-3.5 -Pbackends-velox -Pceleborn -Piceberg -Pdelta -Pspark-ut \
1176
- -DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/ -Dspark.gluten.sql.columnar.forceShuffledHashJoin=false" \
1225
+ export JAVA_HOME=/usr/lib/jvm/java-17-openjdk
1226
+ $MVN_CMD clean test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Pceleborn -Piceberg -Pdelta -Pspark-ut \
1227
+ -DargLine="-Dspark.test.home=/opt/shims/spark35/spark_home/ -Dspark.gluten.sql.columnar.forceShuffledHashJoin=false ${EXTRA_FLAGS}" \
1177
1228
-DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
1178
1229
- name : Upload test report
1179
1230
uses : actions/upload-artifact@v4
@@ -1183,7 +1234,7 @@ jobs:
1183
1234
1184
1235
run-cpp-test-udf-test :
1185
1236
runs-on : ubuntu-20.04
1186
- container : apache/gluten:centos-8
1237
+ container : apache/gluten:centos-8-jdk17
1187
1238
steps :
1188
1239
- uses : actions/checkout@v2
1189
1240
- name : Get Ccache
@@ -1219,5 +1270,6 @@ jobs:
1219
1270
- name : Run UDF test
1220
1271
run : |
1221
1272
# Depends on --build_example=ON.
1222
- $MVN_CMD test -Pspark-3.5 -Pbackends-velox -Piceberg -Pdelta -DtagsToExclude=None \
1223
- -DtagsToInclude=org.apache.gluten.tags.UDFTest
1273
+ export JAVA_HOME=/usr/lib/jvm/java-17-openjdk
1274
+ $MVN_CMD test -Pspark-3.5 -Pjava-17 -Pbackends-velox -Piceberg -Pdelta -DtagsToExclude=None \
1275
+ -DtagsToInclude=org.apache.gluten.tags.UDFTest -DargLine="${EXTRA_FLAGS}"
0 commit comments