Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Spark35 #3

Merged
merged 524 commits into from
Oct 23, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
524 commits
Select commit Hold shift + click to select a range
f0752f2
[SPARK-47770][INFRA] Fix `GenerateMIMAIgnore.isPackagePrivateModule` …
dongjoon-hyun Apr 9, 2024
d424a4b
[SPARK-47774][INFRA][3.5] Remove redundant rules from `MimaExcludes`
dongjoon-hyun Apr 9, 2024
73e10b4
[SPARK-47182][BUILD] Exclude `commons-(io|lang3)` transitive dependen…
dongjoon-hyun Feb 27, 2024
7bec387
[SPARK-47083][BUILD] Upgrade `commons-codec` to 1.16.1
panbingkun Feb 19, 2024
2c43d92
[SPARK-47790][BUILD][3.5] Upgrade `commons-io` to 2.16.1
dongjoon-hyun Apr 10, 2024
8a77a01
[SPARK-47704][SQL] JSON parsing fails with "java.lang.ClassCastExcept…
sadikovi Apr 11, 2024
0818ce7
[MINOR][DOCS] Clarify relation between grouping API and `spark.sql.ex…
HyukjinKwon Apr 11, 2024
7d1e77c
[MINOR][DOCS] Make the link of spark properties with YARN more accurate
beliefer Apr 11, 2024
d18659d
[SPARK-47824][PS] Fix nondeterminism in pyspark.pandas.series.asof
markj-db Apr 12, 2024
b8e2498
[SPARK-47318][CORE][3.5] Adds HKDF round to AuthEngine key derivation…
dongjoon-hyun Apr 13, 2024
3b3903d
[SPARK-47828][CONNECT][PYTHON][3.5] DataFrameWriterV2.overwrite fails…
zhengruifeng Apr 16, 2024
6c67c61
[SPARK-47840][SS] Disable foldable propagation across Streaming Aggre…
sahnib Apr 16, 2024
d54f24c
[SPARK-46335][BUILD][3.5] Upgrade Maven to 3.9.6
dongjoon-hyun Apr 16, 2024
7aea21e
[SPARK-47895][SQL] group by all should be idempotent
cloud-fan Apr 18, 2024
e439e29
[SPARK-47463][SQL][3.5] Use V2Predicate to wrap expression with retur…
wForget Apr 18, 2024
afd99d1
[SPARK-47897][SQL][3.5] Fix ExpressionSet performance regression in s…
wForget Apr 19, 2024
6a358ff
[SPARK-47825][DSTREAMS][3.5] Make `KinesisTestUtils` & `WriteInputFor…
panbingkun Apr 19, 2024
d7c3794
[SPARK-47904][SQL][3.5] Preserve case in Avro schema when using enabl…
sadikovi Apr 22, 2024
cdd077a
[SPARK-47819][CONNECT][3.5] Use asynchronous callback for execution c…
xi-db Apr 24, 2024
ce19bfc
[SPARK-47633][SQL][3.5] Include right-side plan output in `LateralJoi…
bersprockets Apr 24, 2024
fdc0cee
[SPARK-47921][CONNECT] Fix ExecuteJobTag creation in ExecuteHolder
allisonwang-db Apr 25, 2024
7a573b9
[SPARK-48019] Fix incorrect behavior in ColumnVector/ColumnarArray wi…
gene-db Apr 28, 2024
33768f6
[SPARK-47927][SQL] Fix nullability attribute in UDF decoder
Apr 28, 2024
616c216
[SPARK-48034][TESTS] NullPointerException in MapStatusesSerDeserBench…
yaooqinn Apr 29, 2024
e78ee2c
[SPARK-48016][SQL] Fix a bug in try_divide function when with decimals
gengliangwang Apr 29, 2024
9bc2ab0
[SPARK-47129][CONNECT][SQL][3.5] Make ResolveRelations cache connect …
zhengruifeng Apr 30, 2024
953d7f9
[SPARK-48068][PYTHON] `mypy` should have `--python-executable` parameter
dongjoon-hyun May 1, 2024
fc0ef07
Revert "[SPARK-48016][SQL] Fix a bug in try_divide function when with…
dongjoon-hyun May 1, 2024
6a4475c
[SPARK-48016][SQL][3.5] Fix a bug in try_divide function when with de…
gengliangwang May 1, 2024
b6b8d61
[SPARK-45988][SPARK-45989][PYTHON] Fix typehints to handle `list` Gen…
dongjoon-hyun Nov 19, 2023
3d72063
[SPARK-48081] Fix ClassCastException in NTile.checkInputDataTypes() w…
JoshRosen May 2, 2024
d82403f
Revert "[SPARK-48081] Fix ClassCastException in NTile.checkInputDataT…
dongjoon-hyun May 2, 2024
9cd3125
[SPARK-48081][SQL][3.5] Fix ClassCastException in NTile.checkInputDat…
JoshRosen May 2, 2024
71cb930
[SPARK-48088][PYTHON][CONNECT][TESTS][3.5] Skip tests that fail in 3.…
HyukjinKwon May 3, 2024
2f2347f
[SPARK-48128][SQL] For BitwiseCount / bit_count expression, fix codeg…
JoshRosen May 4, 2024
45befc0
[SPARK-48019][SQL][FOLLOWUP] Use primitive arrays over object arrays …
gene-db May 5, 2024
e699a1e
[SPARK-48088][PYTHON][CONNECT][TESTS][FOLLOW-UP][3.5] Skips another t…
HyukjinKwon May 7, 2024
9971001
[SPARK-48083][SPARK-48084][ML][TESTS] Remove JIRA comments for reenab…
WeichenXu123 May 7, 2024
f92580a
[SPARK-48090][SS][PYTHON][TESTS] Shorten the traceback in the test ch…
HyukjinKwon May 7, 2024
ec7e888
[SPARK-48167][CONNECT][TESTS] Skip known behaviour change by SPARK-46122
HyukjinKwon May 7, 2024
1735d7d
[SPARK-48086][PYTHON][TESTS][3.5] Remove obsolete comment
HyukjinKwon May 7, 2024
03bc2b1
[SPARK-48167][PYTHON][TESTS][FOLLOWUP][3.5] Reformat test_readwriter.…
dongjoon-hyun May 7, 2024
a24ec1d
[SPARK-48179][INFRA][3.5] Pin `nbsphinx` to `0.9.3`
dongjoon-hyun May 7, 2024
2f8e7cb
[SPARK-48173][SQL][3.5] CheckAnalysis should see the entire query plan
cloud-fan May 7, 2024
15b5d2a
[SPARK-48178][INFRA][3.5] Run `build/scala-213/java-11-17` jobs of `b…
dongjoon-hyun May 7, 2024
704f956
[MINOR][PYTHON][TESTS] Remove the doc in error message tests to allow…
HyukjinKwon May 8, 2024
58b7130
[SPARK-48037][CORE][3.5] Fix SortShuffleWriter lacks shuffle write re…
cxzl25 May 8, 2024
36da89d
[SPARK-48138][CONNECT][TESTS] Disable a flaky `SparkSessionE2ESuite.i…
dongjoon-hyun May 6, 2024
ac4c5d6
[SPARK-48087][PYTHON][TESTS][3.5] Remove obsolete comment about UDTF …
HyukjinKwon May 8, 2024
a762f31
[SPARK-48184][PYTHON][CONNECT] Always set the seed of `Dataframe.samp…
zhengruifeng May 8, 2024
ff691fa
[SPARK-48116][INFRA][3.5] Run `pyspark-pandas*` only in PR builder an…
dongjoon-hyun May 8, 2024
81775a0
[SPARK-48116][INFRA][FOLLOWUP] Fix `if` statement to check repository
dongjoon-hyun May 4, 2024
26dccf0
[SPARK-48109][INFRA] Enable `k8s-integration-tests` only for `kuberne…
dongjoon-hyun May 3, 2024
9454607
[SPARK-48132][INFRA] Run `k8s-integration-tests` only in PR builder a…
dongjoon-hyun May 5, 2024
6dbbf08
[SPARK-48133][INFRA] Run `sparkr` only in PR builders and Daily CIs
dongjoon-hyun May 5, 2024
8277921
[SPARK-48192][INFRA] Enable TPC-DS tests in forked repository
HyukjinKwon May 8, 2024
541e1c4
[SPARK-48197][SQL] Avoid assert error for invalid lambda function
cloud-fan May 9, 2024
da4c808
[SPARK-48197][SQL][TESTS][FOLLOWUP][3.5] Regenerate golden files
dongjoon-hyun May 9, 2024
dc49117
[SPARK-48089][SS][CONNECT] Fix 3.5 <> 4.0 StreamingQueryListener comp…
WweiL May 9, 2024
c048653
[SPARK-47847][CORE] Deprecate `spark.network.remoteReadNioBufferConve…
pan3793 May 10, 2024
e9a1b42
[SPARK-48237][BUILD] Clean up `dev/pr-deps` at the end of `test-depen…
panbingkun May 11, 2024
ab511a7
[SPARK-48248][PYTHON] Fix nested array to respect legacy conf of infe…
HyukjinKwon May 13, 2024
19d12b2
[SPARK-48241][SQL][3.5] CSV parsing failure with char/varchar type co…
liujiayi771 May 14, 2024
34588a8
[SPARK-48265][SQL] Infer window group limit batch should do constant …
AngersZhuuuu May 14, 2024
172a23f
[SPARK-48267][SS] Regression e2e test with SPARK-47305
HeartSaVioR May 14, 2024
f37fa43
[SPARK-48172][SQL] Fix escaping issues in JDBC Dialects
mihailom-db May 14, 2024
74724d6
Revert "[SPARK-48172][SQL] Fix escaping issues in JDBC Dialects"
yaooqinn May 15, 2024
07e08c0
[SPARK-48105][SS][3.5] Fix the race condition between state store unl…
huanliwang-db May 15, 2024
210ed25
[SPARK-48172][SQL] Fix escaping issues in JDBCDialects
mihailom-db May 15, 2024
c1dd4a5
[SPARK-48297][SQL] Fix a regression TRANSFORM clause with char/varchar
yaooqinn May 16, 2024
1a45428
[SPARK-48294][SQL][3.5] Handle lowercase in nestedTypeMissingElementT…
michaelzhan-db May 17, 2024
0a45bec
[SPARK-48424][INFRA] Make dev/is-changed.py to return true it it fails
HyukjinKwon May 27, 2024
f42c029
[SPARK-41049][SQL][FOLLOW-UP] Mark map related expressions as statefu…
amaliujia May 28, 2024
7313d71
[SPARK-48273][SQL] Fix late rewrite of PlanWithUnresolvedIdentifier
nikolamand-db May 28, 2024
043944e
[SPARK-48273][SQL][FOLLOWUP] Explicitly create non-Hive table in iden…
cloud-fan May 29, 2024
c87b648
[SPARK-41049][SQL][FOLLOW-UP][3.5] stateful expressions test uses dif…
amaliujia May 30, 2024
dc8f652
[SPARK-48273][SQL][FOLLOWUP] Update golden file
cloud-fan May 30, 2024
d64f96c
[SPARK-48172][SQL][FOLLOWUP] Fix escaping issues in JDBCDialects
mihailom-db May 31, 2024
7d39000
[SPARK-48484][SQL] Fix: V2Write use the same TaskAttemptId for differ…
jackylee-ch May 31, 2024
744b070
[SPARK-48391][CORE] Using addAll instead of add function in fromAccum…
monkeyboy123 May 31, 2024
7e0c314
[SPARK-48481][SQL][SS] Do not apply OptimizeOneRowPlan against stream…
HeartSaVioR Jun 1, 2024
7f99f2c
[SPARK-48394][3.5][CORE] Cleanup mapIdToMapIndex on mapoutput unregister
Ngone51 Jun 3, 2024
d3a324d
[SPARK-48535][SS] Update config docs to indicate possibility of data …
anishshri-db Jun 5, 2024
a00c115
[SPARK-48286] Fix analysis of column with exists default expression -…
urosstan-db Jun 6, 2024
88582e1
[SPARK-48610][SQL] refactor: use auxiliary idMap instead of OP_ID_TAG
liuzqt Jun 17, 2024
15e6369
[SPARK-48642][CORE] False SparkOutOfMemoryError caused by killing tas…
pan3793 Jun 17, 2024
a3f4cc7
[SPARK-48286][SQL][3.5] Fix column default value check - Add error class
urosstan-db Jun 17, 2024
145c045
[SPARK-47910][CORE] close stream when DiskBlockObjectWriter closeReso…
JacobZheng0927 Jun 18, 2024
f57af38
[SPARK-47910][CORE][3.5][FOLLOWUP] Remove the use of MDC
JacobZheng0927 Jun 19, 2024
4741f7a
[SPARK-48498][SQL][3.5] Always do char padding in predicates
jackylee-ch Jun 24, 2024
0db5bde
[SPARK-48639][CONNECT][PYTHON] Add Origin to Relation.RelationCommon
HyukjinKwon Jun 24, 2024
405b23b
[SPARK-47172][CORE][3.5] Add support for AES-GCM for RPC encryption
sweisdb Jun 27, 2024
e6d5ddd
Revert "[SPARK-48639][CONNECT][PYTHON] Add Origin to Relation.Relatio…
HyukjinKwon Jun 27, 2024
6cee8e1
[SPARK-47927][SQL][FOLLOWUP] fix ScalaUDF output nullability
cloud-fan Jun 27, 2024
d35af54
[SPARK-48428][SQL] Fix IllegalStateException in NestedColumnAliasing
Jun 27, 2024
789ac5b
[SPARK-46957][CORE] Decommission migrated shuffle files should be abl…
Ngone51 Jun 27, 2024
7aa12b6
Revert "[SPARK-46957][CORE] Decommission migrated shuffle files shoul…
yaooqinn Jun 27, 2024
b28ddb1
[SPARK-46957][CORE][3.5][3.4] Decommission migrated shuffle files sho…
Ngone51 Jun 27, 2024
3d7a20a
[SPARK-48709][SQL][3.5] Fix varchar type resolution mismatch for Data…
wangyum Jun 27, 2024
ade9dbf
[SPARK-48494][BUILD][3.5] Update `airlift:aircompressor` to 0.27
bjornjorgensen Jun 28, 2024
686f59c
[SPARK-48586][SS][3.5] Remove lock acquisition in doMaintenance() by …
riyaverm-db Jun 28, 2024
df70cc1
[SPARK-48292][CORE][3.5] Revert [SPARK-39195][SQL] Spark OutputCommit…
AngersZhuuuu Jul 1, 2024
ef4e456
[SPARK-48759][SQL] Add migration doc for CREATE TABLE AS SELECT behav…
asl3 Jul 2, 2024
44eba46
[SPARK-48710][PYTHON][3.5] Limit NumPy version to supported range (>=…
codesorcery Jul 3, 2024
1cc0043
[SPARK-48806][SQL] Pass actual exception when url_decode fails
wForget Jul 4, 2024
67047cd
[SPARK-48719][SQL][3.5] Fix the calculation bug of RegrSlope & RegrIn…
wayneguow Jul 8, 2024
1e15e3f
[SPARK-48843] Prevent infinite loop with BindParameters
nemanja-boric-databricks Jul 10, 2024
8afce30
[SPARK-48863][SQL] Fix ClassCastException when parsing JSON with "spa…
sadikovi Jul 11, 2024
2bb46e0
[MINOR][DOCS] Add `docs/_generated/` to .gitignore
yaooqinn Apr 23, 2024
d517a63
[MINOR][SQL][TESTS] Remove a duplicate test case in `CSVExprUtilsSuite`
wayneguow Jul 11, 2024
b15a872
[SPARK-48871] Fix INVALID_NON_DETERMINISTIC_EXPRESSIONS validation in…
zhipengmao-db Jul 12, 2024
596f680
[SPARK-48845][SQL] GenericUDF catch exceptions from children
jackylee-ch Jul 12, 2024
56dec39
[SPARK-48666][SQL] Do not push down filter if it contains PythonUDFs
HyukjinKwon Jul 12, 2024
4d1bbfd
[SPARK-47172][DOCS][FOLLOWUP] Fix spark.network.crypto.ciphersince ve…
yaooqinn Jul 16, 2024
f1f5bb6
[SPARK-47307][SQL][3.5] Add a config to optionally chunk base64 strings
wForget Jul 16, 2024
033f6a3
[SPARK-47307][DOCS][FOLLOWUP] Add a migration guide for the behavior …
wForget Jul 17, 2024
ef1f377
[SPARK-48889][SS] testStream to unload state stores before finishing
siying Jul 17, 2024
443825a
[SPARK-48920][BUILD][3.5] Upgrade ORC to 1.9.4
williamhyun Jul 17, 2024
44f8766
[SPARK-48930][CORE] Redact `awsAccessKeyId` by including `accesskey` …
dongjoon-hyun Jul 17, 2024
b151012
Preparing Spark release v3.5.2-rc1
yaooqinn Jul 18, 2024
d556740
Preparing development version 3.5.3-SNAPSHOT
yaooqinn Jul 18, 2024
f07a547
[SPARK-48791][CORE][3.5] Fix perf regression caused by the accumulato…
Ngone51 Jul 18, 2024
e247915
[SPARK-48921][SQL][3.5] ScalaUDF encoders in subquery should be resol…
viirya Jul 18, 2024
66dce6d
[SPARK-48934][SS] Python datetime types converted incorrectly for set…
siying Jul 18, 2024
5a09a7d
Revert "[SPARK-47307][DOCS][FOLLOWUP] Add a migration guide for the b…
yaooqinn Jul 19, 2024
28d33e3
[SPARK-48498][SQL][FOLLOWUP] do padding for char-char comparison
cloud-fan Jul 19, 2024
ea53ea7
[SPARK-47307][SQL][FOLLOWUP][3.5] Promote spark.sql.legacy.chunkBase6…
wForget Jul 20, 2024
6d8f511
Preparing Spark release v3.5.2-rc2
yaooqinn Jul 22, 2024
9574cec
Preparing development version 3.5.3-SNAPSHOT
yaooqinn Jul 22, 2024
ada5dc2
[SPARK-48991][SQL] Move path initialization into try-catch block in F…
yaooqinn Jul 24, 2024
66b7cb9
[SPARK-48089][SS][CONNECT][FOLLOWUP][3.5] Disable Server Listener fai…
WweiL Jul 25, 2024
c4ef321
[SPARK-48308][CORE][3.5] Unify getting data schema without partition …
johanl-db Jul 25, 2024
ebda6a6
Preparing Spark release v3.5.2-rc3
yaooqinn Jul 25, 2024
61a90e5
Preparing development version 3.5.3-SNAPSHOT
yaooqinn Jul 25, 2024
4c38e90
[SPARK-49005][K8S][3.5] Use `17-jammy` tag instead of `17` to prevent…
dongjoon-hyun Jul 25, 2024
1edbddf
Preparing Spark release v3.5.2-rc4
yaooqinn Jul 26, 2024
75029e1
Preparing development version 3.5.3-SNAPSHOT
yaooqinn Jul 26, 2024
268c13e
[SPARK-49054][SQL][3.5] Column default value should support current_*…
gengliangwang Jul 30, 2024
c6df890
[SPARK-49000][SQL] Fix "select count(distinct 1) from t" where t is e…
uros-db Jul 31, 2024
a0f88f5
[SPARK-44638][SQL][TESTS][3.5] Add test for Char/Varchar in JDBC cust…
yaooqinn Jul 31, 2024
d23b70b
[SPARK-49066][SQL][TESTS] Refactor `OrcEncryptionSuite` and make `spa…
LuciferYang Jul 31, 2024
36f9a4b
Revert "[SPARK-49066][SQL][TESTS] Refactor `OrcEncryptionSuite` and m…
dongjoon-hyun Jul 31, 2024
94558f6
Revert "[SPARK-49000][SQL] Fix "select count(distinct 1) from t" wher…
yaooqinn Aug 1, 2024
a1e7fb1
[SPARK-49065][SQL] Rebasing in legacy formatters/parsers must support…
sumeet-db Aug 1, 2024
4f9dbc3
[SPARK-49066][SQL][TESTS][3.5] Refactor `OrcEncryptionSuite` and make…
LuciferYang Aug 1, 2024
0008bd1
[SPARK-49000][SQL][3.5] Fix "select count(distinct 1) from t" where t…
uros-db Aug 2, 2024
98eaaa5
[SPARK-49094][SQL] Fix ignoreCorruptFiles non-functioning for hive or…
yaooqinn Aug 3, 2024
b33a3ee
[SPARK-48791][CORE][FOLLOW-UP][3.5] Fix regression caused by immutabl…
Ngone51 Aug 5, 2024
f2e2601
[SPARK-49099][SQL] CatalogManager.setCurrentNamespace should respect …
amaliujia Aug 6, 2024
d13808c
[SPARK-49099][SQL][FOLLOWUP][3.5] recover tests in DDLSuite
cloud-fan Aug 6, 2024
bb7846d
Preparing Spark release v3.5.2-rc5
yaooqinn Aug 6, 2024
52ca284
Preparing development version 3.5.3-SNAPSHOT
yaooqinn Aug 6, 2024
0e2d757
[SPARK-49182][DOCS][PYTHON] Stop publish site/docs/{version}/api/pyth…
yaooqinn Aug 9, 2024
7bfb4f0
[SPARK-49176][K8S][DOCS] Fix `spark.ui.custom.executor.log.url` docs …
dongjoon-hyun Aug 9, 2024
204dd81
[SPARK-49183][SQL] V2SessionCatalog.createTable should respect PROP_I…
cloud-fan Aug 12, 2024
4a9dae9
[SPARK-48204][INFRA][FOLLOW] fix release scripts for the "finalize" step
cloud-fan Jun 3, 2024
deac780
[SPARK-46632][SQL] Fix subexpression elimination when equivalent tern…
zml1206 Aug 12, 2024
4d69021
[SPARK-49200][SQL] Fix null type non-codegen ordering exception
ulysses-you Aug 12, 2024
6a75640
[SPARK-49193][SQL] Improve the performance of RowSetUtils.toColumnBas…
wangyum Aug 11, 2024
7799811
[SPARK-49183][SQL][FOLLOWUP][3.5] Fix compilation
cloud-fan Aug 12, 2024
3237b8e
[SPARK-49197][CORE] Redact `Spark Command` output in `launcher` module
dongjoon-hyun Aug 12, 2024
d824219
[SPARK-49152][SQL] V2SessionCatalog should use V2Command
amaliujia Aug 13, 2024
6a250e7
[SPARK-49179][SQL] Fix v2 multi bucketed inner joins throw AssertionE…
ulysses-you Aug 13, 2024
6859ef9
Revert "[SPARK-49179][SQL] Fix v2 multi bucketed inner joins throw As…
dongjoon-hyun Aug 13, 2024
5463bfc
[SPARK-49179][SQL][3.5] Fix v2 multi bucketed inner joins throw Asser…
ulysses-you Aug 13, 2024
bd2cbd6
[SPARK-49038][SQL][3.5] SQLMetric should report the raw value in the …
cloud-fan Aug 14, 2024
8d05bf2
[SPARK-49152][SQL][FOLLOWUP][3.5] table location string should be Had…
cloud-fan Aug 15, 2024
3d96901
[SPARK-45590][BUILD][3.5] Upgrade okio to 1.17.6 from 1.15.0
roczei Aug 16, 2024
ac2219e
[SPARK-49039][UI] Reset checkbox when executor metrics are loaded in …
cxzl25 Aug 16, 2024
3148cfa
[SPARK-49211][SQL][3.5] V2 Catalog can also support built-in data sou…
amaliujia Aug 16, 2024
c06906d
[SPARK-49017][SQL][3.5] Insert statement fails when multiple paramete…
mihailom-db Aug 20, 2024
027a14b
[SPARK-49246][SQL] TableCatalog#loadTable should indicate if it's for…
cloud-fan Aug 21, 2024
481bc58
[SPARK-46444][SQL] V2SessionCatalog#createTable should not load the t…
cloud-fan Dec 26, 2023
eaadb39
[SPARK-49300][CORE][3.5] Fix Hadoop delegation token leak when tokenR…
zhangshuyan0 Aug 22, 2024
b2a25be
[SPARK-49333][SQL] Shutdown timeout thread while cleaning up SparkExe…
wangyum Aug 23, 2024
f4bcefb
[SPARK-49352][SQL][3.5] Avoid redundant array transform for identical…
viirya Aug 24, 2024
5eca953
[SPARK-49359][SQL] Allow StagedTableCatalog implementations to fall b…
cloud-fan Aug 23, 2024
870f1d8
[SPARK-49381][K8S][TESTS] Fix `SPARK-41388` test case to assert the r…
dongjoon-hyun Aug 26, 2024
2bc4c82
[SPARK-49385][K8S] Fix `getReusablePVCs` to use `podCreationTimeout` …
dongjoon-hyun Aug 26, 2024
c6f4dd7
[SPARK-49359][TESTS][FOLLOWUP][3.5] Fix compilation to rename `util.M…
dongjoon-hyun Aug 26, 2024
c979127
[SPARK-49402][PYTHON] Fix Binder integration in PySpark documentation
HyukjinKwon Aug 27, 2024
ddc99bc
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
2192437
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
ff83dd9
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
315d987
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
39f760c
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
51bfb88
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
e5a5d19
[SPARK-49402][PYTHON][FOLLOW-UP] Fix Binder integration in PySpark do…
HyukjinKwon Aug 27, 2024
93fedc5
[SPARK-49402][PYTHON][FOLLOW-UP] Set upperfound for NumPy
HyukjinKwon Aug 27, 2024
dcfefd0
[SPARK-49402][PYTHON][FOLLOW-UP] Manually load ~/.profile in Spark Co…
HyukjinKwon Aug 28, 2024
2ad11b6
[SPARK-46037][SQL] Correctness fix for Shuffled Hash Join build left …
cloud-fan Aug 29, 2024
dce7c9a
[SPARK-43242][CORE][3.5] Fix throw 'Unexpected type of BlockId' in sh…
CavemanIV Aug 30, 2024
30a75d1
[MINOR] Add `artifacts` to `.gitignore`
zhengruifeng Aug 30, 2024
d5caaaa
[SPARK-49480][CORE] Fix NullPointerException from `SparkThrowableHelp…
jshmchenxi Sep 2, 2024
38ad0e7
[SPARK-49476][SQL] Fix nullability of base64 function
Kimahriman Sep 2, 2024
a1cd99d
Preparing Spark release v3.5.3-rc1
itholic Sep 3, 2024
0616ef2
Preparing development version 3.5.4-SNAPSHOT
itholic Sep 3, 2024
d8adf4b
[SPARK-49476][SQL][3.5][FOLLOWUP] Fix base64 proto test
Kimahriman Sep 3, 2024
7febde1
[SPARK-49275][SQL][3.5] Fix return type nullness of the xpath expression
chenhao-db Sep 4, 2024
7718777
[MINOR][DOCS] Fix site.SPARK_VERSION pattern in RDD Programming Guide
yaooqinn Sep 4, 2024
d83bf8c
[SPARK-49509][CORE] Use `Platform.allocateDirectBuffer` instead of `B…
cxzl25 Sep 4, 2024
e5ec16e
[SPARK-48965][SQL] Use the correct schema in `Dataset#toJSON`
bersprockets Sep 5, 2024
8fbeaf5
[SPARK-49408][SQL] Use IndexedSeq in ProjectingInternalRow
wzx140 Sep 5, 2024
d8f9d8d
[SPARK-49152][SQL][FOLLOWUP] DelegatingCatalogExtension should also u…
cloud-fan Sep 5, 2024
1f8c71f
Preparing Spark release v3.5.3-rc2
itholic Sep 6, 2024
46214da
Preparing development version 3.5.4-SNAPSHOT
itholic Sep 6, 2024
3f22ef1
[SPARK-49246][SQL][FOLLOW-UP] The behavior of SaveAsTable should not …
cloud-fan Sep 9, 2024
6292cfc
Preparing Spark release v3.5.3-rc3
itholic Sep 9, 2024
e923790
Preparing development version 3.5.4-SNAPSHOT
itholic Sep 9, 2024
ba374c6
fix import
cloud-fan Sep 9, 2024
32232e9
Preparing Spark release v3.5.3-rc3
itholic Sep 9, 2024
67421bb
Preparing development version 3.5.4-SNAPSHOT
itholic Sep 9, 2024
1206b52
[SPARK-49501][SQL] Fix double-escaping of table location
cstavr Sep 9, 2024
96eebeb
[SPARK-49595][CONNECT][SQL] Fix `DataFrame.unpivot/melt` in Spark Con…
xinrong-meng Sep 11, 2024
560efed
[SPARK-49261][SQL] Don't replace literals in aggregate expressions wi…
bersprockets Sep 12, 2024
e693e18
[SPARK-49628][SQL] ConstantFolding should copy stateful expression be…
cloud-fan Sep 13, 2024
e7ca790
[SPARK-49699][SS] Disable PruneFilters for streaming workloads
n-young-db Sep 20, 2024
dd76a82
[SPARK-49760][YARN] Correct handling of `SPARK_USER` env variable ove…
cnauroth Sep 24, 2024
b513297
[SPARK-49750][DOC] Mention delegation token support in K8s mode
pan3793 Sep 24, 2024
f1c69a5
[SPARK-49791][SQL] Make DelegatingCatalogExtension more extendable
cloud-fan Sep 26, 2024
b51db8b
[SPARK-49791][SQL][FOLLOWUP][3.5] Fix `import` statement
dongjoon-hyun Sep 26, 2024
1040657
[SPARK-49211][SQL][FOLLOW-UP] Support catalog in QualifiedTableName
amaliujia Sep 27, 2024
50c1783
[SPARK-49803][SQL][TESTS] Increase `spark.test.docker.connectionTimeo…
dongjoon-hyun Sep 27, 2024
be254c5
[SPARK-49804][K8S] Fix to use the exit code of executor container always
Sep 27, 2024
f888d57
[SPARK-46525][SQL][TESTS][3.5] Fix `docker-integration-tests` on Appl…
yaooqinn Sep 27, 2024
a49d6f4
[SPARK-47949][SQL][DOCKER][TESTS] MsSQLServer: Bump up mssql docker i…
yaooqinn Apr 23, 2024
16b8153
[SPARK-49819] Disable CollapseProject for correlated subqueries in pr…
n-young-db Sep 30, 2024
75860a3
[SPARK-49743][SQL] OptimizeCsvJsonExpr should not change schema field…
nikhilsheoran-db Oct 1, 2024
d9467b5
[SPARK-49816][SQL][3.5] Should only update out-going-ref-count for re…
cloud-fan Oct 1, 2024
b8f8456
[SPARK-49841][PYTHON][TESTS][3.5] Skip PySpark `test_cast_to_udt_with…
dongjoon-hyun Oct 1, 2024
ec28154
[SPARK-47702][CORE] Remove Shuffle service endpoint from the location…
Oct 4, 2024
aa6784c
Revert "[SPARK-47702][CORE] Remove Shuffle service endpoint from the …
LuciferYang Oct 4, 2024
d4b34d2
[SPARK-49843][SQL][3.5] Fix change comment on char/varchar columns
stefankandic Oct 4, 2024
9a5ae45
[SPARK-49836][SQL][SS] Fix possibly broken query when window is provi…
HeartSaVioR Oct 4, 2024
ca8407b
[SPARK-47702][CORE][3.5] Remove Shuffle service endpoint from the loc…
Oct 5, 2024
4d93a30
[SPARK-49806][PYTHON][TESTS][FOLLOW-UP] Skip newline difference in Sp…
HyukjinKwon Oct 7, 2024
4472fb2
[SPARK-49909][SQL][3.5] Fix the pretty name of some expressions
panbingkun Oct 9, 2024
5996c48
[SPARK-49782][SQL] ResolveDataFrameDropColumns rule resolves Unresolv…
LantaoJin Oct 10, 2024
36299b3
[SPARK-49959][SQL] Fix ColumnarArray.copy() to read nulls from the co…
chenhao-db Oct 15, 2024
92ec503
[SPARK-49981][CORE][TESTS] Fix `AsyncRDDActionsSuite.FutureAction res…
dongjoon-hyun Oct 16, 2024
d37a8b9
[SPARK-49983][CORE][TESTS] Fix `BarrierTaskContextSuite.successively …
dongjoon-hyun Oct 16, 2024
f5bc928
[SPARK-49979][SQL] Fix AQE hanging issue when collecting twice on a f…
liuzqt Oct 17, 2024
3d2d1c1
[SPARK-48155][SQL][3.5] AQEPropagateEmptyRelation for join should che…
AngersZhuuuu Oct 17, 2024
1fae1d7
[SPARK-49829][SS] Fix the bug on the optimization on adding input to …
HeartSaVioR Oct 18, 2024
6b9b3c0
Revert "[SPARK-49909][SQL][3.5] Fix the pretty name of some expressions"
panbingkun Oct 18, 2024
3a4ebae
[SPARK-50021][CORE][UI][3.5] Fix `ApplicationPage` to hide App UI lin…
dongjoon-hyun Oct 18, 2024
d24393b
[SPARK-50022][CORE][UI][3.5] Fix `MasterPage` to hide App UI links wh…
dongjoon-hyun Oct 18, 2024
d15c318
Merge branch 'branch-3.5' into spark-3.5
ejblanco Oct 23, 2024
18e355d
fix utests
ejblanco Oct 23, 2024
9ac264c
fix
ejblanco Oct 23, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
61 changes: 43 additions & 18 deletions .github/workflows/build_and_test.yml
Original file line number Diff line number Diff line change
Expand Up @@ -79,25 +79,34 @@ jobs:
id: set-outputs
run: |
if [ -z "${{ inputs.jobs }}" ]; then
pyspark=true; sparkr=true; tpcds=true; docker=true;
pyspark_modules=`cd dev && python -c "import sparktestsupport.modules as m; print(','.join(m.name for m in m.all_modules if m.name.startswith('pyspark')))"`
pyspark=`./dev/is-changed.py -m $pyspark_modules`
sparkr=`./dev/is-changed.py -m sparkr`
tpcds=`./dev/is-changed.py -m sql`
docker=`./dev/is-changed.py -m docker-integration-tests`
# 'build', 'scala-213', and 'java-11-17' are always true for now.
# It does not save significant time and most of PRs trigger the build.
if [[ "${{ github.repository }}" != 'apache/spark' ]]; then
pandas=$pyspark
kubernetes=`./dev/is-changed.py -m kubernetes`
sparkr=`./dev/is-changed.py -m sparkr`
tpcds=`./dev/is-changed.py -m sql`
docker=`./dev/is-changed.py -m docker-integration-tests`
else
pandas=false
kubernetes=false
sparkr=false
tpcds=false
docker=false
fi
build=`./dev/is-changed.py -m "core,unsafe,kvstore,avro,utils,network-common,network-shuffle,repl,launcher,examples,sketch,graphx,catalyst,hive-thriftserver,streaming,sql-kafka-0-10,streaming-kafka-0-10,mllib-local,mllib,yarn,mesos,kubernetes,hadoop-cloud,spark-ganglia-lgpl,sql,hive"`
precondition="
{
\"build\": \"true\",
\"build\": \"$build\",
\"pyspark\": \"$pyspark\",
\"pyspark-pandas\": \"$pandas\",
\"sparkr\": \"$sparkr\",
\"tpcds-1g\": \"$tpcds\",
\"docker-integration-tests\": \"$docker\",
\"scala-213\": \"true\",
\"java-11-17\": \"true\",
\"scala-213\": \"$build\",
\"java-11-17\": \"$build\",
\"lint\" : \"true\",
\"k8s-integration-tests\" : \"true\",
\"k8s-integration-tests\" : \"$kubernetes\",
\"breaking-changes-buf\" : \"true\",
}"
echo $precondition # For debugging
Expand Down Expand Up @@ -204,6 +213,8 @@ jobs:
HIVE_PROFILE: ${{ matrix.hive }}
GITHUB_PREV_SHA: ${{ github.event.before }}
SPARK_LOCAL_IP: localhost
SKIP_UNIDOC: true
SKIP_MIMA: true
SKIP_PACKAGING: true
steps:
- name: Checkout Spark repository
Expand Down Expand Up @@ -256,7 +267,7 @@ jobs:
- name: Install Python packages (Python 3.8)
if: (contains(matrix.modules, 'sql') && !contains(matrix.modules, 'sql-'))
run: |
python3.8 -m pip install 'numpy>=1.20.0' pyarrow pandas scipy unittest-xml-reporting 'grpcio==1.56.0' 'protobuf==3.20.3'
python3.8 -m pip install 'numpy>=1.20.0' 'pyarrow==12.0.1' pandas scipy unittest-xml-reporting 'grpcio==1.56.0' 'protobuf==3.20.3'
python3.8 -m pip list
# Run the tests.
- name: Run tests
Expand Down Expand Up @@ -360,6 +371,14 @@ jobs:
pyspark-pandas-connect
- >-
pyspark-pandas-slow-connect
exclude:
# Always run if pyspark-pandas == 'true', even infra-image is skip (such as non-master job)
# In practice, the build will run in individual PR, but not against the individual commit
# in Apache Spark repository.
- modules: ${{ fromJson(needs.precondition.outputs.required).pyspark-pandas != 'true' && 'pyspark-pandas' }}
- modules: ${{ fromJson(needs.precondition.outputs.required).pyspark-pandas != 'true' && 'pyspark-pandas-slow' }}
- modules: ${{ fromJson(needs.precondition.outputs.required).pyspark-pandas != 'true' && 'pyspark-pandas-connect' }}
- modules: ${{ fromJson(needs.precondition.outputs.required).pyspark-pandas != 'true' && 'pyspark-pandas-slow-connect' }}
env:
MODULES_TO_TEST: ${{ matrix.modules }}
HADOOP_PROFILE: ${{ inputs.hadoop }}
Expand Down Expand Up @@ -407,6 +426,8 @@ jobs:
key: pyspark-coursier-${{ hashFiles('**/pom.xml', '**/plugins.sbt') }}
restore-keys: |
pyspark-coursier-
- name: Free up disk space
run: ./dev/free_disk_space_container
- name: Install Java ${{ matrix.java }}
uses: actions/setup-java@v3
with:
Expand Down Expand Up @@ -504,6 +525,8 @@ jobs:
key: sparkr-coursier-${{ hashFiles('**/pom.xml', '**/plugins.sbt') }}
restore-keys: |
sparkr-coursier-
- name: Free up disk space
run: ./dev/free_disk_space_container
- name: Install Java ${{ inputs.java }}
uses: actions/setup-java@v3
with:
Expand Down Expand Up @@ -612,6 +635,8 @@ jobs:
key: docs-maven-${{ hashFiles('**/pom.xml') }}
restore-keys: |
docs-maven-
- name: Free up disk space
run: ./dev/free_disk_space_container
- name: Install Java 8
uses: actions/setup-java@v3
with:
Expand All @@ -621,6 +646,8 @@ jobs:
run: ./dev/check-license
- name: Dependencies test
run: ./dev/test-dependencies.sh
- name: MIMA test
run: ./dev/mima
- name: Scala linter
run: ./dev/lint-scala
- name: Java linter
Expand Down Expand Up @@ -672,16 +699,16 @@ jobs:
# See also https://issues.apache.org/jira/browse/SPARK-35375.
# Pin the MarkupSafe to 2.0.1 to resolve the CI error.
# See also https://issues.apache.org/jira/browse/SPARK-38279.
python3.9 -m pip install 'sphinx<3.1.0' mkdocs pydata_sphinx_theme nbsphinx numpydoc 'jinja2<3.0.0' 'markupsafe==2.0.1' 'pyzmq<24.0.0'
python3.9 -m pip install 'sphinx<3.1.0' mkdocs pydata_sphinx_theme 'sphinx-copybutton==0.5.2' 'nbsphinx==0.9.3' numpydoc 'jinja2<3.0.0' 'markupsafe==2.0.1' 'pyzmq<24.0.0' 'sphinxcontrib-applehelp==1.0.4' 'sphinxcontrib-devhelp==1.0.2' 'sphinxcontrib-htmlhelp==2.0.1' 'sphinxcontrib-qthelp==1.0.3' 'sphinxcontrib-serializinghtml==1.1.5' 'nest-asyncio==1.5.8' 'rpds-py==0.16.2' 'alabaster==0.7.13'
python3.9 -m pip install ipython_genutils # See SPARK-38517
python3.9 -m pip install sphinx_plotly_directive 'numpy>=1.20.0' pyarrow pandas 'plotly>=4.8'
python3.9 -m pip install sphinx_plotly_directive 'numpy>=1.20.0' 'pyarrow==12.0.1' pandas 'plotly>=4.8'
python3.9 -m pip install 'docutils<0.18.0' # See SPARK-39421
apt-get update -y
apt-get install -y ruby ruby-dev
Rscript -e "install.packages(c('devtools', 'testthat', 'knitr', 'rmarkdown', 'markdown', 'e1071', 'roxygen2', 'ggplot2', 'mvtnorm', 'statmod'), repos='https://cloud.r-project.org/')"
Rscript -e "devtools::install_version('pkgdown', version='2.0.1', repos='https://cloud.r-project.org')"
Rscript -e "devtools::install_version('preferably', version='0.4', repos='https://cloud.r-project.org')"
gem install bundler
gem install bundler -v 2.4.22
cd docs
bundle install
- name: R linter
Expand Down Expand Up @@ -1010,9 +1037,7 @@ jobs:
- name: start minikube
run: |
# See more in "Installation" https://minikube.sigs.k8s.io/docs/start/
# curl -LO https://storage.googleapis.com/minikube/releases/latest/minikube-linux-amd64
# TODO(SPARK-44495): Resume to use the latest minikube for k8s-integration-tests.
curl -LO https://storage.googleapis.com/minikube/releases/v1.30.1/minikube-linux-amd64
curl -LO https://storage.googleapis.com/minikube/releases/latest/minikube-linux-amd64
sudo install minikube-linux-amd64 /usr/local/bin/minikube
# Github Action limit cpu:2, memory: 6947MB, limit to 2U6G for better resource statistic
minikube start --cpus 2 --memory 6144
Expand All @@ -1030,7 +1055,7 @@ jobs:
kubectl create clusterrolebinding serviceaccounts-cluster-admin --clusterrole=cluster-admin --group=system:serviceaccounts || true
kubectl apply -f https://raw.githubusercontent.com/volcano-sh/volcano/v1.7.0/installer/volcano-development.yaml || true
eval $(minikube docker-env)
build/sbt -Psparkr -Pkubernetes -Pvolcano -Pkubernetes-integration-tests -Dspark.kubernetes.test.driverRequestCores=0.5 -Dspark.kubernetes.test.executorRequestCores=0.2 -Dspark.kubernetes.test.volcanoMaxConcurrencyJobNum=1 -Dtest.exclude.tags=local "kubernetes-integration-tests/test"
build/sbt -Psparkr -Pkubernetes -Pvolcano -Pkubernetes-integration-tests -Dspark.kubernetes.test.volcanoMaxConcurrencyJobNum=1 -Dtest.exclude.tags=local "kubernetes-integration-tests/test"
- name: Upload Spark on K8S integration tests log files
if: failure()
uses: actions/upload-artifact@v3
Expand Down
2 changes: 2 additions & 0 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -26,6 +26,7 @@
.scala_dependencies
.settings
.vscode
artifacts/
/lib/
R-unit-tests.log
R/unit-tests.out
Expand All @@ -50,6 +51,7 @@ dev/create-release/*final
dev/create-release/*txt
dev/pr-deps/
dist/
docs/_generated/
docs/_site/
docs/api
docs/.local_ruby_bundle
Expand Down
5 changes: 0 additions & 5 deletions LICENSE
Original file line number Diff line number Diff line change
Expand Up @@ -218,11 +218,6 @@ docs/js/vendor/bootstrap.js
connector/spark-ganglia-lgpl/src/main/java/com/codahale/metrics/ganglia/GangliaReporter.java


Python Software Foundation License
----------------------------------

python/docs/source/_static/copybutton.js

BSD 3-Clause
------------

Expand Down
2 changes: 1 addition & 1 deletion R/pkg/DESCRIPTION
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
Package: SparkR
Type: Package
Version: 3.5.0
Version: 3.5.4
Title: R Front End for 'Apache Spark'
Description: Provides an R Front end for 'Apache Spark' <https://spark.apache.org>.
Authors@R:
Expand Down
8 changes: 7 additions & 1 deletion assembly/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@
<parent>
<groupId>org.apache.spark</groupId>
<artifactId>spark-parent_2.12</artifactId>
<version>3.5.0</version>
<version>3.5.4-SNAPSHOT</version>
<relativePath>../pom.xml</relativePath>
</parent>

Expand Down Expand Up @@ -159,6 +159,12 @@
<groupId>org.apache.spark</groupId>
<artifactId>spark-connect_${scala.binary.version}</artifactId>
<version>${project.version}</version>
<exclusions>
<exclusion>
<groupId>org.apache.spark</groupId>
<artifactId>spark-connect-common_${scala.binary.version}</artifactId>
</exclusion>
</exclusions>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
Expand Down
43 changes: 43 additions & 0 deletions binder/Dockerfile
Original file line number Diff line number Diff line change
@@ -0,0 +1,43 @@
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#

FROM python:3.10-slim
# install the notebook package
RUN pip install --no-cache notebook jupyterlab

# create user with a home directory
ARG NB_USER
ARG NB_UID
ENV USER ${NB_USER}
ENV HOME /home/${NB_USER}

RUN adduser --disabled-password \
--gecos "Default user" \
--uid ${NB_UID} \
${NB_USER}
WORKDIR ${HOME}
USER ${USER}

# Make sure the contents of our repo are in ${HOME}
COPY . ${HOME}
USER root
RUN chown -R ${NB_UID} ${HOME}
RUN apt-get update && apt-get install -y openjdk-17-jre git coreutils
USER ${NB_USER}

RUN binder/postBuild

2 changes: 0 additions & 2 deletions binder/apt.txt

This file was deleted.

11 changes: 8 additions & 3 deletions binder/postBuild
100644 → 100755
Original file line number Diff line number Diff line change
Expand Up @@ -20,8 +20,13 @@
# This file is used for Binder integration to install PySpark available in
# Jupyter notebook.

# SPARK-45706: Should fail fast. Otherwise, the Binder image is successfully
# built, and it cannot be rebuilt.
set -o pipefail
set -e

VERSION=$(python -c "exec(open('python/pyspark/version.py').read()); print(__version__)")
TAG=$(git describe --tags --exact-match 2>/dev/null)
TAG=$(git describe --tags --exact-match 2> /dev/null || true)

# If a commit is tagged, exactly specified version of pyspark should be installed to avoid
# a kind of accident that an old version of pyspark is installed in the live notebook environment.
Expand All @@ -33,9 +38,9 @@ else
fi

if [[ ! $VERSION < "3.4.0" ]]; then
pip install plotly "pandas<2.0.0" "pyspark[sql,ml,mllib,pandas_on_spark,connect]$SPECIFIER$VERSION"
pip install plotly "pandas<2.0.0" "numpy>=1.15,<2" "pyspark[sql,ml,mllib,pandas_on_spark,connect]$SPECIFIER$VERSION"
else
pip install plotly "pandas<2.0.0" "pyspark[sql,ml,mllib,pandas_on_spark]$SPECIFIER$VERSION"
pip install plotly "pandas<2.0.0" "numpy>=1.15,<2" "pyspark[sql,ml,mllib,pandas_on_spark]$SPECIFIER$VERSION"
fi

# Set 'PYARROW_IGNORE_TIMEZONE' to surpress warnings from PyArrow.
Expand Down
7 changes: 6 additions & 1 deletion common/kvstore/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@
<parent>
<groupId>org.apache.spark</groupId>
<artifactId>spark-parent_2.12</artifactId>
<version>3.5.0</version>
<version>3.5.4-SNAPSHOT</version>
<relativePath>../../pom.xml</relativePath>
</parent>

Expand Down Expand Up @@ -66,6 +66,11 @@
<artifactId>commons-io</artifactId>
<scope>test</scope>
</dependency>
<dependency>
<groupId>org.apache.commons</groupId>
<artifactId>commons-lang3</artifactId>
<scope>test</scope>
</dependency>

<dependency>
<groupId>org.apache.logging.log4j</groupId>
Expand Down
2 changes: 1 addition & 1 deletion common/network-common/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@
<parent>
<groupId>org.apache.spark</groupId>
<artifactId>spark-parent_2.12</artifactId>
<version>3.5.0</version>
<version>3.5.4-SNAPSHOT</version>
<relativePath>../../pom.xml</relativePath>
</parent>

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -325,7 +325,10 @@ public TransportResponseHandler getHandler() {

@Override
public void close() {
// close is a local operation and should finish with milliseconds; timeout just to be safe
// Mark the connection as timed out, so we do not return a connection that's being closed
// from the TransportClientFactory if closing takes some time (e.g. with SSL)
this.timedOut = true;
// close should not take this long; use a timeout just to be safe
channel.close().awaitUninterruptibly(10, TimeUnit.SECONDS);
}

Expand Down
Loading
Loading