-
Notifications
You must be signed in to change notification settings - Fork 29k
[SPARK-34142][CORE] Support Fallback Storage Cleanup during stopping SparkContext #31215
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
Kubernetes integration test starting |
|
Kubernetes integration test status success |
|
Test build #134157 has finished for PR 31215 at commit
|
| .checkValue(_.endsWith(java.io.File.separator), "Path should end with separator.") | ||
| .createOptional | ||
|
|
||
| private[spark] val STORAGE_DECOMMISSION_FALLBACK_STORAGE_CLEANUP = |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
looks fine
|
Thank you for review, @yikf . |
|
Could you review this please, @viirya ? |
core/src/main/scala/org/apache/spark/storage/FallbackStorage.scala
Outdated
Show resolved
Hide resolved
| .createOptional | ||
|
|
||
| private[spark] val STORAGE_DECOMMISSION_FALLBACK_STORAGE_CLEANUP = | ||
| ConfigBuilder("spark.storage.decommission.fallbackStorage.cleanUp") |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
BTW, should we plan to document this fallback storage somewhere like https://spark.apache.org/docs/latest/configuration.html#kubernetes or https://spark.apache.org/docs/latest/job-scheduling.html#graceful-decommission-of-executors? Should better be done separately though.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Sure, I'll do that separately. Thanks.
|
Looks pretty good to me |
…cala Co-authored-by: Hyukjin Kwon <gurwls223@gmail.com>
|
Thank you for review and approval, @HyukjinKwon . For the last commit, this PR is only tested by |
|
Merged to master for Apache Spark 3.2.0. |
viirya
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
lgtm too.
|
Thank you, @viirya ! :) |
…SparkContext ### What changes were proposed in this pull request? This PR aims to support fallback storage clean-up during stopping `SparkContext`. ### Why are the changes needed? SPARK-33545 added `Support Fallback Storage during worker decommission` for the managed cloud-storages with TTL support. Usually, it's one day. This PR will add an additional clean-up feature during stopping `SparkContext` in order to save some money before TTL or the other HDFS-compatible storage which doesn't have TTL support. ### Does this PR introduce _any_ user-facing change? Yes, but this is a new feature. ### How was this patch tested? Pass the newly added UT. Closes apache#31215 from dongjoon-hyun/SPARK-34142. Authored-by: Dongjoon Hyun <dongjoon@apache.org> Signed-off-by: Dongjoon Hyun <dhyun@apple.com> (cherry picked from commit 415506c) Signed-off-by: Dongjoon Hyun <dhyun@apple.com>
What changes were proposed in this pull request?
This PR aims to support fallback storage clean-up during stopping
SparkContext.Why are the changes needed?
SPARK-33545 added
Support Fallback Storage during worker decommissionfor the managed cloud-storages with TTL support. Usually, it's one day. This PR will add an additional clean-up feature during stoppingSparkContextin order to save some money before TTL or the other HDFS-compatible storage which doesn't have TTL support.Does this PR introduce any user-facing change?
Yes, but this is a new feature.
How was this patch tested?
Pass the newly added UT.