-
Notifications
You must be signed in to change notification settings - Fork 971
[K8S][HELM] Add Spark engine configuration support #4776
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Codecov Report
@@ Coverage Diff @@
## master #4776 +/- ##
============================================
- Coverage 57.99% 57.96% -0.03%
Complexity 13 13
============================================
Files 581 581
Lines 32431 32431
Branches 4309 4309
============================================
- Hits 18807 18799 -8
- Misses 11820 11827 +7
- Partials 1804 1805 +1 see 6 files with indirect coverage changes 📣 We’re building smart automated test selection to slash your CI/CD build times. Learn more |
|
Seems we had such an idea about the structure of In practice, if Spark uses HDFS as storage and HMS as metastore, typically, the user should provide |
| spark.kubernetes.container.image.pullPolicy={{ .Values.engine.spark.image.pullPolicy }} | ||
| spark.kubernetes.container.image.pullSecrets={{ range .Values.imagePullSecrets }}{{ print .name "," }}{{ end }} | ||
|
|
||
| ### Driver resources |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
IMO this is kind of over-engineering stuff.
One advantage of Kyuubi is, it almost transparently supports all Spark features, users who are familiar w/ Spark, should easy to understand how Kyuubi works and how to configure the Spark engine.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Seems we can put everything in spark-defaults.conf to values.yaml's sparkDefaults block, in Spark native configuration.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
You are right, it is over-engineered implementation and it might confuse users about what to configure and where.
I would like to find some balance between convenient basic configuration and flexibility, but obviously it is not the best implementation.
|
Thanks for the comments!
Right, we discussed here. I'll continue with flat structure in a separate PR. As I mentioned above, it is more experimental PR to track my tries and demo different approach.
Got it! I'll add these files as well. Am I right that there is no default |
|
Closed in favor of #5934 |
Why are the changes needed?
The PR is needed to configure default properties to be used by Apache Spark as query engine.
The PR also changes
values.yamlfile structure:How was this patch tested?
Add some test cases that check the changes thoroughly including negative and positive cases if possible
Add screenshots for manual tests if appropriate
Run test locally before make a pull request