Skip to content
This repository has been archived by the owner on Oct 23, 2023. It is now read-only.

Add Databricks config to Spark config #351

Merged
merged 8 commits into from
Dec 17, 2022
Merged

Add Databricks config to Spark config #351

merged 8 commits into from
Dec 17, 2022

Conversation

pingsutw
Copy link
Member

@pingsutw pingsutw commented Dec 7, 2022

TL;DR

This config is used by the dbx plugin in propeller to submit the dbx job request.

Type

  • Bug Fix
  • Feature
  • Plugin

Are all requirements met?

  • Code completed
  • Smoke tested
  • Unit tests added
  • Code documentation added
  • Any pending items have an associated Issue

Complete description

^^^

Tracking Issue

flyteorg/flyte#3173

Signed-off-by: Kevin Su <pingsutw@apache.org>
Signed-off-by: Kevin Su <pingsutw@apache.org>
Signed-off-by: Kevin Su <pingsutw@apache.org>
@codecov
Copy link

codecov bot commented Dec 7, 2022

Codecov Report

Merging #351 (4257652) into master (305c51e) will not change coverage.
The diff coverage is n/a.

@@           Coverage Diff           @@
##           master     #351   +/-   ##
=======================================
  Coverage   73.12%   73.12%           
=======================================
  Files          18       18           
  Lines        1362     1362           
=======================================
  Hits          996      996           
  Misses        315      315           
  Partials       51       51           
Flag Coverage Δ
unittests 73.12% <ø> (ø)

Flags with carried forward coverage won't be shown. Click here to find out more.

Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.

@@ -21,4 +21,8 @@ message SparkJob {
map<string, string> sparkConf = 4;
map<string, string> hadoopConf = 5;
string executorPath = 6; // Executor path for Python jobs.
// databricksConf is base64 encoded string which stores databricks job configuration.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

why is the databricks one a b64 encoded string while the spark and hadoop conf's above just a map of str, str? Is the databricks one a binary object that needs to be encoded?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Because it's nested dict. wait, I think I can use Struct here. Let me update it

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

+1

wild-endeavor
wild-endeavor previously approved these changes Dec 9, 2022
Copy link
Contributor

@wild-endeavor wild-endeavor left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

approve but maybe address question in a comment?

Signed-off-by: Kevin Su <pingsutw@apache.org>
Signed-off-by: Kevin Su <pingsutw@apache.org>
Signed-off-by: Kevin Su <pingsutw@apache.org>
Signed-off-by: Kevin Su <pingsutw@apache.org>
Signed-off-by: Kevin Su <pingsutw@apache.org>
@pingsutw pingsutw merged commit fd208b7 into master Dec 17, 2022
@pingsutw pingsutw deleted the databricks branch December 17, 2022 02:46
eapolinario pushed a commit that referenced this pull request Sep 8, 2023
* databricks plugin

Signed-off-by: Kevin Su <pingsutw@apache.org>

* update comment

Signed-off-by: Kevin Su <pingsutw@apache.org>

* Use struct instead of string

Signed-off-by: Kevin Su <pingsutw@apache.org>

* Add token

Signed-off-by: Kevin Su <pingsutw@apache.org>

* nit

Signed-off-by: Kevin Su <pingsutw@apache.org>

* add instance name

Signed-off-by: Kevin Su <pingsutw@apache.org>

* add instance name

Signed-off-by: Kevin Su <pingsutw@apache.org>

Signed-off-by: Kevin Su <pingsutw@apache.org>
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants