From bd92dae8dc8b201316d1378976d294f42c1d38b9 Mon Sep 17 00:00:00 2001 From: Kyle Wigley Date: Thu, 25 Feb 2021 22:51:47 -0500 Subject: [PATCH 1/3] update package data path --- setup.py | 17 +++++++++++------ 1 file changed, 11 insertions(+), 6 deletions(-) diff --git a/setup.py b/setup.py index 07abbc37a..e5d91465a 100644 --- a/setup.py +++ b/setup.py @@ -37,6 +37,12 @@ def _dbt_spark_version(): f'dbt_version={dbt_version}' ) +odbc_extras = ['pyodbc>=4.0.30'] +pyhive_extras = [ + 'PyHive[hive]>=0.6.0,<0.7.0', + 'thrift>=0.11.0,<0.12.0', +] +all_extras = odbc_extras + pyhive_extras setup( name=package_name, @@ -56,7 +62,8 @@ def _dbt_spark_version(): 'include/spark/dbt_project.yml', 'include/spark/sample_profiles.yml', 'include/spark/macros/*.sql', - 'include/spark/macros/**/*.sql', + 'include/spark/macros/*/*.sql', + 'include/spark/macros/*/*/*.sql', ] }, install_requires=[ @@ -64,10 +71,8 @@ def _dbt_spark_version(): 'sqlparams>=3.0.0', ], extras_require={ - "ODBC": ['pyodbc>=4.0.30'], - "PyHive": [ - 'PyHive[hive]>=0.6.0,<0.7.0', - 'thrift>=0.11.0,<0.12.0', - ], + "ODBC": odbc_extras, + "PyHive": pyhive_extras, + "all": all_extras } ) From 85a061bc867ce8986412f65afcf2a45f41e1d140 Mon Sep 17 00:00:00 2001 From: Kyle Wigley Date: Thu, 25 Feb 2021 23:16:08 -0500 Subject: [PATCH 2/3] this is much easier, include everything --- MANIFEST.in | 1 + setup.py | 10 +--------- 2 files changed, 2 insertions(+), 9 deletions(-) create mode 100644 MANIFEST.in diff --git a/MANIFEST.in b/MANIFEST.in new file mode 100644 index 000000000..78412d5b8 --- /dev/null +++ b/MANIFEST.in @@ -0,0 +1 @@ +recursive-include dbt/include *.sql *.yml *.md \ No newline at end of file diff --git a/setup.py b/setup.py index e5d91465a..091667bbe 100644 --- a/setup.py +++ b/setup.py @@ -57,15 +57,7 @@ def _dbt_spark_version(): url='https://github.com/fishtown-analytics/dbt-spark', packages=find_namespace_packages(include=['dbt', 'dbt.*']), - package_data={ - 'dbt': [ - 'include/spark/dbt_project.yml', - 'include/spark/sample_profiles.yml', - 'include/spark/macros/*.sql', - 'include/spark/macros/*/*.sql', - 'include/spark/macros/*/*/*.sql', - ] - }, + include_package_data=True, install_requires=[ f'dbt-core=={dbt_version}', 'sqlparams>=3.0.0', From f7b1f597895c55521e624ade4248dcc1e3491695 Mon Sep 17 00:00:00 2001 From: Kyle Wigley Date: Thu, 25 Feb 2021 23:32:42 -0500 Subject: [PATCH 3/3] update changelog --- CHANGELOG.md | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 500dd4f1d..098ad96ee 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,4 +1,9 @@ -## dbt-spark 0.19.0 (Release TBD) +## dbt-spark 0.19.0.1 (Release TBD) + +### Fixes +- Fix package distribution to include incremental model materializations ([#151](https://github.com/fishtown-analytics/dbt-spark/pull/151)) + +## dbt-spark 0.19.0 (February 21, 2021) ### Breaking changes - Incremental models have `incremental_strategy: append` by default. This strategy adds new records without updating or overwriting existing records. For that, use `merge` or `insert_overwrite` instead, depending on the file format, connection method, and attributes of your underlying data. dbt will try to raise a helpful error if you configure a strategy that is not supported for a given file format or connection. ([#140](https://github.com/fishtown-analytics/dbt-spark/pull/140), [#141](https://github.com/fishtown-analytics/dbt-spark/pull/141))