Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix: Set Data boost enabled to false #928

Merged
merged 1 commit into from
Apr 12, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion samples/samples/batch_sample.py
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,7 @@ def run_batch_query(instance_id, database_id):
# A Partition object is serializable and can be used from a different process.
# DataBoost option is an optional parameter which can also be used for partition read
# and query to execute the request via spanner independent compute resources.
data_boost_enabled=True,
data_boost_enabled=False,
)

# Create a pool of workers for the tasks
Expand Down
4 changes: 2 additions & 2 deletions tests/system/test_session_api.py
Original file line number Diff line number Diff line change
Expand Up @@ -1890,7 +1890,7 @@ def test_partition_read_w_index(sessions_database, not_emulator):
columns,
spanner_v1.KeySet(all_=True),
index="name",
data_boost_enabled=True,
data_boost_enabled=False,
)
for batch in batches:
p_results_iter = batch_txn.process(batch)
Expand Down Expand Up @@ -2507,7 +2507,7 @@ def test_partition_query(sessions_database, not_emulator):
all_data_rows = set(_row_data(row_count))
union = set()
batch_txn = sessions_database.batch_snapshot(read_timestamp=committed)
for batch in batch_txn.generate_query_batches(sql, data_boost_enabled=True):
for batch in batch_txn.generate_query_batches(sql, data_boost_enabled=False):
p_results_iter = batch_txn.process(batch)
# Lists aren't hashable so the results need to be converted
rows = [tuple(result) for result in p_results_iter]
Expand Down