diff --git a/samples/samples/batch_sample.py b/samples/samples/batch_sample.py index 69913ac4b3..d11dd5f95a 100644 --- a/samples/samples/batch_sample.py +++ b/samples/samples/batch_sample.py @@ -50,7 +50,7 @@ def run_batch_query(instance_id, database_id): # A Partition object is serializable and can be used from a different process. # DataBoost option is an optional parameter which can also be used for partition read # and query to execute the request via spanner independent compute resources. - data_boost_enabled=True, + data_boost_enabled=False, ) # Create a pool of workers for the tasks diff --git a/tests/system/test_session_api.py b/tests/system/test_session_api.py index 7d58324b04..3fd30958b7 100644 --- a/tests/system/test_session_api.py +++ b/tests/system/test_session_api.py @@ -1890,7 +1890,7 @@ def test_partition_read_w_index(sessions_database, not_emulator): columns, spanner_v1.KeySet(all_=True), index="name", - data_boost_enabled=True, + data_boost_enabled=False, ) for batch in batches: p_results_iter = batch_txn.process(batch) @@ -2507,7 +2507,7 @@ def test_partition_query(sessions_database, not_emulator): all_data_rows = set(_row_data(row_count)) union = set() batch_txn = sessions_database.batch_snapshot(read_timestamp=committed) - for batch in batch_txn.generate_query_batches(sql, data_boost_enabled=True): + for batch in batch_txn.generate_query_batches(sql, data_boost_enabled=False): p_results_iter = batch_txn.process(batch) # Lists aren't hashable so the results need to be converted rows = [tuple(result) for result in p_results_iter]