Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Large Array Handling Error in estimate_place_fields Function #36

Closed
ncguilbeault opened this issue Sep 20, 2024 · 1 comment
Closed

Comments

@ncguilbeault
Copy link
Contributor

Description

There is an issue in the estimate_place_fields function in spiking_likelihood_glm.py when using large arrays of spiking data (>2GB). The function fails, and looking through the callstack shows a ValueError: xxx exceeds max_bin_length(2147483647). This error stems from the line results = dask.compute(*results) and is documented in the dask repository several times (for example here and here).

Current Behavior

When using large spike arrays (>2GB), the function fails with the aforementioned ValueError.

Expected Behavior

The function should be able to handle large spike arrays without throwing an error.

Proposed Solution

A simple fix for this would be to call client.scatter on the spikes array before passing it to the dask compute cluster.

Steps to Reproduce

Use the estimate_place_fields function with a spike array larger than 2GB. Observe the ValueError being thrown.

@edeno
Copy link
Contributor

edeno commented Sep 21, 2024

Closed by #37

@edeno edeno closed this as completed Sep 21, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants