Skip to content

Blosc/python-blosc2

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Python-Blosc2

A Python wrapper for the extremely fast Blosc2 compression library

Author:The Blosc development team
Contact:blosc@blosc.org
Github:https://github.com/Blosc/python-blosc2
Actions:actions
PyPi:version
NumFOCUS:numfocus
Code of Conduct:Contributor Covenant

What it is

C-Blosc2 is the latest major version of C-Blosc, and it is backward compatible with both the C-Blosc1 API and its in-memory format. Python-Blosc2 is a Python package that wraps C-Blosc2, the most recent version of the Blosc compressor.

Starting with version 3.0.0, Python-Blosc2 includes a powerful computing engine capable of operating on compressed data stored in-memory, on-disk, or across the network. This engine also supports advanced features such as reductions, filters, user-defined functions, and broadcasting (the latter is still in beta).

You can read some of our tutorials on how to perform advanced computations at the following links:

Additionally, Python-Blosc2 aims to fully leverage the functionality of C-Blosc2, supporting super-chunks (SChunk), multi-dimensional arrays (NDArray), metadata, serialization, and other features introduced in C-Blosc2.

Note: Blosc2 is designed to be backward compatible with Blosc(1) data. This means it can read data generated by Blosc, but the reverse is not true (i.e. there is no forward compatibility).

NDArray: an N-Dimensional store

One of the most useful abstractions in Python-Blosc2 is the NDArray object. It enables highly efficient reading and writing of n-dimensional datasets through a two-level n-dimensional partitioning system. This allows for more fine-grained slicing and manipulation of arbitrarily large and compressed data:

https://github.com/Blosc/python-blosc2/blob/main/images/b2nd-2level-parts.png?raw=true

To pique your interest, here is how the NDArray object performs when retrieving slices orthogonal to the different axis of a 4-dimensional dataset:

https://github.com/Blosc/python-blosc2/blob/main/images/Read-Partial-Slices-B2ND.png?raw=true

We have written a blog post on this topic: https://www.blosc.org/posts/blosc2-ndim-intro

We also have a ~2 min explanatory video on why slicing in a pineapple-style (aka double partition) is useful:

Slicing a dataset in pineapple-style

Operating with NDArrays

The NDArray objects are easy to work with in Python-Blosc2. Here it is a simple example:

import numpy as np
import blosc2

N = 10_000
na = np.linspace(0, 1, N * N, dtype=np.float32).reshape(N, N)
nb = np.linspace(1, 2, N * N).reshape(N, N)
nc = np.linspace(-10, 10, N * N).reshape(N, N)

# Convert to blosc2
a = blosc2.asarray(na)
b = blosc2.asarray(nb)
c = blosc2.asarray(nc)

# Expression
expr = ((a ** 3 + blosc2.sin(c * 2)) < b) & (c > 0)

# Evaluate and get a NDArray as result
out = expr.compute()
print(out.info)

As you can see, the NDArray instances are very similar to NumPy arrays, but behind the scenes, they store compressed data that can be processed efficiently using the new computing engine included in Python-Blosc2.

To pique your interest, here is the performance (measured on a MacBook Air M2 with 24 GB of RAM) you can achieve when the operands fit comfortably in memory:

Performance when operands fit in-memory

In this case, the performance is somewhat below that of top-tier libraries like Numexpr or Numba, but it is still quite good. Using CPUs with more cores than the M2 could further reduce the performance gap. One important point to note is that the memory consumption when using the LazyArray.compute() method is very low because the output is an NDArray object, which is compressed and stored in memory by default. On the other hand, the LazyArray.__getitem__() method returns an actual NumPy array, so it is not recommended for large datasets, as it can consume a significant amount of memory (though it may still be convenient for small outputs).

It is also important to note that the NDArray object can utilize memory-mapped files, and the benchmark above actually uses a memory-mapped file for operand storage. Memory-mapped files are particularly useful when the operands do not fit in-memory, while still maintaining good performance.

And here is the performance when the operands do not fit well in memory:

Performance when operands do not fit in-memory

In this latter case, the memory consumption figures may seem a bit extreme, but this is because the displayed values represent actual memory consumption, not virtual memory. During evaluation, the OS may need to swap some memory to disk. In this scenario, the performance compared to top-tier libraries like Numexpr or Numba is quite competitive.

You can find the benchmark for the examples above at: https://github.com/Blosc/python-blosc2/blob/main/bench/ndarray/lazyarray-expr.ipynb

Installing

Blosc2 now provides Python wheels for the major OS (Win, Mac and Linux) and platforms. You can install the binary packages from PyPi using pip:

pip install blosc2

We are in the process of releasing 3.0.0, along with wheels for various beta versions. For example, to install the first beta version, you can use:

pip install blosc2==3.0.0b1

Documentation

The documentation is available here:

https://blosc.org/python-blosc2/python-blosc2.html

Additionally, you can find some examples at:

https://github.com/Blosc/python-blosc2/tree/main/examples

Building from sources

python-blosc2 includes the C-Blosc2 source code and can be built in place:

git clone https://github.com/Blosc/python-blosc2/
cd python-blosc2
pip install .   # add -e for editable mode

That's it! You can now proceed to the testing section.

Testing

After compiling, you can quickly verify that the package is functioning correctly by running the tests:

pip install .[test]
pytest  (add -v for verbose mode)

Benchmarking

If you are curious, you may want to run a small benchmark that compares a plain NumPy array copy against compression using different compressors in your Blosc build:

python bench/pack_compress.py

License

This software is licensed under a 3-Clause BSD license. A copy of the python-blosc2 license can be found in LICENSE.txt.

Mailing list

Discussion about this module are welcome on the Blosc mailing list:

blosc@googlegroups.com

https://groups.google.es/group/blosc

Mastodon

Please follow @Blosc2 to stay updated on the latest developments. We recently moved from Twitter to Mastodon.

Citing Blosc

You can cite our work on the various libraries under the Blosc umbrella as follows:

@ONLINE{blosc,
  author = {{Blosc Development Team}},
  title = "{A fast, compressed and persistent data store library}",
  year = {2009-2024},
  note = {https://blosc.org}
}

Make compression better!