Bug Reports & Contributions

Contributions and bug reports are welcome from anyone! Some of the best features in h5py, including thread support, dimension scales, and the scale-offset filter, came from user code contributions.

Since we use GitHub, the workflow will be familiar to many people. If you have questions about the process or about the details of implementing your feature, feel free to ask on Github itself, or on the h5py section of the HDF5 forum:

Posting on this forum requires registering for a free account with HDF group.

Anyone can post to this list. Your first message will be approved by a moderator, so don’t worry if there’s a brief delay.

This guide is divided into three sections. The first describes how to file a bug report.

The second describes the mechanics of how to submit a contribution to the h5py project; for example, how to create a pull request, which branch to base your work on, etc. We assume you’re are familiar with Git, the version control system used by h5py. If not, here’s a great place to start.

Finally, we describe the various subsystems inside h5py, and give technical guidance as to how to implement your changes.

How to File a Bug Report

Bug reports are always welcome! The issue tracker is at:

If you’re unsure whether you’ve found a bug

Always feel free to ask on the mailing list (h5py at Google Groups). Discussions there are seen by lots of people and are archived by Google. Even if the issue you’re having turns out not to be a bug in the end, other people can benefit from a record of the conversation.

By the way, nobody will get mad if you file a bug and it turns out to be something else. That’s just how software development goes.

What to include

When filing a bug, there are two things you should include. The first is the output of h5py.version.info:

>>> import h5py
>>> print(h5py.version.info)

The second is a detailed explanation of what went wrong. Unless the bug is really trivial, include code if you can, either via GitHub’s inline markup:

```
    import h5py
    h5py.explode()    # Destroyed my computer!
```

or by uploading a code sample to Github Gist.

How to Get Your Code into h5py

This section describes how to contribute changes to the h5py code base. Before you start, be sure to read the h5py license and contributor agreement in “license.txt”. You can find this in the source distribution, or view it online at the main h5py repository at GitHub.

The basic workflow is to clone h5py with git, make your changes in a topic branch, and then create a pull request at GitHub asking to merge the changes into the main h5py project.

Here are some tips to getting your pull requests accepted:

  1. Let people know you’re working on something. This could mean posting a comment in an open issue, or sending an email to the mailing list. There’s nothing wrong with just opening a pull request, but it might save you time if you ask for advice first.

  2. Keep your changes focused. If you’re fixing multiple issues, file multiple pull requests. Try to keep the amount of reformatting clutter small so the maintainers can easily see what you’ve changed in a diff.

  3. Unit tests are mandatory for new features. This doesn’t mean hundreds (or even dozens) of tests! Just enough to make sure the feature works as advertised. The maintainers will let you know if more are needed.

Clone the h5py repository

The best way to do this is by signing in to GitHub and cloning the h5py project directly. You’ll end up with a new repository under your account; for example, if your username is yourname, the repository would be at http://github.com/yourname/h5py.

Then, clone your new copy of h5py to your local machine:

$ git clone http://github.com/yourname/h5py

Create a topic branch for your feature

Check out a new branch for the bugfix or feature you’re writing:

$ git checkout -b newfeature master

The exact name of the branch can be anything you want. For bug fixes, one approach is to put the issue number in the branch name.

We develop all changes against the master branch. If we’re making a bugfix release, a bot will backport merged pull requests.

Implement the feature!

You can implement the feature as a number of small changes, or as one big commit; there’s no project policy. Double-check to make sure you’ve included all your files; run git status and check the output.

Run the tests

The easiest way to run the tests is with tox:

pip install tox  # Get tox

tox -e py37-test-deps  # Run tests in one environment
tox                    # Run tests in all possible environments
tox -a                 # List defined environments

Write a release note

Changes which could affect people building and using h5py after the next release should have a news entry. You don’t need to do this if your changes don’t affect usage, e.g. adding tests or correcting comments.

In the news/ folder, make a copy of TEMPLATE.rst named after your branch. Edit the new file, adding a sentence or two about what you’ve added or fixed. Commit this to git too.

News entries are merged into the what’s new documents for each release. They should allow someone to quickly understand what a new feature is, or whether a bug they care about has been fixed. E.g.:

Bug fixes
---------

* Fix reading data for region references pointing to an empty selection.

The Building h5py section is for changes which affect how people build h5py from source. It’s not about how we make prebuilt wheels; changes to that which make a visible difference can go in New features or Bug fixes.

Push your changes back and open a pull request

Push your topic branch back up to your GitHub clone:

$ git push origin newfeature

Then, create a pull request based on your topic branch.

Work with the maintainers

Your pull request might be accepted right away. More commonly, the maintainers will post comments asking you to fix minor things, like add a few tests, clean up the style to be PEP-8 compliant, etc.

The pull request page also shows the results of building and testing the modified code on Travis and Appveyor CI and Azure Pipelines. Check back after about 30 minutes to see if the build succeeded, and if not, try to modify your changes to make it work.

When making changes after creating your pull request, just add commits to your topic branch and push them to your GitHub repository. Don’t try to rebase or open a new pull request! We don’t mind having a few extra commits in the history, and it’s helpful to keep all the history together in one place.

How to Modify h5py

This section is a little more involved, and provides tips on how to modify h5py. The h5py package is built in layers. Starting from the bottom, they are:

  1. The HDF5 C API (provided by libhdf5)

  2. Auto-generated Cython wrappers for the C API (api_gen.py)

  3. Low-level interface, written in Cython, using the wrappers from (2)

  4. High-level interface, written in Python, with things like h5py.File.

  5. Unit test code

Rather than talk about the layers in an abstract way, the parts below are guides to adding specific functionality to various parts of h5py. Most sections span at least two or three of these layers.

Adding a function from the HDF5 C API

This is one of the most common contributed changes. The example below shows how one would add the function H5Dget_storage_size, which determines the space on disk used by an HDF5 dataset. This function is already partially wrapped in h5py, so you can see how it works.

It’s recommended that you follow along, if not by actually adding the feature then by at least opening the various files as we work through the example.

First, get ahold of the function signature; the easiest place for this is at the online HDF5 Reference Manual. Then, add the function’s C signature to the file api_functions.txt:

hsize_t   H5Dget_storage_size(hid_t dset_id)

This particular signature uses types (hsize_t, hid_t) which are already defined elsewhere. But if the function you’re adding needs a struct or enum definition, you can add it using Cython code to the file api_types_hdf5.pxd.

The next step is to add a Cython function or method which calls the function you added. The h5py modules follow the naming convention of the C API; functions starting with H5D are wrapped in h5d.pyx.

Opening h5d.pyx, we notice that since this function takes a dataset identifier as the first argument, it belongs as a method on the DatasetID object. We write a wrapper method:

def get_storage_size(self):
    """ () => LONG storage_size

        Determine the amount of file space required for a dataset.  Note
        this only counts the space which has actually been allocated; it
        may even be zero.
    """
    return H5Dget_storage_size(self.id)

The first line of the docstring gives the method signature. This is necessary because Cython will use a “generic” signature like method(*args, **kwds) when the file is compiled. The h5py documentation system will extract the first line and use it as the signature.

Next, we decide whether we want to add access to this function to the high-level interface. That means users of the top-level h5py.Dataset object will be able to see how much space on disk their files use. The high-level interface is implemented in the subpackage h5py._hl, and the Dataset object is in module dataset.py. Opening it up, we add a property on the Dataset object:

@property
def storagesize(self):
    """ Size (in bytes) of this dataset on disk. """
    return self.id.get_storage_size()

You’ll see that the low-level DatasetID object is available on the high-level Dataset object as obj.id. This is true of all the high-level objects, like File and Group as well.

Finally (and don’t skip this step), we write unit tests for this feature. Since the feature is ultimately exposed at the high-level interface, it’s OK to write tests for the Dataset.storagesize property only. Unit tests for the high-level interface are located in the “tests” subfolder, right near dataset.py.

It looks like the right file is test_dataset.py. Unit tests are implemented as methods on custom unittest.UnitTest subclasses; each new feature should be tested by its own new class. In the test_dataset module, we see there’s already a subclass called BaseDataset, which implements some simple set-up and cleanup methods and provides a h5py.File object as obj.f. We’ll base our test class on that:

class TestStorageSize(BaseDataset):

    """
        Feature: Dataset.storagesize indicates how much space is used.
    """

    def test_empty(self):
        """ Empty datasets take no space on disk """
        dset = self.f.create_dataset("x", (100,100))
        self.assertEqual(dset.storagesize, 0)

    def test_data(self):
        """ Storage size is correct for non-empty datasets """
        dset = self.f.create_dataset("x", (100,), dtype='uint8')
        dset[...] = 42
        self.assertEqual(dset.storagesize, 100)

This set of tests would be adequate to get a pull request approved. We don’t test every combination under the sun (different ranks, datasets with more than 2**32 elements, datasets with the string “kumquat” in the name…), but the basic, commonly encountered set of conditions.

To build and test our changes, we have to do a few things. First of all, run the file api_gen.py to re-generate the Cython wrappers from api_functions.txt:

$ python api_gen.py

Then build the project, which recompiles h5d.pyx:

$ python setup.py build

Finally, run the test suite, which includes the two methods we just wrote:

$ python setup.py test

If the tests pass, the feature is ready for a pull request.

Adding a function only available in certain versions of HDF5

At the moment, h5py must be backwards-compatible all the way back to HDF5 1.8.4. Starting with h5py 2.2.0, it’s possible to conditionally include functions which only appear in newer versions of HDF5. It’s also possible to mark functions which require Parallel HDF5. For example, the function H5Fset_mpi_atomicity was introduced in HDF5 1.8.9 and requires Parallel HDF5. Specifiers before the signature in api_functions.txt communicate this:

MPI 1.8.9 herr_t H5Fset_mpi_atomicity(hid_t file_id, hbool_t flag)

You can specify either, both or none of “MPI” or a version number in “X.Y.Z” format.

In the Cython code, these show up as “preprocessor” defines MPI and HDF5_VERSION. So the low-level implementation (as a method on h5py.h5f.FileID) looks like this:

IF MPI and HDF5_VERSION >= (1, 8, 9):

    def set_mpi_atomicity(self, bint atomicity):
        """ (BOOL atomicity)

        For MPI-IO driver, set to atomic (True), which guarantees sequential
        I/O semantics, or non-atomic (False), which improves  performance.

        Default is False.

        Feature requires: 1.8.9 and Parallel HDF5
        """
        H5Fset_mpi_atomicity(self.id, <hbool_t>atomicity)

High-level code can check the version of the HDF5 library, or check to see if the method is present on FileID objects.

Testing MPI-only features/code

Typically to run code under MPI, mpirun must be used to start the MPI processes. Similarly, tests using MPI features (such as collective IO), must also be run under mpirun. h5py uses pytest markers (specifically pytest.mark.mpi and other markers from pytest-mpi) to specify which tests require usage of mpirun, and will handle skipping the tests as needed. A simple example of how to do this is:

@pytest.mark.mpi
def test_mpi_feature():
   import mpi4py
   # test the MPI feature

To run these tests, you’ll need to:

  1. Have tox installed (e.g. via pip install tox)

  2. Have HDF5 built with MPI as per Building against Parallel HDF5

Then running:

$ CC='mpicc' HDF5_MPI=ON tox -e py37-test-deps-mpi4py

should run the tests. You may need to pass HDF5_DIR depending on the location of the HDF5 with MPI support. You can choose which python version to build against by changing py37 (e.g. py36 runs python 3.6, this is a tox feature), and test with the minimum version requirements by using mindeps rather than deps.

If you get an error similar to:

There are not enough slots available in the system to satisfy the 4 slots
that were requested by the application:
  python

Either request fewer slots for your application, or make more slots available
for use.

then you need to reduce the number of MPI processes you are asking MPI to use. If you have already reduced the number of processes requested (or are running the default number which is 2), you will need to look up the documentation for your MPI implementation for handling this error. On OpenMPI (which is usually the default MPI implementation on most systems), running:

$ export OMPI_MCA_rmaps_base_oversubscribe=1

will instruct OpenMPI to allow more MPI processes than available cores on your system.

If you need to pass additional environment variables to your MPI implementation, add these variables to the passenv setting in the tox.ini, and send us a PR with that change noting the MPI implementation.