Skip to content
Snippets Groups Projects
Commit e82bddae authored by Arin Wongprommoon's avatar Arin Wongprommoon
Browse files

Extractor properly handles number of time points from metadata

When an extractor is defined (based on the Extractor object) without
specifying time points (the `tps` argument), the program raises a
`TypeError`.

The definition of `tps` when it is not otherwise defined in `.run()` is
based on the assumption that the number of time points is stored as a
scalar in the metadata of the HDF5
file (`self.meta["time_settings/ntimepoints"]`).  This is wrong -- the
number of time points is stored as a one-element numpy array.  The
reason it is stored in this way is down to the behaviour of the log file
parser (https://git.ecdf.ed.ac.uk/swain-lab/aliby/agora/-/blob/master/logfile_parser/logfile_parser.py)
and how it processes tables as defined in the JSON file that specifies
the grammar.

This issue does not occur when the whole pipeline is run because the
scalar element of the numpy array that stores the number of time points
is properly extracted in
https://git.ecdf.ed.ac.uk/swain-lab/aliby/aliby/-/blob/master/aliby/pipeline.py,
line 89.

The alternative way to fix it was to fix the behaviour of the log file
parser.  The 'proper' way is to store the number of time points as
scalars, not an array with one element.  However:
 (a) It is JSON-parsing code that I am not familiar with, and
 (b) There is a risk of breaking parts of the code base that relies on
 other parts of the metadata.

So, this is essentially a quick-and-dirty fix, and it *should* be okay
given that `time_settings/ntimepoints` only appears twice in the
project.  However, if we decide to handle the metadata
differently (unlikely given the fixed nature of our log files, and we've
kind of exhausted all the useful information already), we will be in for
adding `[0]`s at various places.

Fixes issue #20.
parent de94d475
No related branches found
No related tags found
Loading
Loading
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment