Dear All,
Apologies for cross-posting! If anyone has come across below issue, please
let me know.
I was trying to write compound datatype (containing region reference) in
parallel using collective operation. With h5dump I saw that the region
references were not created correctly. I looked into "Collective Calling
Requirements" document and then realised that writing region reference is
not supported in parallel i/o. I tried simple test example which gives an
error "*H5Dio.c line 664 in H5D__write(): Parallel IO does not support
writing region reference datatypes yet*".
But note that this error does't appear if I am writing a compound datatype
containing region reference. Is this allowed/possible? (i.e. first create
region reference and store it as the member of compound datatype and then
write compound datatype in parallel collectively).
The dataset that I am creating is for neuron cells which are vary diverse
in terms of sizes. Each mpi rank is writing variable-length data to dataset
and I thought region-reference would be helpful but now have the issue
while writing region references.
What are possible alternatives? I can think of:
- write <count,offset> pair to emulate region reference? (but lose
flexibility as other users are going to use tools/python libraries to read
datasets and in this case they have to use count+offset pair)
- another option could be to write datasets first and then single rank can
create and write region references. (each cell has thousands of
region_references and for large simulations this may not scale (?) )
Any other suggestions?
Thanks,
Pramod