Different errors depending on the chunking job
Hi @aho,
I tried to apply my script (the very long one of the previous issue #154 (closed) ) on all my dataset.
I asked for more num_procs in the StartR::load as you suggested (even if when trying on a subset of the data, it didn't change the total execution time).
I asked to chunk by latitude, 1 chunk by number of latitudes of the targeted grid (when loading, I'm doing an interpolation).
Depending on the chunk, I have different situation:
- the first chunk for instance succeeded, cf this log file
/gpfs/scratch/bsc32/bsc32283/startR_hpc/STARTR_CHUNKING_8002622945/STARTR_CHUNKING_8002622945/computation/lon_CHUNK_1/lat_CHUNK_1/var_CHUNK_1/Chunk.1.err
(it seems to be the only one succeeding). - the second chunk for instance failed with this error (
/gpfs/scratch/bsc32/bsc32283/startR_hpc/STARTR_CHUNKING_8002622945/STARTR_CHUNKING_8002622945/computation/lon_CHUNK_1/lat_CHUNK_2/var_CHUNK_1/Chunk.1.err
):
Error in Apply(list(c(NA_real_, NA_real_, NA_real_, NA_real_, NA_real_, :
Found one or more margin dimensions with the same name and different length in some of the input objects in 'data'.
Calls: do.call -> Apply
Execution halted
- the chunk 12 failed with this error (
/gpfs/scratch/bsc32/bsc32283/startR_hpc/STARTR_CHUNKING_8002622945/STARTR_CHUNKING_8002622945/computation/lon_CHUNK_1/lat_CHUNK_12/var_CHUNK_1/Chunk.1.err
):
The data cannot be loaded.
See the original error message:
missing value where TRUE/FALSE needed
Current files in /dev/shm:
/.statelite/tmpfs/gpfs/scratch/bsc32/bsc32283/startR_hpc/STARTR_CHUNKING_8002622945
Files _8002622945_1_12_1_ has been removed.
Error in value[[3L]](cond) :
The job has failed while loading data. See original error reported above.
Calls: tryCatch -> tryCatchList -> tryCatchOne -> <Anonymous>
In addition: There were 21 warnings (use warnings() to see them)
Execution halted
It seems to be related to the chunking parameters, but I'm not totally sure.
I'll try with another number of chunks but could you explain me a bit more in detail what is happening here?
Thanks
Aude