Jobsub ID 379740.178@justin-prod-sched01.dune.hep.ac.uk
Jobsub ID | 379740.178@justin-prod-sched01.dune.hep.ac.uk |
Workflow ID | 6536 |
Stage ID | 1 |
User name | calcuttj@fnal.gov |
HTCondor Group | group_dune |
Requested | Processors | 1 |
GPU | No |
RSS bytes | 4193255424 (3999 MiB) |
Wall seconds limit | 80000 (22 hours) |
Submitted time | 2025-05-01 17:52:40 |
Site | UK_Edinburgh |
Entry | DUNE_UK_SGridECDF_ce1 |
Last heartbeat | 2025-05-01 23:59:20 |
From worker node | Hostname | node2b06.ecdf.ed.ac.uk |
cpuinfo | Intel(R) Xeon(R) Gold 6338 CPU @ 2.00GHz |
OS release | Scientific Linux release 7.9 (Nitrogen) |
Processors | 1 |
RSS bytes | 4193255424 (3999 MiB) |
Wall seconds limit | 171000 (47 hours) |
GPU | |
Inner Apptainer? | True |
Job state | jobscript_error |
Allocator name | justin-allocator-pro.dune.hep.ac.uk |
Started | 2025-05-01 23:59:03 |
Input files | monte-carlo-006536-000236
|
Jobscript | Exit code | 1 |
Real time | 0m (0s) |
CPU time | 0m (0s = 0%) |
Max RSS bytes | 0 (0 MiB) |
Outputting started | |
Output files | |
Finished | 2025-05-01 23:59:20 |
Saved logs | justin-logs:379740.178-justin-prod-sched01.dune.hep.ac.uk.logs.tgz |
List job events Wrapper job log |
Jobscript log (last 10,000 characters)
Setting up larsoft UPS area... /cvmfs/larsoft.opensciencegrid.org
Setting up DUNE UPS area... /cvmfs/dune.opensciencegrid.org/products/dune/
Justin processors: 1
did_pfn_rse monte-carlo-006536-000236 000236 MONTECARLO
178 379740
usage: hadd [-a A] [-k K] [-T T] [-O O] [-v V] [-j J] [-dbg DBG] [-d D] [-n N]
[-cachesize CACHESIZE]
[-experimental-io-features EXPERIMENTAL_IO_FEATURES] [-f F]
[-fk FK] [-ff FF] [-f0 F0] [-f6 F6]
TARGET SOURCES
OPTIONS:
-a Append to the output
-k Skip corrupt or non-existent files, do not exit
-T Do not merge Trees
-O Re-optimize basket size when merging TTree
-v Explicitly set the verbosity level: 0 request no output, 99 is the default
-j Parallelize the execution in multiple processes
-dbg Parallelize the execution in multiple processes in debug mode (Does not delete partial files stored inside working directory)
-d Carry out the partial multiprocess execution in the specified directory
-n Open at most 'maxopenedfiles' at once (use 0 to request to use the system maximum)
-cachesize Resize the prefetching cache use to speed up I/O operations(use 0 to disable)
-experimental-io-features Used with an argument provided, enables the corresponding experimental feature for output trees
-f Gives the ability to specify the compression level of the target file(by default 4)
-fk Sets the target file to contain the baskets with the same compression
as the input files (unless -O is specified). Compresses the meta data
using the compression level specified in the first input or the
compression setting after fk (for example 206 when using -fk206)
-ff The compression level use is the one specified in the first input
-f0 Do not compress the target file
-f6 Use compression level 6. (See TFile::SetCompressionSettings for the support range of value.)
TARGET Target file
SOURCES Source files
Querying usertests:calcuttj_g4bl_prod_full_1_042425-w6401s1p1 for 10 files
Query: files from usertests:calcuttj_g4bl_prod_full_1_042425-w6401s1p1 where dune.output_status=confirmed ordered skip 2350 limit 10
Getting names and metadata
done
{'core.runs': [379740], 'core.runs_subruns': [37974000178]}
Getting paths from rucio
Got 0 paths from 0 files
['hadd', '']
Finishing metadata
Traceback (most recent call last):
File "/cvmfs/fifeuser2.opensciencegrid.org/sw/dune/04714e6ef575ca47529605518ed919ebaf29bea8/merge_g4bl.py", line 136, in <module>
results = finish_metadata(
File "/cvmfs/fifeuser2.opensciencegrid.org/sw/dune/04714e6ef575ca47529605518ed919ebaf29bea8/merge_g4bl.py", line 17, in finish_metadata
results['size'] = os.path.getsize(outname)
File "/cvmfs/larsoft.opensciencegrid.org/products/python/v3_9_15/Linux64bit+3.10-2.17/lib/python3.9/genericpath.py", line 50, in getsize
return os.stat(filename).st_size
FileNotFoundError: [Errno 2] No such file or directory: ''
Exiting with error