justIN           Dashboard       Workflows       Jobs       AWT       Sites       Storages       Docs       Login

Workflow 4105, Stage 1

Priority50
Processors1
Wall seconds80000
RSS bytes4194304000 (4000 MiB)
Max distance for inputs30.0
Enabled input RSEs CERN_PDUNE_EOS, DUNE_CA_SFU, DUNE_CERN_EOS, DUNE_ES_PIC, DUNE_FR_CCIN2P3_DISK, DUNE_IN_TIFR, DUNE_IT_INFN_CNAF, DUNE_UK_GLASGOW, DUNE_UK_LANCASTER_CEPH, DUNE_UK_MANCHESTER_CEPH, DUNE_US_BNL_SDCC, DUNE_US_FNAL_DISK_STAGE, FNAL_DCACHE, FNAL_DCACHE_STAGING, FNAL_DCACHE_TEST, MONTECARLO, NIKHEF, PRAGUE, QMUL, RAL-PP, RAL_ECHO, SURFSARA, T3_US_NERSC
Enabled output RSEs CERN_PDUNE_EOS, DUNE_CA_SFU, DUNE_CERN_EOS, DUNE_ES_PIC, DUNE_FR_CCIN2P3_DISK, DUNE_IN_TIFR, DUNE_IT_INFN_CNAF, DUNE_UK_GLASGOW, DUNE_UK_LANCASTER_CEPH, DUNE_UK_MANCHESTER_CEPH, DUNE_US_BNL_SDCC, DUNE_US_FNAL_DISK_STAGE, FNAL_DCACHE, FNAL_DCACHE_STAGING, FNAL_DCACHE_TEST, NIKHEF, PRAGUE, QMUL, RAL-PP, RAL_ECHO, SURFSARA, T3_US_NERSC
Enabled sites BR_CBPF, CA_SFU, CA_Victoria, CERN, CH_UNIBE-LHEP, CZ_FZU, ES_CIEMAT, ES_PIC, FR_CCIN2P3, IN_TIFR, IT_CNAF, NL_NIKHEF, NL_SURFsara, UK_Bristol, UK_Brunel, UK_Durham, UK_Edinburgh, UK_Imperial, UK_Lancaster, UK_Liverpool, UK_Manchester, UK_Oxford, UK_QMUL, UK_RAL-PPD, UK_RAL-Tier1, UK_Sheffield, US_BNL, US_Caltech, US_Colorado, US_FNAL-FermiGrid, US_FNAL-T1, US_Michigan, US_MIT, US_Nebraska, US_NotreDame, US_PuertoRico, US_SU-ITS, US_Swan, US_UChicago, US_UConn-HPC, US_UCSD, US_Wisconsin
Scopeusertests
Events for this stage

Output patterns

 DestinationPatternLifetimeFor next stage
1https://fndcadoor.fnal.gov:2880/dune/scratch/users/lavaut/04105/1*singleHit*.*

Environment variables

NameValue
INPUT_TAR_DIR_LOCAL/cvmfs/fifeuser3.opensciencegrid.org/sw/dune/b45eb38553651f97b6c3d6135486f05ed5eb726f

File states

Total filesFindingUnallocatedAllocatedOutputtingProcessedNot foundFailed
100004970000503

Job states

TotalSubmittedStartedProcessingOutputtingFinishedNotusedAbortedStalledJobscript errorOutputting failedNone processed
3840000000038400
Replicas per RSE1000485.3234.0692286.7851429166925282.9804085215564101303.80810225145626161.316544403313892328.98169817408484141.8847304796874568353.4926679252071132.090965335370547372.2330320172109128.98683792802768Replicas per RSEDUNE_US_FNAL_DISK_STAGE (50…DUNE_US_FNAL_DISK_STAGE (50%)RAL-PP (34%)PRAGUE (5%)RAL_ECHO (4%)DUNE_CERN_EOS (3%)DUNE_ES_PIC (2%)

RSEs used

NameInputsOutputs
RAL-PP21930
DUNE_US_FNAL_DISK_STAGE5200
PRAGUE3560
RAL_ECHO3550
DUNE_CERN_EOS2070
DUNE_ES_PIC2040

Stats of processed input files as CSV or JSON, and of uploaded output files as CSV or JSON (up to 10000 files included)

Jobscript

#!/bin/bash
:<<'EOF'

To use this jobscript to process 10 files from the dc4-vd-coldbox-bottom
data and put the output in the usertests namespace (MetaCat) and 
scope (Rucio), and in the usertests:output-test-01 dataset in MetaCat and
Rucio, use this command to create the workflow:

justin simple-workflow \
--mql \
"files from dune:all where core.run_type='dc4-vd-coldbox-bottom' and dune.campaign='dc4' limit 10" \
--jobscript dc4-vd-coldbox-bottom.jobscript --max-distance 30 --rss-mb 4000 \
--scope usertests --output-pattern '*_reco_data_*.root:output-test-01' 

The following optional environment variables can be set when creating the
workflow/stage: FCL_FILE, NUM_EVENTS, DUNE_VERSION, DUNE_QUALIFIER 

EOF

# fcl file and DUNE software version/qualifier to be used
FCL_FILE=MyrunSingleHit_PDHD.fcl
DUNE_VERSION=${DUNE_VERSION:-v10_00_02d00}
DUNE_QUALIFIER=${DUNE_QUALIFIER:-e26:prof}

# number of events to process from the input file
if [ "$NUM_EVENTS" != "" ] ; then
 events_option="-n $NUM_EVENTS"
fi

# First get an unprocessed file from this stage
for nf in {1..10}
do
   did_pfn_rse=`$JUSTIN_PATH/justin-get-file`
   ##Check that any file was returned
   if [ "${did_pfn_rse}" == "" ] ; then
     echo "Could not get file"
     #exit 0
     continue
   fi
   FILE=`echo ${did_pfn_rse} | cut -f2 -d' '`
   DID=`echo ${did_pfn_rse} | cut -f1 -d' '`
   echo ${DID}  >> did.list
   echo ${FILE} >> file.list

done

# id_pfn_rse=`$JUSTIN_PATH/justin-get-file`

#if [ "$did_pfn_rse" = "" ] ; then
 # echo "Nothing to process - exit jobscript"
 # exit 0
#fi

# Keep a record of all input DIDs, for pdjson2meta file -> DID mapping
echo "$did_pfn_rse" | cut -f1 -d' ' >>all-input-dids.txt

# pfn is also needed when creating justin-processed-pfns.txt
pfn=`echo $did_pfn_rse | cut -f2 -d' '`
echo "Input PFN = $pfn"

# Setup DUNE environment
source /cvmfs/dune.opensciencegrid.org/products/dune/setup_dune.sh

export PRODUCTS=$INPUT_TAR_DIR_LOCAL:${PRODUCTS}
echo "PRODUCTS: ${PRODUCTS}"
echo "DUNESW_DIR: ${DUNESW_DIR}"


# Properly setup custom code with INPUT_TAR_DIR_LOCAL

export DUNESW_DIR=${INPUT_TAR_DIR_LOCAL}/dunesw/${DUNE_VERSION}
export DUNESW_FQ_DIR=${DUNESW_DIR}/slf7.x86_64.e26.prof
export DUNESW_LIB=${DUNESW_FQ_DIR}/lib

export PROTODUNEANA_DIR=${INPUT_TAR_DIR_LOCAL}/protoduneana/${DUNE_VERSION}
export PROTODUNEANA_LIB=${PROTODUNEANA_DIR}/include
export PROTODUNEANA_FQ_DIR=${PROTODUNEANA_DIR}/slf7.x86_64.e26.prof
export PROTODUNEANA_LIB=${PROTODUNEANA_FQ_DIR}/lib

#export WIRECELL_PATH=${DUNE_PARDATA_DIR}/WireCellData
#export WIRECELL_PATH=${WIRECELL_DIR}/Linux64bit+3.10-2.17-e26-prof/share/wirecell:${WIRECELL_PATH}
#export WIRECELL_PATH=${PROTODUNEANADIR}/wire-cell-cfg:${WIRECELL_PATH}

export DUNEPROTOTYPES_DIR=${INPUT_TAR_DIR_LOCAL}/duneprototypes/${DUNE_VERSION}
export DUNEPROTOTYPES_INC=${DUNEPROTOTYPES_DIR}/include
export DUNEPROTOTYPES_FQ_DIR=${DUNEPROTOTYPES_DIR}/slf7.x86_64.e26.prof
export DUNEPROTOTYPES_LIB=${DUNEPROTOTYPES_FQ_DIR}/lib

export FHICL_FILE_PATH=${DUNESW_DIR}/fcl:${FHICL_FILE_PATH}
export FHICL_FILE_PATH=${DUNEPROTOTYPES_DIR}/fcl:${FHICL_FILE_PATH}
export FHICL_FILE_PATH=${PROTODUNEANA_DIR}/fcl:${FHICL_FILE_PATH}
export FHICL_FILE_PATH=${INPUT_TAR_DIR_LOCAL}/myfcl:${FHICL_FILE_PATH}

echo "DUNESW_DIR: ${DUNESW_DIR}"
echo "PROTODUNEANA_DIR: ${PROTODUNEANA_DIR}"
echo "DUNEPROTOTYPES_DIR: ${DUNEPROTOTYPES_DIR}"

#echo "WIRECELL_PATH: ${WIRECELL_PATH}"
#echo "FHICL_FILE_PATH: ${FHICL_FILE_PATH}"


export OMP_NUM_THREADS=${JUSTIN_PROCESSORS} 

source "${INPUT_TAR_DIR_LOCAL}/localProducts_larsoft_v10_00_02d00_prof_e26/setup2"
setup dunesw "$DUNE_VERSION" -q "$DUNE_QUALIFIER"

echo " VERSION ACTIVE =  VERSION ACTIVE =  VERSION ACTIVE = "
ups active | grep "dunesw\|duneprototypes"
mrbslp

# Construct outFile from input $pfn 
now=$(date -u +"%Y-%m-%dT_%H%M%SZ")
Ffname=`echo $pfn | awk -F/ '{print $NF}'`
fname=`echo $Ffname | awk -F. '{print $1}'`
outFile=${fname}_singleHit_${now}.root

campaign="justIN.r${JUSTIN_WORKFLOW_ID}s${JUSTIN_STAGE_ID}"

(
# Do the scary preload stuff in a subshell!
#export LD_PRELOAD=${XROOTD_LIB}/libXrdPosixPreload.so
#echo "$LD_PRELOAD"
lar -c ${INPUT_TAR_DIR_LOCAL}/myfcl/$FCL_FILE $events_option -S file.list -T $outFile "$pfn" > ${fname}_singleHit${now}.log 2>&1
)

# Subshell exits with exit code of last command
larExit=$?
echo "lar exit code $larExit"
echo "$pfn" > justin-processed-pfns.txt

ls -lRS

# Create compressed tar file with all log files 
tar zcf `echo "$JUSTIN_JOBSUB_ID.logs.tgz" | sed 's/@/_/g'` *.log
#exit $jobscriptExit
exit $larExit

justIN time: 2024-11-17 03:38:45 UTC       justIN version: 01.01.09