justIN           Dashboard       Workflows       Jobs       AWT       Sites       Storages       Docs       Login

Workflow 5470, Stage 1

Priority50
Processors1
Wall seconds80000
Image/cvmfs/singularity.opensciencegrid.org/fermilab/fnal-wn-sl7:latest
RSS bytes4193255424 (3999 MiB)
Max distance for inputs100.0
Enabled input RSEs CERN_PDUNE_EOS, DUNE_CA_SFU, DUNE_CERN_EOS, DUNE_ES_PIC, DUNE_FR_CCIN2P3_DISK, DUNE_IN_TIFR, DUNE_IT_INFN_CNAF, DUNE_UK_GLASGOW, DUNE_UK_LANCASTER_CEPH, DUNE_UK_MANCHESTER_CEPH, DUNE_US_BNL_SDCC, DUNE_US_FNAL_DISK_STAGE, FNAL_DCACHE, FNAL_DCACHE_STAGING, FNAL_DCACHE_TEST, MONTECARLO, NIKHEF, PRAGUE, QMUL, RAL-PP, RAL_ECHO, SURFSARA, T3_US_NERSC
Enabled output RSEs CERN_PDUNE_EOS, DUNE_CA_SFU, DUNE_CERN_EOS, DUNE_ES_PIC, DUNE_FR_CCIN2P3_DISK, DUNE_IN_TIFR, DUNE_IT_INFN_CNAF, DUNE_UK_GLASGOW, DUNE_UK_LANCASTER_CEPH, DUNE_UK_MANCHESTER_CEPH, DUNE_US_BNL_SDCC, DUNE_US_FNAL_DISK_STAGE, FNAL_DCACHE, FNAL_DCACHE_STAGING, FNAL_DCACHE_TEST, NIKHEF, PRAGUE, QMUL, RAL-PP, RAL_ECHO, SURFSARA, T3_US_NERSC
Enabled sites BR_CBPF, CA_SFU, CA_Victoria, CERN, CH_UNIBE-LHEP, CZ_FZU, ES_CIEMAT, ES_PIC, FR_CCIN2P3, IN_TIFR, IT_CNAF, NL_NIKHEF, NL_SURFsara, UK_Bristol, UK_Brunel, UK_Durham, UK_Edinburgh, UK_Glasgow, UK_Imperial, UK_Lancaster, UK_Liverpool, UK_Manchester, UK_Oxford, UK_RAL-PPD, UK_RAL-Tier1, UK_Sheffield, US_Caltech, US_Colorado, US_FNAL-FermiGrid, US_FNAL-T1, US_Michigan, US_MIT, US_Nebraska, US_NotreDame, US_PuertoRico, US_SU-ITS, US_Swan, US_UChicago, US_UConn-HPC, US_UCSD, US_Wisconsin
Scopefardet-hd
Events for this stage

Output patterns

 DestinationPatternLifetimeFor next stageRSE expression
1Rucio fardet-hd:fardet-hd-reco_5470-w5470s1p1*reco.root10368000False

Environment variables

NameValue
INPUT_DIR/cvmfs/fifeuser4.opensciencegrid.org/sw/dune/f6498424e95c5bbeffa30d19fd7ec57044da6683

File states

Total filesFindingUnallocatedAllocatedOutputtingProcessedNot foundFailed
50000000500000

Job states

TotalSubmittedStartedProcessingOutputtingFinishedNotusedAbortedStalledJobscript errorOutputting failedNone processed
65460000585904682001
Files processed0010010020020030030040040050050060060070070080080090090010001000Mar-06 23:00Mar-07 00:00Mar-07 01:00Mar-07 02:00Mar-07 03:00Mar-07 04:00Mar-07 05:00Mar-07 06:00Mar-07 07:00Mar-07 08:00Mar-07 09:00Mar-07 10:00Mar-07 11:00Mar-07 12:00Mar-07 13:00Mar-07 14:00Mar-07 15:00Mar-07 16:00Files processedBin start timesNumber per binCZ_FZUUK_GlasgowUS_FNAL-FermiG…US_FNAL-FermiGridUK_BristolIT_CNAFCERNFR_CCIN2P3ES_PICNL_SURFsaraNL_NIKHEFUK_RAL-Tier1UK_DurhamUS_ColoradoUK_BrunelUK_RAL-PPDUS_UChicagoUS_WisconsinUK_OxfordUS_UCSDUS_UConn-HPCCA_SFUUS_FNAL-T1UK_Liverpool

RSEs used

NameInputsOutputs
MONTECARLO55800
PRAGUE04327
SURFSARA06
RAL-PP05
RAL_ECHO04
DUNE_US_FNAL_DISK_STAGE03
NIKHEF02
DUNE_UK_GLASGOW01

Stats of processed input files as CSV or JSON, and of uploaded output files as CSV or JSON (up to 10000 files included)

File reset events, by site

SiteAllocatedOutputting
NL_SURFsara11213
CERN640
UK_Liverpool590
NL_NIKHEF5412
ES_PIC331
US_UCSD300
US_PuertoRico280
UK_RAL-PPD231
CZ_FZU210
UK_Glasgow185
US_FNAL-FermiGrid1711
US_UConn-HPC150
UK_Durham103
US_UChicago82
IT_CNAF50
US_Colorado51
UK_RAL-Tier157
CA_SFU30
UK_Brunel30
UK_Oxford11
US_Wisconsin12
UK_Bristol06

Jobscript

#!/bin/bash
#
source /cvmfs/dune.opensciencegrid.org/products/dune/setup_dune.sh
setup metacat
export METACAT_SERVER_URL=https://metacat.fnal.gov:9443/dune_meta_prod/app
export METACAT_AUTH_SERVER_URL=https://metacat.fnal.gov:8143/auth/dune

#Setup recent lar software suite
setup dunesw \
   "${DUNE_VERSION:-v09_91_04d00}" \
   -q "${DUNE_QUALIFIER:-e26:prof}"
echo "printing env"


if [ -z ${JUSTIN_PROCESSORS} ]; then
  JUSTIN_PROCESSORS=1
fi

echo "Justin processors: ${JUSTIN_PROCESSORS}"

export TF_NUM_THREADS=${JUSTIN_PROCESSORS}   
export OPENBLAS_NUM_THREADS=${JUSTIN_PROCESSORS} 
export JULIA_NUM_THREADS=${JUSTIN_PROCESSORS} 
export MKL_NUM_THREADS=${JUSTIN_PROCESSORS} 
export NUMEXPR_NUM_THREADS=${JUSTIN_PROCESSORS} 
export OMP_NUM_THREADS=${JUSTIN_PROCESSORS}  

#env

echo "Will use justin-get-file"
DID_PFN_RSE=`$JUSTIN_PATH/justin-get-file`
if [ "${DID_PFN_RSE}" == "" ] ; then
  echo "Could not get file"
  exit 0
fi
pfn=`echo ${DID_PFN_RSE} | cut -f2 -d' '` 

if [ -z ${LINENUMBER} ] ; then
  LINENUMBER=$pfn
fi

# echo "linenumber " $LINENUMBER
ex_code=0
pstep=10
tstep=0

now=$(date -u +"%Y%m%dT%H%M%SZ")
namespace=${JUSTIN_SCOPE:-"usertests"} 

echo "===============JUSTIN_JOBSUB_ID"
runid=$JUSTIN_WORKFLOW_ID
CLUSTER=`echo $JUSTIN_JOBSUB_ID | awk '{split($0,a,"."); print a[1]}'`
echo $CLUSTER

# define run number and set number of events
firstsubrun=0
#nevts=1
nevts=50
e_pfn=$(echo "$pfn" | sed 's/^0*//')
start_e_pfn=$((e_pfn -1))
firstevent=$((start_e_pfn * nevts + 1))

echo "======checking===="
echo ${DID_PFN_RSE}
echo ${pfn}
echo $runid
echo $firstsubrun
echo $nevts
echo $firstevent


# Run Generator
echo "============generator=========================" 
prodname="prodbackground_radiological_decay0_dune10kt_1x2x6_lateralAPA_${now}_gen_${pfn}" 
istep=1
tstep=$((tstep+1))
nstep=$(($pstep*$tstep)) 
tcode=$((nstep+istep))
tcode=$((tcode*1000)) 

# lar -c prodbackground_radiological_decay0_dune10kt_1x2x6_lateralAPA.fcl -o ${prodname}.root -n 50 
lar -c prodbackground_radiological_decay0_dune10kt_1x2x6_lateralAPA.fcl -o ${prodname}.root  -n $nevts -e ${runid}:${firstsubrun}:${firstevent}

exit_code=$?  
ex_code=$((exit_code+tcode))
files=`ls *_${now}_*`  
if [ $exit_code -ne 0 ]; then
 echo "ERROR: lar (generation) exit code: $ex_code " 
 echo "output files size: "
     for f in $files 
      do
       size=`stat -c %s $f`
       echo   $f $size 
      done   
 return $ex_code 
fi
 
# Stage 1 G4
echo "============G4 stage1=========================" 
g4_name="${prodname}_supernova_g4_halfActiveVol"
tstep=$((tstep+1))
nstep=$(($pstep*$tstep)) 
istep=1 
tcode=$((nstep+istep))
tcode=$((tcode*1000))

lar -c supernova_g4_halfActiveVol_dune10kt_1x2x6.fcl  ${prodname}.root -o ${g4_name}.root -n -1

exit_code=$?    
ex_code=$((exit_code+tcode))
files=`ls *_${now}_*`  
if [ $exit_code -ne 0 ]; then
  echo "ERROR: lar (geant4 step1) exit code: $ex_code " 
   echo "output files size: "
     for f in $files 
      do
       size=`stat -c %s $f`
       echo   $f $size 
      done   
  return $ex_code 
fi
 

# Stage 2 G4  
echo "============G4 stage2=========================" 
tstep=$((tstep+1))
nstep=$(($pstep*$tstep)) 
istep=1 
tcode=$((nstep+istep))
tcode=$((tcode*1000))
 
 
# Detsim
echo "============detsim========================="  
detsim_name="${g4_name}_detsim"
tstep=$((tstep+1))
nstep=$(($pstep*$tstep)) 
istep=1  
tcode=$((nstep+istep))
tcode=$((tcode*1000))

lar -c standard_detsim_dune10kt_1x2x6.fcl ${g4_name}.root  -o ${detsim_name}.root -n -1

exit_code=$?    
ex_code=$((exit_code+tcode))
files=`ls *_${now}_*`  
if [ $exit_code -ne 0 ]; then 
  echo "ERROR: lar (detsim) exit code: $ex_code " 
  echo "output files size: "
     for f in $files 
      do
       size=`stat -c %s $f`
       echo   $f $size 
      done   
  return $ex_code 
fi 
 
 
# Reco
echo "============reco ========================="  
reco_name="${detsim_name}_reco"
tstep=$((tstep+1))
nstep=$(($pstep*$tstep)) 
istep=1 
tcode=$((nstep+istep))
tcode=$((tcode*1000)) 

lar -c reco1_supernova_dune10kt_1x2x6.fcl ${detsim_name}.root -o ${reco_name}.root -n -1


exit_code=$?    
ex_code=$((exit_code+tcode))
files=`ls *_${now}_*`  
if [ $exit_code -ne 0 ]; then
  echo "ERROR: lar (reco) exit code: $ex_code " 
  echo "output files size: "
     for f in $files 
      do
       size=`stat -c %s $f`
       echo   $f $size 
      done   
  return $ex_code 
fi 

istep=$((istep+1))
tcode=$((nstep+istep))
tcode=$((tcode*1000))  
extractor_prod.py --infile ${reco_name}.root --no_crc    --appfamily art --appname reco --appversion  v09_91_04d00  --requestid ritm2205749  --strip_parents --input_json ${INPUT_DIR}/le_input.json> ${reco_name}.root.ext.json  && sed -i -e 's/stepfcl/reco1_supernova_dune10kt_1x2x6.fcl/g' ${reco_name}.root.ext.json 

exit_code=$?   
ex_code=$((exit_code+tcode))
files=`ls *_${now}_*`  
if [ $exit_code -gt 1 ]; then 
  echo "ERROR: metadata generation   $ex_code " 
   echo "output files size: "
     for f in $files 
      do
       size=`stat -c %s $f`
       echo   $f $size 
      done   
 return $ex_code 
fi 
 

rm -fr all-input-dids.txt
echo "noparents:noparents.root" > all-input-dids.txt 
python ${INPUT_DIR}/pdjson2metadata ${reco_name}.root.ext.json  all-input-dids.txt usertests > ${reco_name}.root.temp.json 
sed '/DUNE/ s/.*/\L&/' ${reco_name}.root.temp.json > ${reco_name}.root.json
 
exit_code=$?    
ex_code=$((exit_code+tcode))
files=`ls *_${now}_*` 
if [ $exit_code -ne 0 ]; then
  echo "ERROR: metadata writing  $ex_code " 
   echo "output files size: "
     for f in $files 
      do
       size=`stat -c %s $f`
       echo   $f $size 
      done   
  return $ex_code 
fi  


if [ $? -ne 0 ]
then
  echo "Exiting with error"
  return 1
else
  files=`ls *_${now}_*` 
   for f in $files 
      do
       size=`stat -c %s $f`
       echo "written output file: $f $size"
      done  

  echo "$pfn" > justin-processed-pfns.txt
fi
justIN time: 2025-04-03 07:34:33 UTC       justIN version: 01.03.00