justIN           Dashboard       Workflows       Jobs       AWT       Sites       Storages       Docs       Login

21 July 2025: This instance at RAL is read-only. Please do not try submitting new workflows for now.

Workflow 7773, Stage 1

Priority50
Processors1
Wall seconds80000
Image/cvmfs/singularity.opensciencegrid.org/fermilab/fnal-wn-sl7:latest
RSS bytes6815744000 (6500 MiB)
Max distance for inputs100.0
Enabled input RSEs CERN_PDUNE_EOS, DUNE_CA_SFU, DUNE_CERN_EOS, DUNE_ES_PIC, DUNE_FR_CCIN2P3_DISK, DUNE_IN_TIFR, DUNE_IT_INFN_CNAF, DUNE_UK_GLASGOW, DUNE_UK_LANCASTER_CEPH, DUNE_UK_MANCHESTER_CEPH, DUNE_US_BNL_SDCC, DUNE_US_FNAL_DISK_STAGE, FNAL_DCACHE, FNAL_DCACHE_STAGING, FNAL_DCACHE_TEST, MONTECARLO, NIKHEF, PRAGUE, QMUL, RAL-PP, RAL_ECHO, SURFSARA, T3_US_NERSC
Enabled output RSEs CERN_PDUNE_EOS, DUNE_CA_SFU, DUNE_CERN_EOS, DUNE_ES_PIC, DUNE_FR_CCIN2P3_DISK, DUNE_IN_TIFR, DUNE_IT_INFN_CNAF, DUNE_UK_GLASGOW, DUNE_UK_LANCASTER_CEPH, DUNE_UK_MANCHESTER_CEPH, DUNE_US_BNL_SDCC, DUNE_US_FNAL_DISK_STAGE, FNAL_DCACHE, FNAL_DCACHE_STAGING, FNAL_DCACHE_TEST, NIKHEF, PRAGUE, QMUL, RAL-PP, RAL_ECHO, SURFSARA, T3_US_NERSC
Enabled sites BR_CBPF, CA_Victoria, CERN, CH_UNIBE-LHEP, CZ_FZU, ES_CIEMAT, ES_PIC, FR_CCIN2P3, IN_TIFR, IT_CNAF, NL_NIKHEF, NL_SURFsara, UK_Bristol, UK_Brunel, UK_Durham, UK_Edinburgh, UK_Lancaster, UK_Manchester, UK_Oxford, UK_QMUL, UK_RAL-PPD, UK_RAL-Tier1, UK_Sheffield, US_Caltech, US_Colorado, US_FNAL-FermiGrid, US_FNAL-T1, US_Michigan, US_MIT, US_Nebraska, US_NotreDame, US_PuertoRico, US_SU-ITS, US_Swan, US_UChicago, US_UConn-HPC, US_UCSD, US_Wisconsin
Scopeusertests
Events for this stage

Output patterns

 DestinationPatternLifetimeFor next stageRSE expression
1Rucio usertests:pdhd-LE-test-w7773s1p1prod_*_reco_*.root2592000False

Environment variables

NameValue
CAMPAIGNpdhd-LE-test
DUNESW_TAR/cvmfs/fifeuser2.opensciencegrid.org/sw/dune/617fd47779cef0b5c77acdc4f87dee57a40c3e6b
JSONFILEpdhd_LE_test.json
MCJob1
NEVENTS100
pipyaml1
UTIL_TAR/cvmfs/fifeuser1.opensciencegrid.org/sw/dune/61faa6339f82a4480b0e4f4ee9f5b4919f8f029e
YAMLFILEpdhd_radiologicals_pds_gen.yaml

File states

Total filesFindingUnallocatedAllocatedOutputtingProcessedNot foundFailed
10000009406

Job states

TotalSubmittedStartedProcessingOutputtingFinishedNotusedAbortedStalledJobscript errorOutputting failedNone processed
511000028900212037
Files processed0010102020303040405050Jun-19 07:00Jun-19 17:00Jun-20 03:00Jun-20 13:00Jun-20 23:00Jun-21 09:00Jun-21 19:00Jun-22 05:00Jun-22 15:00Jun-23 01:00Jun-23 11:00Jun-23 21:00Jun-24 07:00Jun-24 17:00Jun-25 03:00Jun-25 13:00Jun-25 23:00Jun-26 09:00Jun-26 19:00Jun-27 05:00Jun-27 15:00Files processedBin start timesNumber per binUK_OxfordCERNES_PICUK_DurhamCZ_FZUUK_RAL-Tier1UK_RAL-PPDUK_SheffieldUK_ManchesterUK_QMULUK_BrunelFR_CCIN2P3

RSEs used

NameInputsOutputs
MONTECARLO2510
DUNE_UK_GLASGOW080
RAL_ECHO051
DUNE_FR_CCIN2P3_DISK042
PRAGUE036
DUNE_UK_MANCHESTER_CEPH028
DUNE_CERN_EOS018
RAL-PP015
QMUL012

Stats of processed input files as CSV or JSON, and of uploaded output files as CSV or JSON (up to 10000 files included)

File reset events, by site

SiteAllocatedOutputting
CZ_FZU280
UK_Manchester230
UK_Sheffield210
UK_RAL-Tier1200
UK_RAL-PPD140
ES_PIC131
FR_CCIN2P3120
UK_QMUL60
UK_Durham40
US_UChicago40
CERN30
UK_Brunel10
UK_Lancaster10

Jobscript

#!/bin/bash
#

#These must be defined
#if [ -z $FCL_FILE ]; then
#  echo "Fatal Must provide FCL_FILE env var"
#  exit 1
#fi

if [ -z $UTIL_TAR ]; then
  echo "Fatal must provide UTIL_TAR env var"
  exit 1
fi
export PYTHONPATH=$UTIL_TAR:$PYTHONPATH

source /cvmfs/dune.opensciencegrid.org/products/dune/setup_dune.sh
setup metacat
export METACAT_SERVER_URL=https://metacat.fnal.gov:9443/dune_meta_prod/app
export METACAT_AUTH_SERVER_URL=https://metacat.fnal.gov:8143/auth/dune

if [ -z ${JUSTIN_PROCESSORS} ]; then
  JUSTIN_PROCESSORS=1
fi

echo "Justin processors: ${JUSTIN_PROCESSORS}"

export TF_NUM_THREADS=${JUSTIN_PROCESSORS}   
export OPENBLAS_NUM_THREADS=${JUSTIN_PROCESSORS} 
export JULIA_NUM_THREADS=${JUSTIN_PROCESSORS} 
export MKL_NUM_THREADS=${JUSTIN_PROCESSORS} 
export NUMEXPR_NUM_THREADS=${JUSTIN_PROCESSORS} 
export OMP_NUM_THREADS=${JUSTIN_PROCESSORS}  

echo "Justin specific env vars"
env | grep JUSTIN
now=$(date -u +"%Y%m%dT%H%M%SZ")
jobid=`echo "${JUSTIN_JOBSUB_ID:-1}" | cut -f1 -d'@' | sed -e "s/\./_/"`
stageid=${JUSTIN_STAGE_ID:-1}
a=1
 if [ -n "$DUNESW_TAR" ]; then
   stat ${DUNESW_TAR}
   if [ $? -ne 0 ]; then
     echo "Failed to stat $DUNESW_TAR. Exiting safely"
     exit 0
   fi
 
   export PRODUCTS=$DUNESW_TAR:$PRODUCTS 
   echo "Set dunesw to $DUNESW_TAR"
 fi

 if [ -n "$WIRECELL_TAR" ]; then
   stat ${WIRECELL_TAR}
   if [ $? -ne 0 ]; then
     echo "Failed to stat $WIRECELL_TAR. Exiting safely"
     exit 0
   fi
 
   export PRODUCTS=$WIRECELL_TAR:$PRODUCTS 
   echo "Set wirecell to $WIRECELL_TAR"
 fi

 #Setup recent lar software suite
 DUNESW_VERSION=${DUNESW_VERSION:-v10_06_00d01}
 setup dunesw \
    "${DUNESW_VERSION}" \
    -q "${DUNE_QUALIFIER:-e26:prof}"
 
 setup_exit=$?
 if [ $? -ne 0 ]; then
   echo "Failed to setup dunesw $DUNESW_VERSION $DUNE_QUALIFIER"
   exit $setup_exit
 fi

 echo "DUNESW loc:"
 ups active | grep dunesw

 echo "WIRECELL loc:"
 ups active | grep wirecell
 
export FHICL_FILE_PATH=${DUNESW_TAR}:${FHICL_FILE_PATH}
 
 
 echo "printing env"
 env
 echo "DUNE specific env vars"
 env | grep DUNE

# User can specify to pip install pyyaml
if [ -n "$pipyaml" ]; then
  echo "Installing yaml"
  pip install --user pyyaml
  echo "Done"
# elif [ -n "${YAML_TAR}" ]; then
#   echo "Using venv at ${YAML_TAR}"
#   source ${YAML_TAR}/yaml_venv/bin/activate
fi


 echo "Will use justin-get-file"
 DID_PFN_RSE=`$JUSTIN_PATH/justin-get-file`
 if [ "${DID_PFN_RSE}" == "" ] ; then
   echo "Could not get file"
   exit 0
 fi
 pfn=`echo ${DID_PFN_RSE} | cut -f2 -d' '`
 did=`echo ${DID_PFN_RSE} | cut -f1 -d' '`
 
 input_filename=`echo $did | cut -f2 -d':'`
 echo "input file: $input_filename"

 echo "jobsub_id: ${JUSTIN_JOBSUB_ID:-1}"
 
 #Get run/subrun from parent if using input -- else from justin job ID/process"
 echo "MCJob: $MCJob"
 if [ -z "$MCJob" ]; then
   echo "Getting run subrun from $did"
   run_subrun=`python -m beam_job_utils get_run_subrun -i $did`
   get_run_ret=$?
   if [ $get_run_ret -ne 0 ]; then
     echo "error in get_run_subrun"
   fi
 
   echo $run_subrun
   run=`echo $run_subrun | cut -f1 -d' '`
   subrun=`echo $run_subrun | cut -f2 -d' '`
 else
   run=`echo $jobid | cut -f1 -d'_'`  ##TODO -- check order here
   subrun=`echo $jobid | cut -f2 -d'_'`
 fi
 nevents=${NEVENTS:--1} #Set from env override or -1 
 NSKIP=${NSKIP:-0} 
 echo "nevents: $nevents"
 
 overrides="core.data_tier=full-reconstructed \
  core.application.version=${DUNESW_VERSION} \
  core.application.name=reco \
  core.application.family=dunesw \
  core.application=dunesw.reco \
 "

#User specifies if they'd like to use an input file or not 
if [ -n "$MCJob" ]; then
  inputflag=""
else
  inputflag="-i $pfn --parent $did"
fi

echo "input flag: $inputflag"

echo "Will execute:"
echo "python -m beam_job_utils run_job \
      --json ${UTIL_TAR}/${JSONFILE} \
      --overrides ${overrides} \
      --run $run \
      --subrun $subrun \
      --yaml ${UTIL_TAR}/${YAMLFILE} \
      --nevents $nevents \
      --event 0 \
      $inputflag"

starttime=`date +"%s"`.0

# TODO -- metadata file configurable
python -m beam_job_utils run_job \
      --json ${UTIL_TAR}/${JSONFILE} \
      --overrides ${overrides} \
      --run $run \
      --subrun $subrun \
      --yaml ${UTIL_TAR}/${YAMLFILE} \
      --nevents $nevents \
      --event 0 \
      $inputflag

endtime=`date +"%s"`.0

#output_g4_files=`ls *_g4_stage2_*.root`
output_reco_files=`ls *_reco_*.root`
#output_ana_files=`ls *_ana_*.root`

echo "Output files:"
#echo "   G4: ${output_g4_files}"
echo "   Reco: ${output_reco_files}"
#echo "   Ana: ${output_ana_files}"


#echo "Forming g4 metadata"
#for FILE in ${output_g4_files[@]}
#do
#	echo "  G4 file: $FILE"
#
#	#Retrieve the key in output filename that corresponds to YAML config. stage
#        IFS=_ read -ra substrs <<< $(echo ${FILE} | cut -d '.' -f 1)
#	key_found='false'
#	KEY='g4_stage2_'
#	iter=0
#	declare -i nsub=${#substrs[@]}-1
#	for substr in ${substrs[@]}; do
#        	if [ $key_found == 'true' ]; then
#                	KEY+=$substr
#                	if [ $iter -lt ${nsub} ]; then
#                        	KEY+='_'
#                	fi      
#        	fi
#        	if [ $substr == 'stage2' ]; then
#                	key_found='true'
#        	fi      
#        	let iter=$iter+1
#	done    
#	echo "    KEY: ${KEY}"
##
#	#Retrieve the FHICL file that corresponds to the output file key
#	FCL=$(grep ${UTIL_TAR}/${YAMLFILE} -e "${KEY}:" -A 1 | grep -e "fcl" | cut -d ':' -f 2)
#        echo "    FCL: ${FCL}"
#
#	python -m meta_maker --start_time $starttime --end_time $endtime --file_format "artroot" \
#                     --app_family "dunesw" --app_name "reco" --app_version ${DUNESW_VERSION} \
#                     --data_tier "detector-simulated" --get_events \
#                     --campaign $CAMPAIGN \
#                     --runs 1 --subruns ${subrun} --fcl $FCL \
#                     -f "${JUSTIN_SCOPE}:${FILE}" -j "${FILE}.json"
#	if [ $? -ne 0 ]; then
#	  echo "  Error in g4 metadata"
#	  exit 1
#	fi
#        echo "  Ran successfully"
#        cat ${FILE}.json
#done

subrun=1

echo "Forming reco metadata"
for FILE in ${output_reco_files[@]}
do
	echo "  subrun: ${subrun} -> reco file: $FILE"

	#Retrieve the key in output filename that corresponds to YAML config. stage
        IFS=_ read -ra substrs <<< $(echo ${FILE} | cut -d '.' -f 1)
        key_found='false'
        KEY='reco_'
        iter=0
        declare -i nsub=${#substrs[@]}-1
        for substr in ${substrs[@]}; do
                if [ $key_found == 'true' ]; then
                        KEY+=$substr
                        if [ $iter -lt ${nsub} ]; then
                                KEY+='_'
                        fi
                fi
                if [ $substr == 'reco' ]; then
                        key_found='true'
                fi
                let iter=$iter+1
        done
        echo "    KEY: ${KEY}"

        #Retrieve the FHICL file that corresponds to the output file key
        FCL=$(grep ${UTIL_TAR}/${YAMLFILE} -e "${KEY}:" -A 1 | grep -e "fcl" | cut -d ':' -f 2)
        echo "    FCL: ${FCL}"

	python -m meta_maker --start_time $starttime --end_time $endtime --file_format "artroot" \
                     --app_family "dunesw" --app_name "reco" --app_version ${DUNESW_VERSION} \
                     --data_tier "full-reconstructed" --get_events \
                     --campaign $CAMPAIGN --runs 1 --subruns ${subrun} \
                     --fcl $FCL \
                     -f "${JUSTIN_SCOPE}:${FILE}" -j "${FILE}.json"
	if [ $? -ne 0 ]; then
	  echo "  Error in reco metadata"
	  exit 1
	fi
        echo "  Ran successfully"
        cat ${FILE}.json
        let subrun=${subrun}+1
done


#subrun=1
#echo "Forming ana metadata"
#for FILE in ${output_ana_files[@]}
#do
#	echo "  subrun: ${subrun} -> ana file: $FILE"
#        #Retrieve the key in output filename that corresponds to YAML config. stage
#        IFS=_ read -ra substrs <<< $(echo ${FILE} | cut -d '.' -f 1)
#        key_found='false'
#        KEY='ana_'
#        iter=0
#        declare -i nsub=${#substrs[@]}-2
#        for substr in ${substrs[@]}; do
#                if [ $key_found == 'true' ] && [ $substr != 'tfile' ]; then
#                        KEY+=$substr
#                        if [ $iter -lt ${nsub} ]; then
#                                KEY+='_'
#                        fi
#                fi
#                if [ $substr == 'ana' ]; then
#                        key_found='true'
#                fi
#                let iter=$iter+1
#        done
#
#	echo "    KEY: ${KEY}"
#        FCL=$(grep ${UTIL_TAR}/${YAMLFILE} -e "${KEY}:" -A 1 | grep -e "fcl" | cut -d ':' -f 2)
#        echo "    FCL: ${FCL}"
#
#	python -m meta_maker --start_time $starttime --end_time $endtime --file_format "tfile" \
#                     --app_family "dunesw" --app_name "reco" --app_version ${DUNESW_VERSION} \
#                     --data_tier "root-tuple" \
#                     --campaign $CAMPAIGN --runs 1 --subruns ${subrun} \
#                     --fcl $FCL \
#                     -f "${JUSTIN_SCOPE}:${FILE}" -j "${FILE}.json"
#	if [ $? -ne 0 ]; then
#	  echo "  Error in ana metadata"
#	  exit 1
#	fi
#	echo "  Ran successfully"
#	cat ${FILE}.json
#	let subrun=${subrun}+1
#done

exitcode=$?
if [ $exitcode -ne 0 ]; then
  echo "Error running. Exiting with ${exitcode}"
  exit $exitcode
fi

echo "$pfn" > justin-processed-pfns.txt
justIN time: 2025-08-14 16:31:58 UTC       justIN version: 01.03.02