Workflow 9631, Stage 1
| Priority | 50 |
| Processors | 1 |
| Wall seconds | 80000 |
| Image | /cvmfs/singularity.opensciencegrid.org/fermilab/fnal-wn-sl7:latest |
| RSS bytes | 6815744000 (6500 MiB) |
| Max distance for inputs | 30.0 |
| Enabled input RSEs |
CERN_PDUNE_EOS, DUNE_CA_SFU, DUNE_CERN_EOS, DUNE_ES_PIC, DUNE_FR_CCIN2P3_DISK, DUNE_IN_TIFR, DUNE_IT_INFN_CNAF, DUNE_UK_GLASGOW, DUNE_UK_LANCASTER_CEPH, DUNE_UK_MANCHESTER_CEPH, DUNE_US_BNL_SDCC, DUNE_US_FNAL_DISK_STAGE, FNAL_DCACHE, FNAL_DCACHE_STAGING, FNAL_DCACHE_TEST, MONTECARLO, NIKHEF, PRAGUE, QMUL, RAL-PP, RAL_ECHO, SURFSARA, T3_US_NERSC |
| Enabled output RSEs |
CERN_PDUNE_EOS, DUNE_CA_SFU, DUNE_CERN_EOS, DUNE_ES_PIC, DUNE_FR_CCIN2P3_DISK, DUNE_IN_TIFR, DUNE_IT_INFN_CNAF, DUNE_UK_GLASGOW, DUNE_UK_LANCASTER_CEPH, DUNE_UK_MANCHESTER_CEPH, DUNE_US_BNL_SDCC, DUNE_US_FNAL_DISK_STAGE, FNAL_DCACHE, FNAL_DCACHE_STAGING, FNAL_DCACHE_TEST, NIKHEF, PRAGUE, QMUL, RAL-PP, RAL_ECHO, SURFSARA, T3_US_NERSC |
| Enabled sites |
BR_CBPF, CA_SFU, CERN, CH_UNIBE-LHEP, CZ_FZU, ES_CIEMAT, ES_PIC, FR_CCIN2P3, IT_CNAF, NL_NIKHEF, NL_SURFsara, UK_Bristol, UK_Brunel, UK_Durham, UK_Edinburgh, UK_Lancaster, UK_Liverpool, UK_Manchester, UK_Oxford, UK_QMUL, UK_RAL-PPD, UK_RAL-Tier1, UK_Sheffield, US_Colorado, US_FNAL-FermiGrid, US_FNAL-T1, US_Michigan, US_PuertoRico, US_SU-ITS, US_Swan, US_UChicago, US_UConn-HPC, US_UCSD, US_Wisconsin |
| Scope | usertests |
| Events for this stage |
Output patterns
| | Destination | Pattern | Lifetime | For next stage | RSE expression |
|---|
| 1 | https://fndcadoor.fnal.gov:2880/dune/scratch/users/ykermaid/fnal/09631/1 | *_ana.root | | | |
Environment variables
| Name | Value |
|---|
| FCL_FILE | runSingleHit_PDVD.fcl |
| NTUPLE_DIR | /cvmfs/fifeuser1.opensciencegrid.org/sw/dune/f7cf64ee30ed1538f19cf08c7a2134cbc2c64147 |
Condor Class Ads
| Name | Value |
|---|
| HAS_CVMFS_dune_osgstorage_org | true |
File states
| Total files | Finding | Unallocated | Allocated | Outputting | Processed | Not found | Failed |
|---|
|
| 10 | 0 | 10 | 0 | 0 | 0 | 0 | 0 |
Job states
| Total | Submitted | Started | Processing | Outputting | Finished | Notused | Aborted | Stalled | Jobscript error | Outputting failed | None processed |
|---|
| 76 | 0 | 0 | 0 | 0 | 40 | 0 | 0 | 4 | 0 | 0 | 32 |
RSEs used
| Name | Inputs | Outputs |
|---|
| DUNE_US_FNAL_DISK_STAGE | 36 | 0 |
| None | 0 | 32 |
Stats of processed input files as CSV or JSON, and of uploaded output files as CSV or JSON (up to 10000 files included)
File reset events, by site
| Site | Allocated | Outputting |
|---|
| CA_SFU | 4 | 0 |
Jobscript
#!/bin/bash
#
#These must be defined
if [ -z $FCL_FILE ]; then
echo "Fatal Must provide FCL_FILE env var"
exit 1
fi
if [ -z $NTUPLE_DIR ]; then
echo "Fatal Must provide NTUPLE_DIR env var"
exit 1
fi
stat ${NTUPLE_DIR}
if [ $? -ne 0 ]; then
echo "Failed to $NTUPLE_DIR. Exiting safely"
exit 0
fi
source /cvmfs/dune.opensciencegrid.org/products/dune/setup_dune.sh
setup metacat
export METACAT_SERVER_URL=https://metacat.fnal.gov:9443/dune_meta_prod/app
export METACAT_AUTH_SERVER_URL=https://metacat.fnal.gov:8143/auth/dune
#Setup recent lar software suite
DUNE_VERSION=${DUNE_VERSION:-v10_12_01d00}
setup dunesw \
"${DUNE_VERSION}" \
-q "${DUNE_QUALIFIER:-e26:prof}"
setup_exit=$?
if [ $? -ne 0 ]; then
echo "Failed to setup dunesw $DUNE_VERSION $DUNE_QUALIFIER"
exit $setup_exit
fi
echo "DUNESW loc:"
ups active | grep dunesw
if [ -z ${JUSTIN_PROCESSORS} ]; then
JUSTIN_PROCESSORS=1
fi
echo "Justin processors: ${JUSTIN_PROCESSORS}"
export TF_NUM_THREADS=${JUSTIN_PROCESSORS}
export OPENBLAS_NUM_THREADS=${JUSTIN_PROCESSORS}
export JULIA_NUM_THREADS=${JUSTIN_PROCESSORS}
export MKL_NUM_THREADS=${JUSTIN_PROCESSORS}
export NUMEXPR_NUM_THREADS=${JUSTIN_PROCESSORS}
export OMP_NUM_THREADS=${JUSTIN_PROCESSORS}
echo "printing env"
env
echo "Justin specific env vars"
env | grep JUSTIN
echo "Will use justin-get-file"
DID_PFN_RSE=`$JUSTIN_PATH/justin-get-file`
##Check that any file was returned
if [ "${DID_PFN_RSE}" == "" ] ; then
echo "Could not get file"
exit 0
fi
pfn=`echo ${DID_PFN_RSE} | cut -f2 -d' '`
did=`echo ${DID_PFN_RSE} | cut -f1 -d' '`
echo "pfn: ${pfn}"
echo "did: ${did}"
now=$(date -u +"%Y%m%dT%H%M%SZ")
##TODO -- edit this
#jobid=`echo "${JUSTIN_JOBSUB_ID:-1}" | awk -F '.' '{print $1}'`
jobid=`echo "${JUSTIN_JOBSUB_ID:-1}" | cut -f1 -d'@' | sed -e "s/\./_/"`
PREFIX=$(basename ${pfn} .root)
OUTFILE="${PREFIX}_${now}_ana.root"
echo "Output file: ${OUTFILE}"
nevents=${NEVENTS:--1}
echo "Running singlehit"
touch ana.log
starttime=`date +"%s"`.0
lar -c ${FCL_FILE} \
-n ${nevents} \
-T ${OUTFILE} \
${pfn} >ana.log 2>&1
#-s ${pfn_list} >ana.log 2>&1
larExit=$?
endtime=`date +"%s"`.0
if [ $larExit -ne 0 ]; then
echo "Error in ntuple production"
cat ana.log
exit $larExit
fi
echo "Ran successfully"
#echo "Forming metadata"
### Get the metadata
#python $NTUPLE_DIR/ntuple_prod_utils.py \
# metadata \
# --root_file ${OUTFILE} \
# --dids $NTUPLE_DIR/input_dids_${KEY}.list \
# --version ${DUNE_VERSION} \
# --fcl_name ${FCL_FILE} \
# --tree_name ${TREE_NAME:-"r4/ClusterTree"} \
# -o $OUTFILE.json
# #--log_file md.log \
#mdExit=$?
#
#if [ $mdExit -ne 0 ]; then
# echo "Error in ntuple metdata production"
# cat md.log
# exit $mdExit
#fi
#
#echo "formed"
#cat ${OUTFILE}.json
echo "$pfn_list" > justin-processed-pfns.txt