From 7df82e13326f00739f23cb64b57bc79141bcab56 Mon Sep 17 00:00:00 2001 From: RussTreadon-NOAA Date: Mon, 27 Nov 2023 21:51:43 +0000 Subject: [PATCH 1/2] initial commit for hercules GDASApp build (#773) --- build.sh | 2 +- modulefiles/GDAS/hercules.lua | 98 +++++++++++++++++++++++++++++++++++ ush/detect_machine.sh | 2 + ush/module-setup.sh | 7 +++ 4 files changed, 108 insertions(+), 1 deletion(-) create mode 100644 modulefiles/GDAS/hercules.lua diff --git a/build.sh b/build.sh index 448bd0a50..ec99bbe1e 100755 --- a/build.sh +++ b/build.sh @@ -70,7 +70,7 @@ while getopts "p:t:c:hvdfa" opt; do done case ${BUILD_TARGET} in - hera | orion) + hera | orion | hercules) echo "Building GDASApp on $BUILD_TARGET" source $dir_root/ush/module-setup.sh module use $dir_root/modulefiles diff --git a/modulefiles/GDAS/hercules.lua b/modulefiles/GDAS/hercules.lua new file mode 100644 index 000000000..e989af14a --- /dev/null +++ b/modulefiles/GDAS/hercules.lua @@ -0,0 +1,98 @@ +help([[ +Load environment for running the GDAS application with Intel compilers and MPI. +]]) + +local pkgName = myModuleName() +local pkgVersion = myModuleVersion() +local pkgNameVer = myModuleFullName() + +prepend_path("MODULEPATH", '/work/noaa/epic/role-epic/spack-stack/hercules/spack-stack-1.5.1/envs/unified-env/install/modulefiles/Core') +prepend_path("MODULEPATH", '/work2/noaa/da/python/opt/modulefiles/stack') + +-- below two lines get us access to the spack-stack modules +load("stack-intel/2021.9.0") +load("stack-intel-oneapi-mpi/2021.9.0") +--load("stack-python/3.10.8") +-- JCSDA has 'jedi-fv3-env/unified-dev', but we should load these manually as needed +load("cmake/3.23.1") +load("curl/8.1.2") +load("zlib/1.2.13") +load("git/2.40.0") +--load("pkg-config/0.27.1") +load("hdf5/1.14.0") +load("parallel-netcdf/1.12.2") +load("netcdf-c/4.9.2") +load("nccmp/1.9.0.1") +load("netcdf-fortran/4.6.0") +load("nco/5.0.6") +load("parallelio/2.5.10") +load("wget/1.21.1") +load("boost/1.78") +load("bufr/12.0.1") +load("git-lfs/3.1.2") +load("ecbuild/3.7.2") +load("openjpeg/2.4.0") +load("eccodes/2.27.0") +load("eigen/3.4.0") +load("openblas/0.3.19") +load("eckit/1.24.4") +load("fftw/3.3.10") +load("fckit/0.11.0") +load("fiat/1.2.0") +load("ectrans/1.2.0") +load("atlas/0.35.0") +load("sp/2.3.3") +load("gsl-lite/0.37.0") +load("libjpeg/2.1.0") +load("krb5/1.20.1") +load("libtirpc/1.2.6") +load("hdf/4.2.15") +load("jedi-cmake/1.4.0") +load("libpng/1.6.37") +load("libxt/1.1.5") +load("libxmu/1.1.4") +load("libxpm/3.5.12") +load("libxaw/1.0.13") +load("udunits/2.2.28") +load("ncview/2.1.8") +load("netcdf-cxx4/4.3.1") +load("py-pybind11/2.8.1") +--load("crtm/v2.4_jedi") +load("contrib/0.1") +load("noaatools/3.1") +load("rocoto/1.3.5") + +load("hpc/1.2.0") +unload("python/3.10.8") +unload("python/3.9.2") +load("miniconda3/4.6.14") +load("gdasapp/1.0.0") + +load("boost/1.78.0") +load("bufr/12.0.1") +load("fckit/0.11.0") +load("atlas/0.35.0") +load("py-pybind11/2.8.1") + +-- below is a hack because of cmake finding the wrong python... +setenv("CONDA_PREFIX", "/work2/noaa/da/python/opt/core/miniconda3/4.6.14/envs/gdasapp/") + +setenv("CC","mpiicc") +setenv("FC","mpiifort") +setenv("CXX","mpiicpc") +local mpiexec = '/opt/slurm/bin/srun' +local mpinproc = '-n' +setenv('MPIEXEC_EXEC', mpiexec) +setenv('MPIEXEC_NPROC', mpinproc) + +setenv('R2D2_CONFIG', '/work2/noaa/da/cmartin/GDASApp/R2D2_SHARED/config_orion.yaml') +setenv("CRTM_FIX","/work2/noaa/da/cmartin/GDASApp/fix/crtm/2.4.0") +setenv("GDASAPP_TESTDATA","/work2/noaa/da/cmartin/CI/GDASApp/data") +prepend_path("PATH","/apps/contrib/NCEP/libs/hpc-stack/intel-2018.4/prod_util/1.2.2/bin") + +execute{cmd="ulimit -s unlimited",modeA={"load"}} + +whatis("Name: ".. pkgName) +whatis("Version: ".. pkgVersion) +whatis("Category: GDASApp") +whatis("Description: Load all libraries needed for GDASApp") diff --git a/ush/detect_machine.sh b/ush/detect_machine.sh index 949526019..8610af15c 100755 --- a/ush/detect_machine.sh +++ b/ush/detect_machine.sh @@ -25,6 +25,8 @@ case $(hostname -f) in Orion-login-[1-4].HPC.MsState.Edu) MACHINE_ID=orion ;; ### orion1-4 + Hercules-login-[1-4].HPC.MsState.Edu) MACHINE_ID=hercules ;; ### hercules1-4 + cheyenne[1-6].cheyenne.ucar.edu) MACHINE_ID=cheyenne ;; ### cheyenne1-6 cheyenne[1-6].ib0.cheyenne.ucar.edu) MACHINE_ID=cheyenne ;; ### cheyenne1-6 chadmin[1-6].ib0.cheyenne.ucar.edu) MACHINE_ID=cheyenne ;; ### cheyenne1-6 diff --git a/ush/module-setup.sh b/ush/module-setup.sh index 469fd4a3c..ef7f73699 100755 --- a/ush/module-setup.sh +++ b/ush/module-setup.sh @@ -22,6 +22,13 @@ elif [[ $MACHINE_ID = orion* ]] ; then fi module purge +elif [[ $MACHINE_ID = hercules* ]] ; then + # We are on Hercules + if ( ! eval module help > /dev/null 2>&1 ) ; then + source /apps/lmod/init/bash + fi + module purge + elif [[ $MACHINE_ID = s4* ]] ; then # We are on SSEC Wisconsin S4 if ( ! eval module help > /dev/null 2>&1 ) ; then From 8e67f9baa514e59be76965f6132fd7b046f6df92 Mon Sep 17 00:00:00 2001 From: RussTreadon-NOAA Date: Tue, 9 Jan 2024 21:14:33 +0000 Subject: [PATCH 2/2] add hercules to atm and soca ctests (#773) --- test/atm/global-workflow/jjob_ens_final.sh | 10 +++++----- test/atm/global-workflow/jjob_ens_init.sh | 14 +++++++------- test/atm/global-workflow/jjob_ens_run.sh | 10 +++++----- test/atm/global-workflow/jjob_var_final.sh | 10 +++++----- test/atm/global-workflow/jjob_var_init.sh | 18 +++++++++--------- test/atm/global-workflow/jjob_var_run.sh | 10 +++++----- test/soca/gw/CMakeLists.txt | 11 ++++++++--- ush/soca/run_jjobs.py | 2 +- 8 files changed, 45 insertions(+), 40 deletions(-) diff --git a/test/atm/global-workflow/jjob_ens_final.sh b/test/atm/global-workflow/jjob_ens_final.sh index ae1928a18..f0a0a9117 100755 --- a/test/atm/global-workflow/jjob_ens_final.sh +++ b/test/atm/global-workflow/jjob_ens_final.sh @@ -33,11 +33,11 @@ export PYTHONPATH machine=$(echo `grep 'machine=' $EXPDIR/config.base | cut -d"=" -f2` | tr -d '"') # Set NETCDF and UTILROOT variables (used in config.base) -if [ $machine = 'HERA' ]; then +if [[ $machine = 'HERA' ]]; then NETCDF=$( which ncdump ) export NETCDF export UTILROOT="/scratch2/NCEPDEV/ensemble/save/Walter.Kolczynski/hpc-stack/intel-18.0.5.274/prod_util/1.2.2" -elif [ $machine = 'ORION' ]; then +elif [[ $machine = 'ORION' || $machine = 'HERCULES' ]]; then ncdump=$( which ncdump ) NETCDF=$( echo "${ncdump}" | cut -d " " -f 3 ) export NETCDF @@ -45,10 +45,10 @@ elif [ $machine = 'ORION' ]; then fi # Execute j-job -if [ $machine = 'HERA' ]; then +if [[ $machine = 'HERA' ]]; then + sbatch --ntasks=1 --account=$ACCOUNT --qos=batch --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATMENS_ANALYSIS_FINALIZE +elif [[ $machine = 'ORION' || $machine = 'HERCULES' ]]; then sbatch --ntasks=1 --account=$ACCOUNT --qos=batch --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATMENS_ANALYSIS_FINALIZE -elif [ $machine = 'ORION' ]; then - sbatch --ntasks=1 --account=$ACCOUNT --qos=batch --partition=orion --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATMENS_ANALYSIS_FINALIZE else ${HOMEgfs}/jobs/JGLOBAL_ATMENS_ANALYSIS_FINALIZE fi diff --git a/test/atm/global-workflow/jjob_ens_init.sh b/test/atm/global-workflow/jjob_ens_init.sh index 8966f2b69..e46ea90da 100755 --- a/test/atm/global-workflow/jjob_ens_init.sh +++ b/test/atm/global-workflow/jjob_ens_init.sh @@ -38,11 +38,11 @@ export PYTHONPATH machine=$(echo `grep 'machine=' $EXPDIR/config.base | cut -d"=" -f2` | tr -d '"') # Set NETCDF and UTILROOT variables (used in config.base) -if [ $machine = 'HERA' ]; then +if [[ $machine = 'HERA' ]]; then NETCDF=$( which ncdump ) export NETCDF export UTILROOT="/scratch2/NCEPDEV/ensemble/save/Walter.Kolczynski/hpc-stack/intel-18.0.5.274/prod_util/1.2.2" -elif [ $machine = 'ORION' ]; then +elif [[ $machine = 'ORION' || $machine = 'HERCULES' ]]; then ncdump=$( which ncdump ) NETCDF=$( echo "${ncdump}" | cut -d " " -f 3 ) export NETCDF @@ -69,7 +69,7 @@ dpath=gdas.$PDY/$cyc/obs mkdir -p $COM_OBS flist="amsua_n19.$CDATE.nc4 sondes.$CDATE.nc4" for file in $flist; do - ln -fs $GDASAPP_TESTDATA/lowres/$dpath/${oprefix}.$file $COM_OBS/ + ln -fs $GDASAPP_TESTDATA/lowres/$dpath/${oprefix}.$file $COM_OBS/${oprefix}.$file done # Link radiance bias correction files @@ -77,7 +77,7 @@ dpath=gdas.$gPDY/$gcyc/analysis/atmos mkdir -p $COM_ATMOS_ANALYSIS_PREV flist="amsua_n19.satbias.nc4 amsua_n19.satbias_cov.nc4 amsua_n19.tlapse.txt" for file in $flist; do - ln -fs $GDASAPP_TESTDATA/lowres/$dpath/$gprefix.$file $COM_ATMOS_ANALYSIS_PREV/ + ln -fs $GDASAPP_TESTDATA/lowres/$dpath/$gprefix.$file $COM_ATMOS_ANALYSIS_PREV/$gprefix.$file done # Link member atmospheric background on tiles and atmf006 @@ -104,10 +104,10 @@ for imem in $(seq 1 $NMEM_ENS); do done # Execute j-job -if [ $machine = 'HERA' ]; then +if [[ $machine = 'HERA' ]]; then + sbatch --ntasks=1 --account=$ACCOUNT --qos=batch --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATMENS_ANALYSIS_INITIALIZE +elif [[ $machine = 'ORION' || $machine = 'HERCULES' ]]; then sbatch --ntasks=1 --account=$ACCOUNT --qos=batch --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATMENS_ANALYSIS_INITIALIZE -elif [ $machine = 'ORION' ]; then - sbatch --ntasks=1 --account=$ACCOUNT --qos=batch --partition=orion --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATMENS_ANALYSIS_INITIALIZE else ${HOMEgfs}/jobs/JGLOBAL_ATMENS_ANALYSIS_INITIALIZE fi diff --git a/test/atm/global-workflow/jjob_ens_run.sh b/test/atm/global-workflow/jjob_ens_run.sh index 597afcd43..89bff09fe 100755 --- a/test/atm/global-workflow/jjob_ens_run.sh +++ b/test/atm/global-workflow/jjob_ens_run.sh @@ -36,11 +36,11 @@ export PYTHONPATH machine=$(echo `grep 'machine=' $EXPDIR/config.base | cut -d"=" -f2` | tr -d '"') # Set NETCDF and UTILROOT variables (used in config.base) -if [ $machine = 'HERA' ]; then +if [[ $machine = 'HERA' ]]; then NETCDF=$( which ncdump ) export NETCDF export UTILROOT="/scratch2/NCEPDEV/ensemble/save/Walter.Kolczynski/hpc-stack/intel-18.0.5.274/prod_util/1.2.2" -elif [ $machine = 'ORION' ]; then +elif [[ $machine = 'ORION' || $machine = 'HERCULES' ]]; then ncdump=$( which ncdump ) NETCDF=$( echo "${ncdump}" | cut -d " " -f 3 ) export NETCDF @@ -48,10 +48,10 @@ elif [ $machine = 'ORION' ]; then fi # Execute j-job -if [ $machine = 'HERA' ]; then +if [[ $machine = 'HERA' ]]; then + sbatch --nodes=1 --ntasks=36 --account=$ACCOUNT --qos=batch --time=00:30:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATMENS_ANALYSIS_RUN +elif [[ $machine = 'ORION' || $machine = 'HERCULES' ]]; then sbatch --nodes=1 --ntasks=36 --account=$ACCOUNT --qos=batch --time=00:30:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATMENS_ANALYSIS_RUN -elif [ $machine = 'ORION' ]; then - sbatch --nodes=1 --ntasks=36 --account=$ACCOUNT --qos=batch --partition=orion --time=00:30:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATMENS_ANALYSIS_RUN else ${HOMEgfs}/jobs/JGLOBAL_ATMENS_ANALYSIS_RUN fi diff --git a/test/atm/global-workflow/jjob_var_final.sh b/test/atm/global-workflow/jjob_var_final.sh index f9b64afbe..01b3c555f 100755 --- a/test/atm/global-workflow/jjob_var_final.sh +++ b/test/atm/global-workflow/jjob_var_final.sh @@ -33,11 +33,11 @@ export PYTHONPATH machine=$(echo `grep 'machine=' $EXPDIR/config.base | cut -d"=" -f2` | tr -d '"') # Set NETCDF and UTILROOT variables (used in config.base) -if [ $machine = 'HERA' ]; then +if [[ $machine = 'HERA' ]]; then NETCDF=$( which ncdump ) export NETCDF export UTILROOT="/scratch2/NCEPDEV/ensemble/save/Walter.Kolczynski/hpc-stack/intel-18.0.5.274/prod_util/1.2.2" -elif [ $machine = 'ORION' ]; then +elif [[ $machine = 'ORION' || $machine = 'HERCULES' ]]; then ncdump=$( which ncdump ) NETCDF=$( echo "${ncdump}" | cut -d " " -f 3 ) export NETCDF @@ -45,10 +45,10 @@ elif [ $machine = 'ORION' ]; then fi # Execute j-job -if [ $machine = 'HERA' ]; then +if [[ $machine = 'HERA' ]]; then + sbatch --ntasks=1 --account=$ACCOUNT --qos=batch --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATM_ANALYSIS_FINALIZE +elif [[ $machine = 'ORION' || $machine = 'HERCULES' ]]; then sbatch --ntasks=1 --account=$ACCOUNT --qos=batch --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATM_ANALYSIS_FINALIZE -elif [ $machine = 'ORION' ]; then - sbatch --ntasks=1 --account=$ACCOUNT --qos=batch --partition=orion --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATM_ANALYSIS_FINALIZE else ${HOMEgfs}/jobs/JGLOBAL_ATM_ANALYSIS_FINALIZE fi diff --git a/test/atm/global-workflow/jjob_var_init.sh b/test/atm/global-workflow/jjob_var_init.sh index 0d88c2174..6d304e0a8 100755 --- a/test/atm/global-workflow/jjob_var_init.sh +++ b/test/atm/global-workflow/jjob_var_init.sh @@ -38,11 +38,11 @@ export PYTHONPATH machine=$(echo `grep 'machine=' $EXPDIR/config.base | cut -d"=" -f2` | tr -d '"') # Set NETCDF and UTILROOT variables (used in config.base) -if [ $machine = 'HERA' ]; then +if [[ $machine = 'HERA' ]]; then NETCDF=$( which ncdump ) export NETCDF export UTILROOT="/scratch2/NCEPDEV/ensemble/save/Walter.Kolczynski/hpc-stack/intel-18.0.5.274/prod_util/1.2.2" -elif [ $machine = 'ORION' ]; then +elif [[ $machine = 'ORION' || $machine = 'HERCULES' ]]; then ncdump=$( which ncdump ) NETCDF=$( echo "${ncdump}" | cut -d " " -f 3 ) export NETCDF @@ -71,7 +71,7 @@ dpath=gdas.$PDY/$cyc/obs mkdir -p $COM_OBS flist="amsua_n19.$CDATE.nc4 sondes.$CDATE.nc4" for file in $flist; do - ln -fs $GDASAPP_TESTDATA/lowres/$dpath/${oprefix}.$file $COM_OBS/ + ln -fs $GDASAPP_TESTDATA/lowres/$dpath/${oprefix}.$file $COM_OBS/${oprefix}.$file done # Link radiance bias correction files @@ -79,7 +79,7 @@ dpath=gdas.$gPDY/$gcyc/analysis/atmos mkdir -p $COM_ATMOS_ANALYSIS_PREV flist="amsua_n19.satbias.nc4 amsua_n19.satbias_cov.nc4 amsua_n19.tlapse.txt" for file in $flist; do - ln -fs $GDASAPP_TESTDATA/lowres/$dpath/$gprefix.$file $COM_ATMOS_ANALYSIS_PREV/ + ln -fs $GDASAPP_TESTDATA/lowres/$dpath/$gprefix.$file $COM_ATMOS_ANALYSIS_PREV/$gprefix.$file done # Link atmospheric background on gaussian grid @@ -87,7 +87,7 @@ dpath=gdas.$gPDY/$gcyc/model_data/atmos/history mkdir -p $COM_ATMOS_HISTORY_PREV flist="atmf006.nc" for file in $flist; do - ln -fs $GDASAPP_TESTDATA/lowres/$dpath/${gprefix}.${file} $COM_ATMOS_HISTORY_PREV/ + ln -fs $GDASAPP_TESTDATA/lowres/$dpath/${gprefix}.${file} $COM_ATMOS_HISTORY_PREV/${gprefix}.${file} done # Link atmospheric bacgkround on tiles @@ -96,7 +96,7 @@ COM_ATMOS_RESTART_PREV_DIRNAME=$(dirname $COM_ATMOS_RESTART_PREV) mkdir -p $COM_ATMOS_RESTART_PREV_DIRNAME flist="restart" for file in $flist; do - ln -fs $GDASAPP_TESTDATA/lowres/$dpath/$file $COM_ATMOS_RESTART_PREV_DIRNAME/ + ln -fs $GDASAPP_TESTDATA/lowres/$dpath/$file $COM_ATMOS_RESTART_PREV_DIRNAME/$file done @@ -125,10 +125,10 @@ done # Execute j-job -if [ $machine = 'HERA' ]; then +if [[ $machine = 'HERA' ]]; then + sbatch --ntasks=1 --account=$ACCOUNT --qos=batch --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATM_ANALYSIS_INITIALIZE +elif [[ $machine = 'ORION' || $machine = 'HERCULES' ]]; then sbatch --ntasks=1 --account=$ACCOUNT --qos=batch --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATM_ANALYSIS_INITIALIZE -elif [ $machine = 'ORION' ]; then - sbatch --ntasks=1 --account=$ACCOUNT --qos=batch --partition=orion --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATM_ANALYSIS_INITIALIZE else ${HOMEgfs}/jobs/JGLOBAL_ATM_ANALYSIS_INITIALIZE fi diff --git a/test/atm/global-workflow/jjob_var_run.sh b/test/atm/global-workflow/jjob_var_run.sh index a78b1e10e..e13aa2daa 100755 --- a/test/atm/global-workflow/jjob_var_run.sh +++ b/test/atm/global-workflow/jjob_var_run.sh @@ -36,11 +36,11 @@ export PYTHONPATH machine=$(echo `grep 'machine=' $EXPDIR/config.base | cut -d"=" -f2` | tr -d '"') # Set NETCDF and UTILROOT variables (used in config.base) -if [ $machine = 'HERA' ]; then +if [[ $machine = 'HERA' ]]; then NETCDF=$( which ncdump ) export NETCDF export UTILROOT="/scratch2/NCEPDEV/ensemble/save/Walter.Kolczynski/hpc-stack/intel-18.0.5.274/prod_util/1.2.2" -elif [ $machine = 'ORION' ]; then +elif [[ $machine = 'ORION' || $machine = 'HERCULES' ]]; then ncdump=$( which ncdump ) NETCDF=$( echo "${ncdump}" | cut -d " " -f 3 ) export NETCDF @@ -48,10 +48,10 @@ elif [ $machine = 'ORION' ]; then fi # Execute j-job -if [ $machine = 'HERA' ]; then +if [[ $machine = 'HERA' ]]; then + sbatch --ntasks=6 --account=$ACCOUNT --qos=batch --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATM_ANALYSIS_RUN +elif [[ $machine = 'ORION' || $machine = 'HERCULES' ]]; then sbatch --ntasks=6 --account=$ACCOUNT --qos=batch --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATM_ANALYSIS_RUN -elif [ $machine = 'ORION' ]; then - sbatch --ntasks=6 --account=$ACCOUNT --qos=batch --partition=orion --time=00:10:00 --export=ALL --wait ${HOMEgfs}/jobs/JGLOBAL_ATM_ANALYSIS_RUN else ${HOMEgfs}/jobs/JGLOBAL_ATM_ANALYSIS_RUN fi diff --git a/test/soca/gw/CMakeLists.txt b/test/soca/gw/CMakeLists.txt index 68a5035a7..14f512c6d 100644 --- a/test/soca/gw/CMakeLists.txt +++ b/test/soca/gw/CMakeLists.txt @@ -12,9 +12,14 @@ add_test(NAME test_gdasapp_soca_prep # Identify machine set(MACHINE "container") -IF (IS_DIRECTORY /work2/noaa/da) - set(MACHINE "orion") - set(PARTITION "orion") +IF (IS_DIRECTORY /work2) + IF (IS_DIRECTORY /apps/other) + set(MACHINE "hercules") + set(PARTITION "hercules") + ELSE() + set(MACHINE "orion") + set(PARTITION "orion") + ENDIF() ENDIF() IF (IS_DIRECTORY /scratch2/NCEPDEV/) set(MACHINE "hera") diff --git a/ush/soca/run_jjobs.py b/ush/soca/run_jjobs.py index 4e0212da2..f9b73c845 100755 --- a/ush/soca/run_jjobs.py +++ b/ush/soca/run_jjobs.py @@ -6,7 +6,7 @@ import argparse from datetime import datetime, timedelta -machines = {"container", "hera", "orion"} +machines = {"container", "hera", "orion", "hercules"} # Assume the default conda environement is gdassapp ENVS = {'JGDAS_GLOBAL_OCEAN_ANALYSIS_VRFY': 'eva'}