fxa data directory
This commit is contained in:
parent
b63959e95e
commit
74d128cbd0
9 changed files with 20 additions and 25 deletions
|
@ -5,7 +5,7 @@ path_to_script=`readlink -f $0`
|
|||
RUN_FROM_DIR=`dirname ${path_to_script}`
|
||||
BASE_ENV_DIR=`dirname ${RUN_FROM_DIR}`
|
||||
|
||||
#DR 18113 rehost. /awips/fxa/... Has kicked the bit-bucket.
|
||||
#DR 18113 rehost. /awips2/fxa/... Has kicked the bit-bucket.
|
||||
export TMCP_HOME=/awips2/cave/caveEnvironment
|
||||
export FXA_HOME=/awips2/cave/caveEnvironment
|
||||
|
||||
|
|
|
@ -13,7 +13,7 @@
|
|||
# Usage: VBconversion.sh > outputfile
|
||||
# Expected Inputs: browserFieldMenu.txt and testGridKeyServer_v.txt
|
||||
# (testGridKeyServer_v.txt contains the A1 output of
|
||||
# /awips/fxa/bin/testGridKeyServer v)
|
||||
# /awips2/fxa/bin/testGridKeyServer v)
|
||||
#
|
||||
# KNOWN ISSUES:
|
||||
# 1) This script is inefficient...it may take a minute to finish.
|
||||
|
|
|
@ -1689,12 +1689,7 @@ Scripts = [
|
|||
"-d {productDB} ",
|
||||
|
||||
"Make and Send HTI:" +
|
||||
"/awips2/GFESuite/hti/bin/make_hti.sh {site}",
|
||||
|
||||
"Official Grids to LDAD: " +
|
||||
"ifpAG -h {host} -r {port} -o - -d {productDB} | gzip -9 > " +
|
||||
" /data/fxa/LDAD/ifp/Official/.incoming; " +
|
||||
"mv /data/fxa/LDAD/ifp/Official/.incoming /data/fxa/LDAD/ifp/Official/{ztime} &"
|
||||
"/awips2/GFESuite/hti/bin/make_hti.sh {site}"
|
||||
]
|
||||
|
||||
## Note: Please define TextProducts through
|
||||
|
|
|
@ -687,7 +687,7 @@ class Procedure (SmartScript.SmartScript):
|
|||
|
||||
# if your path does not include FXA_HOME/bin,
|
||||
# this line may work instead of the above line.
|
||||
# cmd = "/awips/fxa/bin/textdb -r " + productID
|
||||
# cmd = "/awips2/fxa/bin/textdb -r " + productID
|
||||
|
||||
(stdout, stdin, stderr) = popen2.popen3(cmd)
|
||||
|
||||
|
|
|
@ -829,7 +829,7 @@ class Forecaster(GridUtilities):
|
|||
if wenameLevel not in JUtil.javaStringListToPylist(self.newdb().getKeys()):
|
||||
msg = wenameLevel + " not in " + \
|
||||
self.newdb().getModelIdentifier() + " " + "SKIPPING"
|
||||
LogStream.logProblem(msg)
|
||||
LogStream.logEvent(msg)
|
||||
continue
|
||||
rval = filter(lambda x,y=wenameLevel : x[0] != y, rval)
|
||||
rval.append((wenameLevel, mthd, fargs))
|
||||
|
|
|
@ -260,7 +260,7 @@ IFR = 'ifr'
|
|||
MVFR = 'mvfr'
|
||||
VFR = 'vfr'
|
||||
|
||||
_TextDB = os.environ.get('FXA_HOME','/awips/fxa') + '/bin/textdb'
|
||||
_TextDB = os.environ.get('FXA_HOME','/awips2/fxa') + '/bin/textdb'
|
||||
TAFWorkPIL = os.environ.get('DEFAULT_CCC','XXX') +'WRKTAF'
|
||||
|
||||
PLUGIN_NAME = 'com.raytheon.uf.common.aviation'
|
||||
|
@ -510,4 +510,4 @@ def mkgmtime(t):
|
|||
caught by Python or the underlying C libraries). The earliest date for which
|
||||
it can generate a time is platform-dependent.
|
||||
"""
|
||||
return time.mktime(t) - time.timezone
|
||||
return time.mktime(t) - time.timezone
|
||||
|
|
|
@ -292,7 +292,7 @@ shefdecode_host : dx1f # controlling UNIX system.
|
|||
shefdecode_dir : $(apps_dir)/shefdecode # main directory location
|
||||
shefdecode_bin : $(shefdecode_dir)/bin # executable programs location
|
||||
shefdecode_input : $(shefdecode_dir)/input # SHEF parameter file location
|
||||
shef_data_dir : /data/fxa/ispan/hydro # input products location
|
||||
shef_data_dir : /awips2/fxa/data/fxa/ispan/hydro # input products location
|
||||
|
||||
shefdecode_log : $(shefdecode_dir)/logs/decoder # daily log files location
|
||||
shef_error_dir : $(shefdecode_dir)/logs/product # product log files location
|
||||
|
@ -684,7 +684,7 @@ pproc_util_log_dir : $(pproc_log)/misc # miscellaneous logs
|
|||
# DecodeDPA tokens (formerly DecodeHDP tokens that looked like hdp_*)
|
||||
|
||||
dpa_log_dir : $(pproc_log)/decodedpa # DPA Decoder logs
|
||||
dpa_prod_dir : /data/fxa/ispan/hdp # DPA input directory
|
||||
dpa_prod_dir : /awips2/fxa/data/fxa/ispan/hdp # DPA input directory
|
||||
dpa_gather : $(pproc_local_data)/dpa_gather # DPA gather directory
|
||||
dpa_error_dir : $(pproc_local_data)/stage1_error # DPA error files
|
||||
dpa_arch_dir : $(pproc_local_data)/stage1_archive # DPA archives
|
||||
|
@ -924,7 +924,7 @@ mpe_freezing_data_dir : $(mpe_gageqc_dir)/freezing_level
|
|||
mpe_maz_dir : $(mpe_freezing_data_dir)/MAZ
|
||||
mpe_grid_freezing_dir : $(mpe_freezing_data_dir)/grid
|
||||
mpe_point_freezing_dir : $(mpe_freezing_data_dir)/point
|
||||
ruc_model_data_dir : /data/fxa/Grid/SBN/netCDF/CONUS211/RUC
|
||||
ruc_model_data_dir : /awips2/fxa/data/fxa/Grid/SBN/netCDF/CONUS211/RUC
|
||||
|
||||
### MPE 1 hour mosaics and fields and supporting reference fields.
|
||||
mpe_avgrmosaic_dir : $(mpe_fieldgen_product_dir)/avgrmosaic
|
||||
|
@ -1065,7 +1065,7 @@ mpe_generate_areal_qpe : OFF
|
|||
gaq_rfc_list : MBRFC,NCRFC
|
||||
gaq_dur_list : 1,6,24
|
||||
gaq_app_dir : $(pproc_local_data)/app/gen_areal_qpe
|
||||
gaq_input_dir : /data/fxa/Grid/SBN/netCDF/HRAP/QPE
|
||||
gaq_input_dir : /awips2/fxa/data/fxa/Grid/SBN/netCDF/HRAP/QPE
|
||||
gaq_log_dir : $(pproc_log)/gen_areal_qpe
|
||||
gaq_rfc_mask_dir : $(gaq_app_dir)
|
||||
gaq_temp_xmrg_dir : $(rfcwide_output_dir)/rfcqpe_temp
|
||||
|
@ -1083,11 +1083,11 @@ adjust_PC_startingtime : 4 #allow PC starting time tolerance
|
|||
|
||||
### tokens for sending MPE mean field bias data to the ORPG
|
||||
|
||||
bias_message_dir : $(apps_dir)/data/fxa/radar/envData
|
||||
bias_message_dir : $(apps_dir)/awips2/fxa/data/fxa/radar/envData
|
||||
|
||||
### tokens for Lightning Data processing
|
||||
|
||||
lightning_input_dir : /data/fxa/point/binLightning/netcdf
|
||||
lightning_input_dir : /awips2/fxa/data/fxa/point/binLightning/netcdf
|
||||
|
||||
lightning_log_dir : $(pproc_log)/lightning_proc
|
||||
|
||||
|
@ -1616,7 +1616,7 @@ vsys_rls : $(verify_rls) #Not really needed, but consi
|
|||
|
||||
# ================== RFC Archive Database tokens ===============================
|
||||
|
||||
archive_shefdata_dir : /data/fxa/ispan/hydro_adbs # directory for archive data
|
||||
archive_shefdata_dir : /awips2/fxa/data/fxa/ispan/hydro_adbs # directory for archive data
|
||||
archive_enable : OFF # ON/OFF - Enable or Disable
|
||||
# archive data feed (OFF by default)
|
||||
metar_output_dir : $(whfs_local_data_dir)/metar_output # metar2shef temp output directory
|
||||
|
@ -1881,8 +1881,8 @@ hdb_map_projection : FLAT # The initial map projection used by HDB.
|
|||
# Possible values: FLAT, POLAR, HRAP
|
||||
# ====================== DHM Token =======================
|
||||
dhm_data_dir : $(ofs_files)/$(ofs_level)/dhmdata # DHM data dir
|
||||
dhm_d2d_data_dir : /data/fxa/Grid/LOCAL/netCDF/DHM # d2d data dir
|
||||
dhm_d2d_notify_bin_dir : /awips/fxa/bin # d2d notify bin dir
|
||||
dhm_d2d_data_dir : /awips2/fxa/data/fxa/Grid/LOCAL/netCDF/DHM # d2d data dir
|
||||
dhm_d2d_notify_bin_dir : /awips2/fxa/bin # d2d notify bin dir
|
||||
rdhm_input_dir : $(geo_data)
|
||||
dhm_rain_plus_melt_data_dir: $(geo_data)
|
||||
|
||||
|
|
|
@ -125,10 +125,10 @@
|
|||
# supposed to stop. Instead it kept trying. vtm #
|
||||
# 06/19/08: Changed the directory on AWIPS where the netcdf file is created. #
|
||||
# Now using a drive local to machine for performance reasons. #
|
||||
# New directory is now /awips/fxa/netcdf. This directory will be #
|
||||
# New directory is now /awips2/fxa/netcdf. This directory will be #
|
||||
# created by the script if it doesn't exist. vtm #
|
||||
# 06/19/08: Changed script to remove AWIPS netcdf and log files sooner. vtm #
|
||||
# 07/10/08: Made the /awips/fxa/netcdf configurable for by DX and LX machines.#
|
||||
# 07/10/08: Made the /awips2/fxa/netcdf configurable for by DX and LX machines.#
|
||||
# vtm #
|
||||
# 07/11/08: Pointed most all of script feedback to a log file that can be #
|
||||
# found in the /awips/GFESuite/primary/data/logfiles/yyyymmdd/ #
|
||||
|
|
|
@ -35,8 +35,8 @@ while ((i < $#)); do
|
|||
fi
|
||||
(( i+= 1 ))
|
||||
done
|
||||
FXA_HOME=/awips/fxa
|
||||
. /awips/fxa/readenv.sh
|
||||
FXA_HOME=/awips2/fxa
|
||||
. /awips2/fxa/readenv.sh
|
||||
#If WFO, launch legacy A1 hmMonitor
|
||||
if [ "$SITE_TYPE" = "wfo" ]; then
|
||||
$FXA_HOME/bin/hmMonitor.tcl --forTextWS "${extra_textws_args[@]}" &
|
||||
|
|
Loading…
Add table
Reference in a new issue