Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
D
DART-WRF
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Wiki
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Deploy
Releases
Package registry
Model registry
Operate
Terraform modules
Monitor
Incidents
Service Desk
Analyze
Value stream analytics
Contributor analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
GitLab community forum
Contribute to GitLab
Provide feedback
Terms and privacy
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
DataAssimilation
DART-WRF
Commits
fe007678
Commit
fe007678
authored
4 months ago
by
Lukas Kugler
Browse files
Options
Downloads
Patches
Plain Diff
docs, configs
parent
57ff4b1f
No related branches found
No related tags found
No related merge requests found
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
dartwrf/workflows.py
+46
-49
46 additions, 49 deletions
dartwrf/workflows.py
with
46 additions
and
49 deletions
dartwrf/workflows.py
+
46
−
49
View file @
fe007678
...
@@ -18,22 +18,14 @@ from dartwrf.utils import Config
...
@@ -18,22 +18,14 @@ from dartwrf.utils import Config
class
WorkFlows
(
object
):
class
WorkFlows
(
object
):
def
__init__
(
self
,
cfg
:
Config
):
def
__init__
(
self
,
cfg
:
Config
):
"""
Set up the experiment folder in `archivedir`.
"""
Create the archive directory, copy scripts to archive
1. Copy the selected config files
2. Import configurations
3. Prepare obskind.py file (dictionary of observation types)
4. Copy the scripts and configs to `archivedir`
5. Set python path
6. Set log path and slurm scripts path
Args:
exp_config (str): Path to exp config file
server_config (str): Path to the cluster config file
Attributes:
Attributes:
cluster (obj): cluster configuration as defined in server_config file
use_slurm: if True, run jobs with SLURM
exp (obj): experiment configuration as defined in exp_config file
dir_log: Logging directory for slurm
dir_slurm: Scripts directory for slurm
dir_dartwrf_run: as defined in Config
python: python command with prepended PYTHONPATH
"""
"""
print
(
'
------------------------------------------------------
'
)
print
(
'
------------------------------------------------------
'
)
print
(
'
>>> Experiment name:
"'
+
cfg
.
name
+
'"'
)
print
(
'
>>> Experiment name:
"'
+
cfg
.
name
+
'"'
)
...
@@ -42,7 +34,7 @@ class WorkFlows(object):
...
@@ -42,7 +34,7 @@ class WorkFlows(object):
# Copy scripts and config files to `self.archivedir` folder
# Copy scripts and config files to `self.archivedir` folder
dirs_exist_ok
=
False
dirs_exist_ok
=
False
if
os
.
path
.
exists
(
cfg
.
dir_archive
+
'
/DART-WRF/
'
):
if
os
.
path
.
exists
(
cfg
.
dir_archive
+
'
/DART-WRF/
'
):
if
input
(
'
The experiment name already exists! Overwrite existing
experiment
? (Y/n)
'
)
in
[
'
Y
'
,
'
y
'
]:
if
input
(
'
The experiment name already exists! Overwrite existing
scripts
? (Y/n)
'
)
in
[
'
Y
'
,
'
y
'
]:
dirs_exist_ok
=
True
dirs_exist_ok
=
True
shutil
.
copytree
(
cfg
.
dir_dartwrf_dev
,
shutil
.
copytree
(
cfg
.
dir_dartwrf_dev
,
...
@@ -68,10 +60,9 @@ class WorkFlows(object):
...
@@ -68,10 +60,9 @@ class WorkFlows(object):
print
(
'
------------------------------------------------------
'
)
print
(
'
------------------------------------------------------
'
)
# use this python path
# use this python path
run_wrf_from_this_folder
=
cfg
.
dir_dartwrf_run
+
'
/../
'
pythonpath_archive
=
cfg
.
dir_dartwrf_run
+
'
/../
'
self
.
dir_dartwrf_run
=
cfg
.
dir_dartwrf_run
self
.
dir_dartwrf_run
=
cfg
.
dir_dartwrf_run
self
.
python
=
'
export PYTHONPATH=
'
+
run_wrf_from_this_folder
+
'
;
'
+
cfg
.
python
self
.
python
=
'
export PYTHONPATH=
'
+
pythonpath_archive
+
'
;
'
+
cfg
.
python
self
.
cfg
=
cfg
def
run_job
(
self
,
cmd
,
cfg
,
depends_on
=
None
,
**
kwargs
):
def
run_job
(
self
,
cmd
,
cfg
,
depends_on
=
None
,
**
kwargs
):
"""
Run scripts in a shell
"""
Run scripts in a shell
...
@@ -92,7 +83,10 @@ class WorkFlows(object):
...
@@ -92,7 +83,10 @@ class WorkFlows(object):
from
slurmpy
import
Slurm
from
slurmpy
import
Slurm
# name of calling function
# name of calling function
path_to_script
=
inspect
.
stack
()[
1
].
function
path_to_script
=
inspect
.
stack
()[
1
].
function
jobname
=
path_to_script
.
split
(
'
/
'
)[
-
1
]
+
'
-
'
+
cfg
.
f_cfg_current
.
split
(
'
/
'
)[
-
1
].
replace
(
'
.py
'
,
''
)
if
'
time
'
in
cfg
:
jobname
=
path_to_script
.
split
(
'
/
'
)[
-
1
]
+
'
-
'
+
cfg
.
time
.
strftime
(
'
%H:%M
'
)
else
:
jobname
=
path_to_script
.
split
(
'
/
'
)[
-
1
]
+
'
-
'
+
cfg
.
name
print
(
'
> SLURM job:
'
,
jobname
)
print
(
'
> SLURM job:
'
,
jobname
)
slurm_kwargs
=
cfg
.
slurm_kwargs
.
copy
()
slurm_kwargs
=
cfg
.
slurm_kwargs
.
copy
()
...
@@ -175,12 +169,11 @@ class WorkFlows(object):
...
@@ -175,12 +169,11 @@ class WorkFlows(object):
Returns:
Returns:
str: job ID of the submitted job
str: job ID of the submitted job
"""
"""
cmd
=
script_to_str
(
self
.
cfg
.
WRF_ideal_template
cmd
=
script_to_str
(
cfg
.
WRF_ideal_template
).
replace
(
'
<expname>
'
,
cfg
.
name
).
replace
(
'
<wrf_rundir_base>
'
,
cfg
.
dir_wrf_run
.
replace
(
'
<ens>
'
,
'
$IENS
'
)
).
replace
(
'
<wrf_rundir_base>
'
,
cfg
.
dir_wrf_run
).
replace
(
'
<wrf_modules>
'
,
cfg
.
wrf_modules
,
).
replace
(
'
<wrf_modules>
'
,
cfg
.
wrf_modules
,
)
)
id
=
self
.
run_job
(
cmd
,
cfg
,
depends_on
=
[
depends_on
],
time
=
"
30
"
)
id
=
self
.
run_job
(
cmd
,
cfg
,
depends_on
=
[
depends_on
],
time
=
"
30
"
,
array
=
"
1-
"
+
str
(
cfg
.
ensemble_size
)
)
return
id
return
id
def
run_WRF
(
self
,
cfg
,
depends_on
=
None
):
def
run_WRF
(
self
,
cfg
,
depends_on
=
None
):
...
@@ -191,31 +184,32 @@ class WorkFlows(object):
...
@@ -191,31 +184,32 @@ class WorkFlows(object):
end
=
cfg
.
WRF_end
end
=
cfg
.
WRF_end
# SLURM configuration for WRF
# SLURM configuration for WRF
slurm_kwargs
=
{
"
array
"
:
"
1-
"
+
str
(
self
.
cfg
.
ensemble_size
),
slurm_kwargs
=
{
"
array
"
:
"
1-
"
+
str
(
cfg
.
ensemble_size
),
"
nodes
"
:
"
1
"
,
"
nodes
"
:
"
1
"
,
"
ntasks
"
:
str
(
self
.
cfg
.
max_nproc_for_each_ensemble_member
),
"
ntasks
"
:
str
(
cfg
.
max_nproc_for_each_ensemble_member
),
"
ntasks-per-core
"
:
"
1
"
,
"
mem
"
:
"
90G
"
,
}
"
ntasks-per-core
"
:
"
1
"
,
"
mem
"
:
"
90G
"
,
"
ntasks-per-node
"
:
str
(
cfg
.
max_nproc_for_each_ensemble_member
),}
# command from template file
# command from template file
wrf_cmd
=
script_to_str
(
self
.
cfg
.
WRF_exe_template
wrf_cmd
=
script_to_str
(
cfg
.
WRF_exe_template
).
replace
(
'
<dir_wrf_run>
'
,
self
.
cfg
.
dir_wrf_run
.
replace
(
'
<ens>
'
,
'
$IENS
'
)
).
replace
(
'
<dir_wrf_run>
'
,
cfg
.
dir_wrf_run
.
replace
(
'
<ens>
'
,
'
$IENS
'
)
).
replace
(
'
<wrf_modules>
'
,
self
.
cfg
.
wrf_modules
,
).
replace
(
'
<wrf_modules>
'
,
cfg
.
wrf_modules
,
).
replace
(
'
<WRF_number_of_processors>
'
,
str
(
self
.
cfg
.
max_nproc_for_each_ensemble_member
),
).
replace
(
'
<WRF_number_of_processors>
'
,
str
(
cfg
.
max_nproc_for_each_ensemble_member
),
)
)
# prepare namelist
# prepare namelist
path_to_script
=
self
.
dir_dartwrf_run
+
'
/prepare_namelist.py
'
path_to_script
=
self
.
dir_dartwrf_run
+
'
/prepare_namelist.py
'
cmd
=
'
'
.
join
([
self
.
python
,
path_to_script
,
self
.
cfg
.
f_cfg_current
])
cmd
=
'
'
.
join
([
self
.
python
,
path_to_script
,
cfg
.
f_cfg_current
])
id
=
self
.
run_job
(
cmd
,
cfg
,
depends_on
=
[
depends_on
])
id
=
self
.
run_job
(
cmd
,
cfg
,
depends_on
=
[
depends_on
])
# run WRF ensemble
# run WRF ensemble
time_in_simulation_hours
=
(
end
-
start
).
total_seconds
()
/
3600
time_in_simulation_hours
=
(
end
-
start
).
total_seconds
()
/
3600
runtime_wallclock_mins_expected
=
int
(
# runtime_wallclock_mins_expected = int(time_in_simulation_hours*15*1.5 + 15)
time_in_simulation_hours
*
30
+
10
)
# usually <15 min/hour
runtime_wallclock_mins_expected
=
int
(
time_in_simulation_hours
*
30
*
1.5
+
15
)
# usually max 15 min/hour + 50% margin + 15 min buffer
slurm_kwargs
.
update
({
"
time
"
:
str
(
runtime_wallclock_mins_expected
)})
slurm_kwargs
.
update
({
"
time
"
:
str
(
runtime_wallclock_mins_expected
)})
if
runtime_wallclock_mins_expected
>
20
:
if
runtime_wallclock_mins_expected
>
25
:
slurm_kwargs
.
update
({
"
partition
"
:
"
amd
"
})
slurm_kwargs
.
update
({
"
partition
"
:
"
amd
"
})
#
#cfg_update.update({"exclude": "jet03"})
#cfg_update.update({"exclude": "jet03"})
id
=
self
.
run_job
(
wrf_cmd
,
cfg
,
depends_on
=
[
id
],
**
slurm_kwargs
)
id
=
self
.
run_job
(
wrf_cmd
,
cfg
,
depends_on
=
[
id
],
**
slurm_kwargs
)
return
id
return
id
...
@@ -271,18 +265,30 @@ class WorkFlows(object):
...
@@ -271,18 +265,30 @@ class WorkFlows(object):
id
=
self
.
run_job
(
cmd
,
cfg
,
depends_on
=
[
depends_on
],
time
=
"
10
"
)
id
=
self
.
run_job
(
cmd
,
cfg
,
depends_on
=
[
depends_on
],
time
=
"
10
"
)
return
id
return
id
def
create_satimages
(
self
,
cfg
,
depends_on
=
None
):
def
run_RTTOV
(
self
,
cfg
,
depends_on
=
None
):
"""
Run a job array, one job per ensemble member, to create satellite images
"""
"""
Run a job array, one job per ensemble member, to create satellite images
"""
prefix
=
'
module purge; module load rttov/v13.2-gcc-8.5.0; python
'
prefix
=
'
module purge; module load rttov/v13.2-gcc-8.5.0; python
'
path_to_script
=
'
~/RTTOV-WRF/run_init.py
'
path_to_script
=
'
~/RTTOV-WRF/run_init.py
'
cmd
=
'
'
.
join
([
prefix
,
path_to_script
,
cfg
.
f_cfg_current
])
cmd
=
'
'
.
join
([
prefix
,
path_to_script
,
cfg
.
dir_archive
+
cfg
.
time
.
strftime
(
'
/%Y-%m-%d_%H:%M/
'
),
'
$SLURM_ARRAY_TASK_ID
'
])
id
=
self
.
run_job
(
cmd
,
cfg
,
depends_on
=
[
depends_on
],
id
=
self
.
run_job
(
cmd
,
cfg
,
depends_on
=
[
depends_on
],
**
{
"
ntasks
"
:
"
1
"
,
"
time
"
:
"
60
"
,
"
mem
"
:
"
10G
"
,
**
{
"
ntasks
"
:
"
1
"
,
"
time
"
:
"
60
"
,
"
mem
"
:
"
10G
"
,
"
array
"
:
"
1-
"
+
str
(
self
.
cfg
.
ensemble_size
)})
"
array
"
:
"
1-
"
+
str
(
cfg
.
ensemble_size
)})
return
id
return
id
def
verify
(
self
,
cfg
:
Config
,
depends_on
=
None
):
"""
Not included in DART-WRF
"""
cmd
=
'
'
.
join
([
'
python /jetfs/home/lkugler/osse_analysis/plot_from_raw/analyze_fc.py
'
,
cfg
.
name
,
cfg
.
verify_against
,
#cfg.time.strftime('%y%m%d_%H:%M'),
'
sat
'
,
'
wrf
'
,
'
has_node
'
,
'
np=10
'
,
'
mem=250G
'
])
self
.
run_job
(
cmd
,
cfg
,
depends_on
=
[
depends_on
],
**
{
"
time
"
:
"
03:00:00
"
,
"
mail-type
"
:
"
FAIL,END
"
,
"
ntasks
"
:
"
10
"
,
"
ntasks-per-node
"
:
"
10
"
,
"
ntasks-per-core
"
:
"
1
"
,
"
mem
"
:
"
250G
"
})
def
evaluate_obs_posterior_after_analysis
(
self
,
cfg
,
depends_on
=
None
):
def
evaluate_obs_posterior_after_analysis
(
self
,
cfg
,
depends_on
=
None
):
path_to_script
=
self
.
dir_dartwrf_run
+
'
/evaluate_obs_space.py
'
path_to_script
=
self
.
dir_dartwrf_run
+
'
/evaluate_obs_space.py
'
...
@@ -298,12 +304,3 @@ class WorkFlows(object):
...
@@ -298,12 +304,3 @@ class WorkFlows(object):
# "time": "15", "mail-type": "FAIL"},
# "time": "15", "mail-type": "FAIL"},
# depends_on=[id])
# depends_on=[id])
return
id
return
id
\ No newline at end of file
def
verify
(
self
,
cfg
,
depends_on
=
None
):
"""
Not included in DART-WRF
"""
cmd
=
'
'
.
join
([
'
python /jetfs/home/lkugler/osse_analysis/plot_from_raw/analyze_fc.py
'
,
cfg
.
f_cfg_current
])
id
=
self
.
run_job
(
cmd
,
cfg
,
depends_on
=
[
depends_on
],
**
{
"
time
"
:
"
210
"
,
"
mail-type
"
:
"
FAIL,END
"
,
"
ntasks
"
:
"
10
"
,
"
ntasks-per-node
"
:
"
10
"
,
"
ntasks-per-core
"
:
"
1
"
,
"
mem
"
:
"
250G
"
})
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment