Skip to content

Commit

Permalink
Took a first pass at reducing the resources to be requested for the v…
Browse files Browse the repository at this point in the history
…arious workflow tasks.
  • Loading branch information
drewoldag committed Sep 10, 2024
1 parent b6f28c6 commit c8c2023
Showing 1 changed file with 11 additions and 9 deletions.
20 changes: 11 additions & 9 deletions src/kbmod_wf/resource_configs/klone_configuration.py
Original file line number Diff line number Diff line change
Expand Up @@ -8,8 +8,8 @@
walltimes = {
"compute_bigmem": "01:00:00",
"large_mem": "04:00:00",
"sharded_reproject": "04:00:00",
"gpu_max": "08:00:00",
"sharded_reproject": "01:00:00",
"gpu_max": "01:00:00",
}


Expand All @@ -21,7 +21,7 @@ def klone_resource_config():
os.path.join("/gscratch/dirac/kbmod/workflow/run_logs", datetime.date.today().isoformat())
),
run_dir=os.path.join("/gscratch/dirac/kbmod/workflow/run_logs", datetime.date.today().isoformat()),
retries=1,
retries=100,
executors=[
HighThroughputExecutor(
label="small_cpu",
Expand All @@ -35,14 +35,15 @@ def klone_resource_config():
parallelism=1,
nodes_per_block=1,
cores_per_node=1, # perhaps should be 8???
mem_per_node=256, # In GB
mem_per_node=32, # In GB
exclusive=False,
walltime=walltimes["compute_bigmem"],
# Command to run before starting worker - i.e. conda activate <special_env>
worker_init="",
),
),
HighThroughputExecutor(
# This executor was used for the pre-TNO reprojection task
label="large_mem",
max_workers=1,
provider=SlurmProvider(
Expand All @@ -62,18 +63,19 @@ def klone_resource_config():
),
),
HighThroughputExecutor(
# This executor is used for reprojecting sharded WorkUnits
label="sharded_reproject",
max_workers=1,
provider=SlurmProvider(
partition="ckpt-g2",
account="astro",
min_blocks=0,
max_blocks=2,
max_blocks=10,
init_blocks=0,
parallelism=1,
nodes_per_block=1,
cores_per_node=32,
mem_per_node=128, # ~2-4 GB per core
cores_per_node=8,
mem_per_node=32, # ~2-4 GB per core
exclusive=False,
walltime=walltimes["sharded_reproject"],
# Command to run before starting worker - i.e. conda activate <special_env>
Expand All @@ -91,8 +93,8 @@ def klone_resource_config():
init_blocks=0,
parallelism=1,
nodes_per_block=1,
cores_per_node=2, # perhaps should be 8???
mem_per_node=512, # In GB
cores_per_node=1,
mem_per_node=128, # In GB
exclusive=False,
walltime=walltimes["gpu_max"],
# Command to run before starting worker - i.e. conda activate <special_env>
Expand Down

0 comments on commit c8c2023

Please sign in to comment.