Skip to content

Commit

Permalink
Remove enable_placement: ... from blueprint examples
Browse files Browse the repository at this point in the history
  • Loading branch information
mr0re1 committed Jan 10, 2025
1 parent 1fc131c commit ebde2b9
Show file tree
Hide file tree
Showing 21 changed files with 0 additions and 33 deletions.
2 changes: 0 additions & 2 deletions community/examples/AMD/hpc-amd-slurm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -178,7 +178,6 @@ deployment_groups:
machine_type: c2d-standard-4
node_count_dynamic_max: 10
bandwidth_tier: gvnic_enabled
enable_placement: false
allow_automatic_updates: false

- id: low_cost_partition
Expand All @@ -194,7 +193,6 @@ deployment_groups:
machine_type: c2d-standard-112
node_count_dynamic_max: 50
bandwidth_tier: gvnic_enabled
enable_placement: true
allow_automatic_updates: false

# Because is_default is set to true, jobs will run on this partition unless an
Expand Down
1 change: 0 additions & 1 deletion community/examples/hpc-slurm-sharedvpc.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,6 @@ deployment_groups:
settings:
node_count_dynamic_max: 4
machine_type: n2-standard-2
enable_placement: false # the default is: true
allow_automatic_updates: false

- id: debug_partition
Expand Down
1 change: 0 additions & 1 deletion community/examples/hpc-slurm-ubuntu2004.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -51,7 +51,6 @@ deployment_groups:
use: [network]
settings:
instance_image: $(vars.slurm_image)
enable_placement: false # the default is: true
node_count_dynamic_max: 4
machine_type: n2-standard-2

Expand Down
4 changes: 0 additions & 4 deletions community/examples/htc-slurm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -82,7 +82,6 @@ deployment_groups:
name: c2s60
node_count_dynamic_max: 200
bandwidth_tier: gvnic_enabled
enable_placement: false
allow_automatic_updates: false

- id: compute_nodeset_c2s30
Expand All @@ -92,7 +91,6 @@ deployment_groups:
node_count_dynamic_max: 200
machine_type: c2-standard-30
bandwidth_tier: gvnic_enabled
enable_placement: false
allow_automatic_updates: false

- id: compute_partition
Expand All @@ -116,7 +114,6 @@ deployment_groups:
machine_type: n2-standard-2
node_count_dynamic_max: 10
bandwidth_tier: gvnic_enabled
enable_placement: false
allow_automatic_updates: false

- id: low_cost_nodeset_n2s4
Expand All @@ -127,7 +124,6 @@ deployment_groups:
machine_type: n2-standard-4
node_count_dynamic_max: 10
bandwidth_tier: gvnic_enabled
enable_placement: false
allow_automatic_updates: false

- id: low_cost_partition
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,6 @@ deployment_groups:
settings:
partition_name: debug
exclusive: false
enable_placement: false
is_default: false

- id: compute_node_group
Expand Down
1 change: 0 additions & 1 deletion docs/tutorials/hpc-slurm-qwiklabs.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,6 @@ deployment_groups:
settings:
node_count_dynamic_max: 4
machine_type: n2d-standard-2
enable_placement: false # the default is: true

- id: debug_partition
source: community/modules/compute/schedmd-slurm-gcp-v6-partition
Expand Down
1 change: 0 additions & 1 deletion examples/hcls-blueprint.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -326,7 +326,6 @@ deployment_groups:
enable_smt: true
node_count_dynamic_max: 20
machine_type: g2-standard-4
enable_placement: False
allow_automatic_updates: false

- id: gpu_partition
Expand Down
2 changes: 0 additions & 2 deletions examples/hpc-enterprise-slurm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -112,7 +112,6 @@ deployment_groups:
node_count_dynamic_max: 4
machine_type: n2-standard-2
instance_image: $(vars.slurm_image)
enable_placement: false # the default is: true
allow_automatic_updates: false

- id: n2_partition
Expand All @@ -132,7 +131,6 @@ deployment_groups:
node_count_dynamic_max: 20
machine_type: c2-standard-60 # this is the default
instance_image: $(vars.slurm_image)
enable_placement: true
bandwidth_tier: tier_1_enabled
disk_type: pd-ssd
disk_size_gb: 100
Expand Down
1 change: 0 additions & 1 deletion examples/hpc-slurm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,6 @@ deployment_groups:
settings:
node_count_dynamic_max: 4
machine_type: n2-standard-2
enable_placement: false # the default is: true
allow_automatic_updates: false

- id: debug_partition
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -96,7 +96,6 @@ deployment_groups:
node_count_static: 0
node_count_dynamic_max: 4
machine_type: n2-standard-2
enable_placement: false

- id: debug_partition
source: community/modules/compute/schedmd-slurm-gcp-v6-partition
Expand Down Expand Up @@ -182,7 +181,6 @@ deployment_groups:
settings:
reservation_name: $(vars.a3_reservation_name)
maintenance_interval: $(vars.a3_maintenance_interval)
enable_placement: false
node_count_static: $(vars.a3_static_cluster_size)
node_count_dynamic_max: 0
disk_type: pd-ssd
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,6 @@ vars:
deployment_name: a3mega-cluster
a3mega_partition_name: a3mega
a3mega_maintenance_interval: ""
enable_placement: false
remote_mount_homefs: /nfsshare
local_mount_homefs: /home
instance_image_custom: true
Expand Down
1 change: 0 additions & 1 deletion examples/ml-slurm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -202,7 +202,6 @@ deployment_groups:
use: [network]
settings:
node_count_dynamic_max: 20
enable_placement: false
bandwidth_tier: gvnic_enabled
machine_type: g2-standard-4
instance_image: $(vars.new_image)
Expand Down
1 change: 0 additions & 1 deletion examples/ps-slurm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -55,7 +55,6 @@ deployment_groups:
settings:
node_count_dynamic_max: 4
machine_type: $(vars.compute_node_machine_type)
enable_placement: false # the default is: true
allow_automatic_updates: false

- id: debug_partition
Expand Down
2 changes: 0 additions & 2 deletions tools/cloud-build/daily-tests/blueprints/lustre-slurm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -23,8 +23,6 @@ vars:
zone: us-central1-a
machine_type: n2-standard-2
disk_type: pd-ssd
# enable_placement: false
# on_host_maintenance: MIGRATE
num_nodes: 1
rocky_image:
family: slurm-gcp-6-8-hpc-rocky-linux-8
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,6 @@ deployment_groups:
node_count_dynamic_max: 0
node_count_static: 5
allow_automatic_updates: false
enable_placement: true

- id: partition
source: community/modules/compute/schedmd-slurm-gcp-v6-partition
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -108,7 +108,6 @@ deployment_groups:
node_count_dynamic_max: 4
machine_type: n2-standard-2
instance_image: $(vars.slurm_image)
enable_placement: false # the default is: true
allow_automatic_updates: false

- id: n2_partition
Expand All @@ -128,7 +127,6 @@ deployment_groups:
node_count_dynamic_max: 20
machine_type: c2-standard-60 # this is the default
instance_image: $(vars.slurm_image)
enable_placement: true
bandwidth_tier: tier_1_enabled
disk_type: pd-ssd
disk_size_gb: 100
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -150,7 +150,6 @@ deployment_groups:
- compute_sa
settings:
allow_automatic_updates: false
enable_placement: false
instance_image: ((var.slurm_image))
instance_image_custom: ((var.instance_image_custom))
labels: ((var.labels))
Expand Down Expand Up @@ -185,7 +184,6 @@ deployment_groups:
bandwidth_tier: tier_1_enabled
disk_size_gb: 100
disk_type: pd-ssd
enable_placement: true
instance_image: ((var.slurm_image))
instance_image_custom: ((var.instance_image_custom))
labels: ((var.labels))
Expand Down
1 change: 0 additions & 1 deletion tools/validate_configs/test_configs/config-ssh.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,6 @@ deployment_groups:
settings:
node_count_dynamic_max: 4
machine_type: n2-standard-2
enable_placement: false # the default is: true

- id: debug_partition
source: community/modules/compute/schedmd-slurm-gcp-v6-partition
Expand Down
3 changes: 0 additions & 3 deletions tools/validate_configs/test_configs/gpu.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -139,7 +139,6 @@ deployment_groups:
use: [network_slurm]
settings:
name: nogpu
enable_placement: false
node_count_dynamic_max: 4
machine_type: n2-standard-2

Expand All @@ -148,7 +147,6 @@ deployment_groups:
use: [network_slurm]
settings:
name: man
enable_placement: false
node_count_dynamic_max: 4
machine_type: a2-ultragpu-2g
guest_accelerator:
Expand All @@ -160,7 +158,6 @@ deployment_groups:
use: [network_slurm]
settings:
name: auto
enable_placement: false
node_count_dynamic_max: 4
machine_type: a2-ultragpu-2g

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,6 @@ deployment_groups:
settings:
node_count_dynamic_max: 4
machine_type: n2-standard-2
enable_placement: false

- id: debug_partition
source: community/modules/compute/schedmd-slurm-gcp-v6-partition
Expand Down
2 changes: 0 additions & 2 deletions tools/validate_configs/test_configs/zone-policies-slurm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,6 @@ deployment_groups:
node_count_dynamic_max: 4
machine_type: n2-standard-2
enable_public_ips: true
enable_placement: false

- id: zonal_partition
source: community/modules/compute/schedmd-slurm-gcp-v6-partition
Expand All @@ -63,7 +62,6 @@ deployment_groups:
settings:
node_count_dynamic_max: 4
machine_type: n2-standard-2
enable_placement: false
zones: $(vars.additional_zones)

- id: multizonal_partition
Expand Down

0 comments on commit ebde2b9

Please sign in to comment.