Skip to content

Commit

Permalink
Cleanup
Browse files Browse the repository at this point in the history
Signed-off-by: Olatunji Ruwase <[email protected]>
  • Loading branch information
tjruwase committed Feb 8, 2025
1 parent 204c4dd commit 78e1915
Showing 1 changed file with 0 additions and 7 deletions.
7 changes: 0 additions & 7 deletions deepspeed/runtime/bf16_optimizer.py
Original file line number Diff line number Diff line change
Expand Up @@ -340,10 +340,6 @@ def _update_hp_grad(self, lp, group_idx, param_idx, clear_lp_grads):
hp_grad.data.add_(lp.grad.data.to(hp_grad.dtype).view(hp_grad.shape))
lp._hp_grad = hp_grad
self.fp32_groups_has_gradients[group_idx][param_idx] = True
# if param_idx == 200:
print_rank_0(f"update_hp_grad self={id(self)} {group_idx=} {param_idx=} gnorm={float(lp.grad.norm().float())}",
force=True)
# import pdb; pdb.set_trace()

# clear gradients
if clear_lp_grads:
Expand Down Expand Up @@ -429,9 +425,6 @@ def update_lp_params(self):
fp32_partition) in enumerate(zip(self.bf16_partitioned_groups, self.fp32_groups_flat_partition)):
partition_id = dist.get_rank(group=self.real_dp_process_group[i])
bf16_partitions[partition_id].data.copy_(fp32_partition.data)
# print_rank_0(f'update_lp_params {i=} {partition_id=}', force=True)
# if i == 0:
# print_rank_0(f'{fp32_partition[:10]=}', force=True)

all_gather_dp_groups(groups_flat=self.bf16_groups_flat,
partitioned_param_groups=self.bf16_partitioned_groups,
Expand Down

0 comments on commit 78e1915

Please sign in to comment.