Skip to content

Commit 762a2a1

Browse files
deepcharmloadamshwchen2017
committed
Avoid graph break by removing another redundant requires grad false (deepspeedai#7263)
This PR is an follow-up to [PR deepspeedai#7158](deepspeedai#7158) handling the same issue in another place. See [PR deepspeedai#7158](deepspeedai#7158) for details. --------- Signed-off-by: Max Kovalenko <[email protected]> Co-authored-by: Logan Adams <[email protected]> Co-authored-by: Hongwei Chen <[email protected]> Signed-off-by: Max Kovalenko <[email protected]>
1 parent a58480b commit 762a2a1

File tree

1 file changed

+0
-1
lines changed

1 file changed

+0
-1
lines changed

deepspeed/runtime/zero/partition_parameters.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1921,7 +1921,6 @@ def _allgather_params(self, param_list, hierarchy=0):
19211921
flat_scale_tensor = torch.empty(scale_tensor_size,
19221922
dtype=param_list[0].ds_tensor.ds_quant_scale.dtype,
19231923
device=self.local_device)
1924-
flat_scale_tensor.requires_grad = False
19251924
scale_partitions = []
19261925
for i in range(self.world_size):
19271926
start = scale_tensor_size * i

0 commit comments

Comments
 (0)