Skip to content

Commit c4b01f6

Browse files
KakaruHayateyxlllc
authored andcommitted
Fix RoPE cache issue about 'find_unused_parameters' when DDP training. (#244)
* Fix issue about 'find_unused_parameters' when DDP training. * annotation * slim * Fix issue about 'find_unused_parameters' when DDP training. annotation slim * Update rotary_embedding_torch.py
1 parent 277c082 commit c4b01f6

1 file changed

Lines changed: 4 additions & 1 deletion

File tree

modules/commons/rotary_embedding_torch.py

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -306,7 +306,10 @@ def forward(
306306
exists(self.cached_freqs) and \
307307
(offset + seq_len) <= self.cached_freqs_seq_len
308308
):
309-
return self.cached_freqs[offset:(offset + seq_len)].detach()
309+
freqs = self.cached_freqs[offset:(offset + seq_len)].detach()
310+
# Fix issue about 'find_unused_parameters' when DDP training.(#244)
311+
freqs = freqs + 0. * self.freqs.sum()
312+
return freqs
310313

311314
freqs = self.freqs
312315

0 commit comments

Comments
 (0)