From 75e9b8d69b7a0a4c4c8d43430c83dd2bd0249e8c Mon Sep 17 00:00:00 2001 From: Kunal Vaishnavi Date: Tue, 3 Oct 2023 18:43:34 +0000 Subject: [PATCH] Add back copy comment --- src/transformers/models/persimmon/modeling_persimmon.py | 1 + 1 file changed, 1 insertion(+) diff --git a/src/transformers/models/persimmon/modeling_persimmon.py b/src/transformers/models/persimmon/modeling_persimmon.py index 8d85c2694aa4c8..250852f7875799 100644 --- a/src/transformers/models/persimmon/modeling_persimmon.py +++ b/src/transformers/models/persimmon/modeling_persimmon.py @@ -71,6 +71,7 @@ def _expand_mask(mask: torch.Tensor, dtype: torch.dtype, tgt_len: Optional[int] return inverted_mask.masked_fill(inverted_mask.to(torch.bool), torch.finfo(dtype).min) +# Copied from transformers.models.llama.modeling_llama.LlamaRotaryEmbedding with Llama->Persimmon class PersimmonRotaryEmbedding(nn.Module): def __init__(self, dim, max_position_embeddings=2048, base=10000, device=None): super().__init__()