From 4189276779bb9e84c20422e11d522c8c9e84f8e8 Mon Sep 17 00:00:00 2001 From: fxmarty <9808326+fxmarty@users.noreply.github.com> Date: Fri, 4 Aug 2023 18:43:22 +0900 Subject: [PATCH] Fix BetterTransformer starcoder init (#1254) fix bug --- optimum/bettertransformer/models/decoder_models.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/optimum/bettertransformer/models/decoder_models.py b/optimum/bettertransformer/models/decoder_models.py index 8e79e2b10d..b52bb4d7bd 100644 --- a/optimum/bettertransformer/models/decoder_models.py +++ b/optimum/bettertransformer/models/decoder_models.py @@ -370,7 +370,7 @@ class GPTBigCodeAttentionLayerBetterTransformer(BetterTransformerBaseLayer, GPTB def __init__(self, layer: nn.Module, config: "PretrainedConfig"): with torch.device("meta"): - super(BetterTransformerBaseLayer, self).__init__(config) + super(BetterTransformerBaseLayer, self).__init__(config, layer.is_cross_attention, layer.layer_idx) self.module_mapping = None submodules = ["c_attn", "c_proj"]