From 7ed3adc2a771312c7329f9c2924bc0f28f716b29 Mon Sep 17 00:00:00 2001 From: Sidd Karamcheti Date: Mon, 21 Aug 2023 05:01:56 -0700 Subject: [PATCH] Lint --- open_flamingo/src/factory.py | 17 +++++++++++++---- 1 file changed, 13 insertions(+), 4 deletions(-) diff --git a/open_flamingo/src/factory.py b/open_flamingo/src/factory.py index 4ac9df40..2e23213e 100644 --- a/open_flamingo/src/factory.py +++ b/open_flamingo/src/factory.py @@ -40,16 +40,23 @@ def create_model_and_transforms( Tokenizer: A tokenizer for the language model """ vision_encoder, _, image_processor = open_clip.create_model_and_transforms( - clip_vision_encoder_path, pretrained=clip_vision_encoder_pretrained, cache_dir=cache_dir + clip_vision_encoder_path, + pretrained=clip_vision_encoder_pretrained, + cache_dir=cache_dir, ) # set the vision encoder to output the visual features vision_encoder.visual.output_tokens = True text_tokenizer = AutoTokenizer.from_pretrained( - tokenizer_path, local_files_only=use_local_files, trust_remote_code=True, cache_dir=cache_dir + tokenizer_path, + local_files_only=use_local_files, + trust_remote_code=True, + cache_dir=cache_dir, ) # add Flamingo special tokens to the tokenizer - text_tokenizer.add_special_tokens({"additional_special_tokens": ["<|endofchunk|>", ""]}) + text_tokenizer.add_special_tokens( + {"additional_special_tokens": ["<|endofchunk|>", ""]} + ) if text_tokenizer.pad_token is None: # Issue: GPT models don't have a pad token, which we use to # modify labels for the loss. @@ -87,7 +94,9 @@ def set_input_embeddings(self, new_embeddings): lang_encoder, text_tokenizer.encode("<|endofchunk|>")[-1], text_tokenizer.encode("")[-1], - vis_dim=open_clip.get_model_config(clip_vision_encoder_path)["vision_cfg"]["width"], + vis_dim=open_clip.get_model_config(clip_vision_encoder_path)["vision_cfg"][ + "width" + ], cross_attn_every_n_layers=cross_attn_every_n_layers, **flamingo_kwargs, )