diff --git a/llmfoundry/__init__.py b/llmfoundry/__init__.py index 86f0bff917..f62ff11637 100644 --- a/llmfoundry/__init__.py +++ b/llmfoundry/__init__.py @@ -10,9 +10,8 @@ Seq2SeqFinetuningCollator, build_finetuning_dataloader, build_text_denoising_dataloader) - from llmfoundry.models.hf import (ComposerHFCausalLM, - ComposerHFCausalLMFromPython, - ComposerHFPrefixLM, ComposerHFT5) + from llmfoundry.models.hf import (ComposerHFCausalLM, ComposerHFPrefixLM, + ComposerHFT5) from llmfoundry.models.layers.attention import ( MultiheadAttention, attn_bias_shape, build_alibi_bias, build_attn_bias, flash_attn_fn, scaled_multihead_dot_product_attention, diff --git a/llmfoundry/models/hf/__init__.py b/llmfoundry/models/hf/__init__.py index a41f659c75..d0ab65dfaf 100644 --- a/llmfoundry/models/hf/__init__.py +++ b/llmfoundry/models/hf/__init__.py @@ -1,8 +1,7 @@ # Copyright 2022 MosaicML LLM Foundry authors # SPDX-License-Identifier: Apache-2.0 -from llmfoundry.models.hf.hf_causal_lm import (ComposerHFCausalLM, - ComposerHFCausalLMFromPython) +from llmfoundry.models.hf.hf_causal_lm import ComposerHFCausalLM from llmfoundry.models.hf.hf_fsdp import (prepare_hf_causal_lm_model_for_fsdp, prepare_hf_enc_dec_model_for_fsdp, prepare_hf_model_for_fsdp) @@ -11,7 +10,6 @@ __all__ = [ 'ComposerHFCausalLM', - 'ComposerHFCausalLMFromPython', 'ComposerHFPrefixLM', 'ComposerHFT5', 'prepare_hf_causal_lm_model_for_fsdp',