diff --git a/litgpt/utils.py b/litgpt/utils.py index a0f16ce213..e682038159 100644 --- a/litgpt/utils.py +++ b/litgpt/utils.py @@ -560,6 +560,7 @@ def instantiate_torch_optimizer(optimizer, model_parameters, **kwargs): # Special care taken where some optimizers do not have parameter "fused" like: # bnb.optim.AdamW8bit # grokadamw.GrokAdamW + # torch.optim.RMSprop if isinstance(optimizer, str): if "." in optimizer: