From 129dcbe4d2aef094b8b9a9d61a9fa2a3f0476995 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Thu, 19 Sep 2024 22:21:00 +0000 Subject: [PATCH] Update wandb requirement from <0.18,>=0.13.2 to >=0.13.2,<0.19 (#3615) Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Saaketh Narayan Co-authored-by: Mihir Patel --- setup.py | 2 +- tests/loggers/test_wandb_logger.py | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/setup.py b/setup.py index befb663b98..6cc65702a7 100644 --- a/setup.py +++ b/setup.py @@ -160,7 +160,7 @@ def package_files(prefix: str, directory: str, extension: str): ] extra_deps['wandb'] = [ - 'wandb>=0.13.2,<0.18', + 'wandb>=0.13.2,<0.19', ] extra_deps['comet_ml'] = [ diff --git a/tests/loggers/test_wandb_logger.py b/tests/loggers/test_wandb_logger.py index e190e39663..b0462fc842 100644 --- a/tests/loggers/test_wandb_logger.py +++ b/tests/loggers/test_wandb_logger.py @@ -269,10 +269,10 @@ def test_wandb_log_metrics(test_wandb_logger): eval_metrics_cross_entropy_count = all_run_text.count('metrics/eval/CrossEntropy') train_loss_count = all_run_text.count('loss/train/total') - expected_number_train_loss_count = (dataset_size / batch_size) + 1 # wandb includes it in the file one extra time + expected_number_train_loss_count = (dataset_size / batch_size) * 2 # wandb includes it twice per step expected_number_train_metrics_count = ( dataset_size / batch_size - ) + 2 # wandb includes it in the file two extra times + ) * 2 + 2 # wandb includes it twice per step plus two extra times expected_number_eval_metrics_count = 2 # wandb includes it in the file twice assert train_metrics_accuracy_count == expected_number_train_metrics_count assert train_loss_count == expected_number_train_loss_count