Skip to content

Commit

Permalink
fix VI: return of the fix
Browse files Browse the repository at this point in the history
  • Loading branch information
milocress committed Apr 19, 2024
1 parent a59e09c commit b7fb56a
Show file tree
Hide file tree
Showing 2 changed files with 4 additions and 5 deletions.
1 change: 0 additions & 1 deletion llmfoundry/data/packing.py
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,6 @@
import numpy as np
import torch
from composer.utils import dist
from omegaconf import DictConfig
from transformers import PreTrainedTokenizerBase

log = logging.getLogger(__name__)
Expand Down
8 changes: 4 additions & 4 deletions tests/data/test_packing.py
Original file line number Diff line number Diff line change
Expand Up @@ -107,9 +107,9 @@ def test_auto_packing(profile_packing: Mock):
profile_packing.return_value = [(1, .9, 0), (2, .8, 0), (3, .7, .5)]

packing_ratio = auto_packing_ratio(
dataloader_cfg=DictConfig({'dataset': {
dataloader_cfg={'dataset': {
'max_seq_len': 2048
}}),
}},
tokenizer=None,
device_batch_size=1,
) # Dummy values, profiling results are already set.
Expand All @@ -134,9 +134,9 @@ def test_dist_auto_packing(profile_packing: Mock):
(3, .7, .5)] # should pick 2

packing_ratio = auto_packing_ratio(
dataloader_cfg=DictConfig({'dataset': {
dataloader_cfg={'dataset': {
'max_seq_len': 2048
}}),
}},
tokenizer=None,
device_batch_size=1,
) # Dummy values, profiling results are already set.
Expand Down

0 comments on commit b7fb56a

Please sign in to comment.