Skip to content

Commit

Permalink
change debugging statement
Browse files Browse the repository at this point in the history
  • Loading branch information
mattyding committed Dec 27, 2024
1 parent 6c3e0a7 commit 2712e2b
Show file tree
Hide file tree
Showing 2 changed files with 5 additions and 2 deletions.
2 changes: 0 additions & 2 deletions llmfoundry/data/finetuning/dataloader.py
Original file line number Diff line number Diff line change
Expand Up @@ -179,8 +179,6 @@ def build_finetuning_dataloader(
padding/waste rates for different `cfg.dataset.packing_ratio` choices,
given a starting workload YAML.
"""
print(f"🚨🚨🚨 build_finetuning_dataloader was called with the following arguments: tokenizer={tokenizer}, device_batch_size={device_batch_size}, dataset={dataset}, num_workers={num_workers}, drop_last={drop_last}, pin_memory={pin_memory}, prefetch_factor={prefetch_factor}, persistent_workers={persistent_workers}, timeout={timeout}")

dataset_cfg = dataset
is_streaming = (
dataset_cfg.get('remote') is not None or
Expand Down
5 changes: 5 additions & 0 deletions llmfoundry/utils/config_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -429,6 +429,7 @@ def calculate_batch_size_info(
int, Literal['auto']]]:

world_size = dist.get_world_size()
print(f"🚨🚨🚨 world_size: {world_size}")
if world_size % data_replication_degree != 0:
raise ValueError(
f'World size {world_size} is not divisible by data replication degree {data_replication_degree}.',
Expand Down Expand Up @@ -457,6 +458,10 @@ def calculate_batch_size_info(
)
else:
raise ValueError(f'Not sure how to parse {device_microbatch_size=}')

print(f"🚨🚨🚨 device_batch_size: {device_batch_size}")
print(f"🚨🚨🚨 device_microbatch_size: {device_microbatch_size}")
print(f"🚨🚨🚨 device_grad_accum: {device_grad_accum}")

return device_batch_size, device_microbatch_size, device_grad_accum

Expand Down

0 comments on commit 2712e2b

Please sign in to comment.