mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00
Clarify batch size displayed when using DataParallel (#24430)
This commit is contained in:
parent
b6295b26c5
commit
2834c17ad2
@ -1671,7 +1671,9 @@ class Trainer:
|
||||
logger.info("***** Running training *****")
|
||||
logger.info(f" Num examples = {num_examples:,}")
|
||||
logger.info(f" Num Epochs = {num_train_epochs:,}")
|
||||
logger.info(f" Instantaneous batch size per device = {self._train_batch_size:,}")
|
||||
logger.info(f" Instantaneous batch size per device = {self.args.per_device_train_batch_size:,}")
|
||||
if self.args.per_device_train_batch_size != self._train_batch_size:
|
||||
logger.info(f" Training with DataParallel so batch size has been adjusted to: {self._train_batch_size:,}")
|
||||
logger.info(f" Total train batch size (w. parallel, distributed & accumulation) = {total_train_batch_size:,}")
|
||||
logger.info(f" Gradient Accumulation steps = {args.gradient_accumulation_steps}")
|
||||
logger.info(f" Total optimization steps = {max_steps:,}")
|
||||
|
Loading…
Reference in New Issue
Block a user