Skip to content

README.md multi-GPU training instructions add bold emphasis to how numGPUs influences batch size #268

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 1 commit into
base: develop
Choose a base branch
from

Conversation

mkrupczak3
Copy link

I found this description to be incredibly subtle, I ran a lot of training runs with the wrong batch size before I saw it

Description

Simple change to README.md to add bold emphasis under Multi-GPU training instructions:

Note that your effective batch size is multiplied by the number of GPUs**, so you may need to adjust your batch_size and grad_accum_steps to maintain the same overall batch size!

No dependencies as it is just a documentation update

Type of change

Please delete options that are not relevant.

  • This change requires a documentation update

How has this change been tested, please provide a testcase or example of how you tested the change?

With my eyes reading the README.md

Any specific deployment considerations

May be worth emphasizing in documentation elsewhere to make sure others don't make the same mistake that I did

Docs

  • Docs updated? What were the changes:
    Just a simple change to README.md

…mGPUs influences batch size

I found this description to be incredibly subtle, I ran a lot of training runs with the wrong batch size before I saw it
@CLAassistant
Copy link

CLA assistant check
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.
You have signed the CLA already but the status is still pending? Let us recheck it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants