-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
generate.py not utilizing GPU in full #476
Comments
Have you tried increasing the batch size? |
Thanks for the reply. I wonder how should I increase the batch size during generation? In the configuration file it has batch size for training, such as "train_micro_batch_size_per_gpu" |
same setting in inference |
@frankxu2004 @sdtblck Changing train_micro_batch_size_per_gpu doesnt work out for Inference mode("Input-file") should I change any other parameter? |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I tried to run text generation with prompts using
generate.py
. I provided a large list of prompts, approximately 20K, and tried to run the generation on 10 RTX 8000 GPUs. However, the GPU utilization by nvidia-smi shows that the GPU utilization during generation is averaging at about 50-60%, which is not ideal. Thank you!My configuration is:
And the model config:
The text was updated successfully, but these errors were encountered: