You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When running the train_rlhf_llama.sh script for LLaMA2-7B, the performance gain from using vLLM as the inference backend is not significant compared to Megatron (approximately a 20% improvement). According to some other reports, the expected improvement should be around 200%. How can it be promoted?
The text was updated successfully, but these errors were encountered:
Our tests show an improvement of about 2 times. You can compare this with the configuration and performance data provided in our documentation to see if there are any differences.
When running the train_rlhf_llama.sh script for LLaMA2-7B, the performance gain from using vLLM as the inference backend is not significant compared to Megatron (approximately a 20% improvement). According to some other reports, the expected improvement should be around 200%. How can it be promoted?
The text was updated successfully, but these errors were encountered: