Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

RuntimeError: CUDA error: CUBLAS_STATUS_INVALID_VALUE when calling cublasSgemm( handle, opa, opb, m, n, k, &alpha, a, lda, b, ldb, &beta, c, ldc) #22

Open
dandeperson opened this issue Jun 8, 2023 · 0 comments

Comments

@dandeperson
Copy link

dandeperson commented Jun 8, 2023

在模型lora微调时编写shell脚本如下,出现图中错误,错误的原因是什么,如何修改?
python train.py
--load_model "/home/intern/workspace/LLM/RWKV-LM-LoRA/weight/RWKV-4-Pile-3B-Chn-testNovel-done-ctx2048-20230312.pth"
--proj_dir "/home/intern/workspace/LLM/RWKV-LM-LoRA/weight/lora"
--data_file "/home/intern/workspace/LLM/RWKV-LM-LoRA/RWKV-v4neo/data/firefly-text_text_document"
--data_type "binidx"
--vocab_size 50277
--ctx_len 2048
--epoch_steps 1000 --epoch_count 1000
--epoch_begin 0 --epoch_save 5 --micro_bsz 2
--n_layer 32 --n_embd 2560
--pre_ffn 0 --head_qk 0 --lr_init 1e-4 --lr_final 1e-4
--warmup_steps 0 --beta1 0.9 --beta2 0.999 --adam_eps 1e-8
--accelerator gpu --devices 1 --precision bf16 --strategy deepspeed_stage_2 --grad_cp 0
--lora --lora_r 8 --lora_alpha 16 --lora_dropout 0.01
--lora_parts=att,ffn,time,ln
image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant