-
Notifications
You must be signed in to change notification settings - Fork 33
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ModuleNotFoundError: No module named 'flash_attn.models.falcon' #22
Comments
Just pushed an update that should fix the import issue. Could you add more details on the preprocess issue? |
tks for ur help! I will try it later. The preprocessing issue is as follows: it can be successfully run by updating
|
Hello, another error when I run the lastest code.
Wandb has been init ,and have this error. I run the code with a 8*A100 GPUs machine. It something wrong with my config? or the wrong version I used.(maybe version with also, the training args have a sooooooo big num epochs, but the num epoch in
|
Yes it's because of the different version of The num_epochs should be ignored - the training terminates on steps. |
tksss for ur help!! I can run the
Sorry for asking so many run-related questions. I really can't resolve these issues.😭 |
I'm not totally sure, but sometimes these errors can be mitigated by uninstalling and doing a fresh install of torch / flash-attn. You could also try running on CPU or with CUDA_LAUNCH_BLOCKING=1 to see if it will give a better trace. It could also be worth checking your CUDA version vs. which CUDA version your pytorch is compiled with. |
The code can't run with CPU mode. And I have rebuilt the conda environment for several times.
Is this right? I can run baseline successful, but wrong in doremi120M |
Do you have a more detailed trace? What part of the code raises this error? |
sure,there are the detail error message. run on a *8A100 server
|
I face the same issue, but it worked when I lowered batch_size. |
I ran the
bash scripts/setup_flash.sh
without error (but it cost just a few minute)But I got a wrong message when I run the
bash scripts/run_pile.sh
what wrong with this?
also,I found something wrong with
bash scripts/run_preprocess_pile.sh
until I update the packagesdatasets
to 2.15.0 , but version insetup.py
is 2.10.1.Is something wrong in my operate?
The text was updated successfully, but these errors were encountered: