-
Notifications
You must be signed in to change notification settings - Fork 251
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
多卡并行报错:Expected all tensors to be on the same device, but found at least two devices, cuda:1 and cuda:0! #419
Comments
我在本地测试是可以正常run的,但是也会爆显存。 |
小于等于2卡就能正常run但是爆显存,大于等于3卡就报上述错误hhh,我后面再试试看看能不能解决 |
hi, do you have any further issues? |
Right now, no. |
@Darknessrky @XeeKee @zxlzr Do you have any better solutions? thank you |
Thank you very much for your attention to easyedit. After testing, we found that 2 GPUs are working fine, but 4 GPUs cause a bug. We are currently working on debugging it, but due to my busy schedule, it may take until this weekend to fix the issue.We sincerely apologize for the inconvenience caused. |
Not yet. I'm running this on A40(48G) now is fine. |
Thank you for your in time response, I'm looking forward to the revised version. |
In editor.py line 140
There should be something wrong here. |
You can inspect the devices of the parameters with
|
Thank you very much for your help. I will try your code tonight. |
Hello, we have fixed this bug. You are welcome to pull the latest code and use it. |
I tried updated code in ./easyeditor/models/memit, still face the same error |
11.14 修改了错别字和表述问题
由于7b和3b模型用双卡3090在跑的时候都会出现第一张卡爆显存的情况,在4卡3090上跑基于llama-7b的MEMIT后出现如题报错,查看类似issue后未能解决问题。
代码如下:
超参如下:
报错如下:
烦请告知是哪里出了问题或者我需要做哪些更改,谢谢!
The text was updated successfully, but these errors were encountered: