-
Notifications
You must be signed in to change notification settings - Fork 536
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Nvidia 4.1 inference code is giving segmentation fault for RTX 4090 (4.0 code works fine) #1847
Comments
Hi, May I ask if it is possible to test this on fp32 model? Specifically, change
then I was unable to build the int8 models on RTX3090 but fp16/32 models worked and got them running. |
That's great. But the same code didn't work for me - still getting segmentation fault for |
That's strange.... personally i think the error is due to the model but being quantized/exported properly or the data was not properly preprocessed. If you still have the commands history, is it possible for you to send all |
@Oseltamivir For me, the below line is the culprit. Commenting it out makes it work for me - though I haven't checked everything. I'll also try on different systems. |
In that case it might be a issue caused by mitten. But the maintainers don't seem to respond to issues posted in the mitten repo. I opened an issue there 3 months ago but got no reply. I ended up having to email Yiheng to ask about their implementation of mitten. |
yes, it is. We'll come back to it as currently we are using Nvidia v4.0 code to collect the inference results via github actions. |
Trying to run Nvidia v4.1 implementation for stable diffusion on RTX 4090.
ran successfully and produced int8 model. Below are the custom configs used for 2x RTX4090.
The text was updated successfully, but these errors were encountered: