-
Notifications
You must be signed in to change notification settings - Fork 228
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Inference nondeterminisms can break GPU compilation #1508
Comments
Thanks for the MWE. I'm not sure we can do much about this though; inference/compilation being dependent on additional methods being defined can be caused by an heuristic kicking in after a limit has been exceeded, or on 'bugs' like JuliaLang/julia#35800. Here, the code pulled in by RandomNumbers.AbstractRNG seems to be pretty massive, relying on device-side broadcast which fails to compile statically after those method definitions. Simplifying that code might be one course of action. That said, I also noticed that the code works again on 1.8, so maybe that's an easier solution? |
I believe I'm having the same issue with the DCGAN example in the Flux model zoo. I haven't yet been able to resolve it. |
FYI, I had the same error when using |
Describe the bug
After a system and Julia update I can't train vgg_cifar10.jl from the Flux model zoo anymore. It generates this error:
Stacktrace (from the MWE below)
To reproduce
I discussed that on discourse and was able to reduce the MWE to that:
But if the
Random.rand!
instruction is done first, then the error does not happen:The actual lines in the original packages that trigger the error are:
Manifest.toml
Expected behavior
No error.
Version info
Details on Julia:
Details on CUDA:
The text was updated successfully, but these errors were encountered: