-
Notifications
You must be signed in to change notification settings - Fork 71
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
NeuronModelForSentenceTransformers and NeuronStableDiffusionPipeline are not compiling with Neuron SDK 2.19 and onward #710
Comments
@yahavb thank you for your feedback. |
Thanks for the speedy response. I upgraded pytroch to
DO we know who set these flags? I tried with older DLC (deep learning containers images) and we did not needed to set those flags. |
gently pinging @JingyaHuang on this. |
Would it be possible that you set up these flags while debugging? (like for snapshotting hlo?) @yahavb. Thes flags are not set by Optimum Neuron, and in my dev env they are not set up if I install optimum neuron in a clean environment. |
I don't. could be from the DLC. No idea but I will close it as I am unblocked. |
System Info
when attempting to compile a model like https://huggingface.co/docs/optimum-neuron/en/inference_tutorials/sentence_transformers#sentence-transformers-on-aws-inferentia-with-optimum-neuron or https://huggingface.co/docs/optimum-neuron/en/inference_tutorials/stable_diffusion
The compilation aborts with the error log:
root@7938e622e5ed:/# python
Python 3.10.12 | packaged by conda-forge | (main, Jun 23 2023, 22:40:32) [GCC 12.3.0] on linux
Type "help", "copyright", "credits" or "license" for more information.
root@7938e622e5ed:/# python
Python 3.10.12 | packaged by conda-forge | (main, Jun 23 2023, 22:40:32) [GCC 12.3.0] on linux
Type "help", "copyright", "credits" or "license" for more information.
2024-10-06 16:27:46.795083: F external/xla/xla/parse_flags_from_env.cc:224] Unknown flags in XLA_FLAGS: --xla_gpu_simplify_all_fp_conversions=false --xla_gpu_force_compilation_parallelism=8
The text was updated successfully, but these errors were encountered: