-
Notifications
You must be signed in to change notification settings - Fork 98
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] Could not run DirectML models on Intel laptop #70
Comments
+1 from my side This is caused by your BIOS only preallocating ~64MB of RAM to your iGPU. All other memory is being allocated dynamically. I have an RX6400 with 4GB of VRAM, yet the AI Dev Gallery detects it as 3.9GB of VRAM, thus the check fails For now, you can patch the check yourself by inputting a fake value
After patching this single file, Phi 3 Mini runs fine on my system. |
I agree we need to enable DML models on integrated GPUs. We won't have time to work on this this month and can add it for next month, but if anyone wants to pick it up before then please feel free to submit a PR. |
Related to #47 I guess Having a "download anyway" button could be an easy fix |
@nmetulev Unless the dev is using a dGPU with enough VRAM dedicated, the system always has to use to shared memory anyway. |
Good point. I had decided to only check dedicated vram because I noticed that language models have a degraded performance when they overflow to shared memory, and in some cases, it would cause blue screens. This is likely a DML bug. However, that should not be the case when using just the shared vram on integrated gpu, so we should be able to just check for either one, and if no dGPU, then check the shared vram. And I like the suggestion from @BobLd of having a "download anyway" option where the user is presented with a warning and they agree to the risks. Thoughts? |
I created microsoft/DirectML#683. Feel free to add to it. @nmetulev Is this bug reported and tracked in Microsoft already? If not, can you please ensure that it is? Lots of models can't be run on the GPU due to this issue (because dedicated VRAM is often too small) and CPU is a factor slower, so having this fixed would be really welcome. |
Describe the bug
My laptop doesn't have any discrete GPU, but an Intel integrated GPU (Meteor Lake). However, I have 32G system memory, and Intel iGPU may use up to half of it. The application disallows to download DirectML models.
To Reproduce
Steps to reproduce the behavior:
Expected behavior
DirectML based models could run on integrated GPUs.
Screenshots

Please complete the following information:
Self-built from GitHub
Additional context
NA
The text was updated successfully, but these errors were encountered: