-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: High RAM usage in iGPU #28009
Comments
Hi @yaniv5678 , how did you check the memory usage when compiling the model with CPU and GPU? Meanwhile, GPU performance relies on the OpenCL kernels for the implementation. You can refer GPU Performance Checklist. |
Hi @Aznie-Intel! Thanks for your prompt response. I've checked using the task manager. I made sure to only read and compile the model, and then put the process to sleep to be sure it isn't related to other code.
I've referred the GPU perf checklist, thanks. Do you know what is the memory layout of these 500MB RAM? |
@yaniv5678 Below is my observation for both CPU and GPU. There is no significant memory consumption compared to CPU and GPU. Can you provide the following information for further investigation:
|
|
@yaniv5678 Thanks for the information. I will check this with the relevant team and update you soon. |
Ref. 159902 |
OpenVINO Version
2024.5.0
Operating System
Windows System
Device used for inference
GPU
Framework
None
Model used
deberta-v3-mini
Issue description
Hi,
I converted deberta-v3-mini to OpenVINO using optimum-cli, with weights compressed to int8. The file size on disk is ~160MB.
And then compiled the model using both pythonic openvino & openvino-rs.
In both scenarios, model took ~500MB of RAM.
When I used inference precision hint of "int8" to my iGPU (intel iris xe, with core i5)
It didn't help, it even took more RAM (around 1.2GB!!)
When I compiled this model to CPU, it only took ~40MB somehow.
Can you help me understand why, and how to decrease RAM usage in GPU case?
Is this a bug?
Thanks!
Step-by-step reproduction
Relevant log output
No response
Issue submission checklist
The text was updated successfully, but these errors were encountered: