Frequently asked questions¶
Does the tensor data consume extra memory when compiled into C++ code?¶
When compiled into C++ code, the tensor data will be mmaped by the system loader. For the CPU runtime, the tensor data are used without memory copy. For the GPU and DSP runtime, the tensor data are used once during model initialization. The operating system is free to swap the pages out, however, it still consumes virtual memory addresses. So generally speaking, it takes no extra physical memory. If you are short of virtual memory space (this should be very rare), you can use the option to load the tensor data from data file (can be manually unmapped after initialization) instead of compiled code.
Why is the generated static library file size so huge?¶
The static library is simply an archive of a set of object files which are intermediate and contain much extra information, please check whether the final binary file size is as expected.
How to set the input shape in your model deployment file(.yml) when your model support multiple input shape?¶
Set the largest input shape of your model. The input shape is used for memory optimization.
OpenCL allocator failed with CL_OUT_OF_RESOURCES¶
OpenCL runtime usually requires continuous virtual memory for its image buffer, the error will occur when the OpenCL driver can't find the continuous space due to high memory usage or fragmentation. Several solutions can be tried:
- Change the model by reducing its memory usage
- Split the Op with the biggest single memory buffer
- Change from armeabi-v7a to arm64-v8a to expand the virtual address space
- Reduce the memory consumption of other modules of the same process
Why is the performance worse than the official result for the same model?¶
The power options may not set properly, see
Why is the UI getting poor responsiveness when running model with GPU runtime?¶
Try to set
1. If still not resolved, try to
modify the source code to use even smaller time intervals or changed to CPU
or DSP runtime.
For GPUs such as Arm Mali, sometimes even setting
a small time interval can not solve the problem. At this time, you can try to
opencl_queue_window_size, such as 16. This parameter means that the GPU
command queue will contain only
opencl_queue_window_size commands at most.
You can adjust this parameter to achieve a balance between performance and UI
response. You should not use this parameter unless you have to.
Why is MACE not working on DSP?¶
Running models on Hexagon DSP need a few prerequisites for DSP developers:
- You need to make sure SOCs of your phone is manufactured by Qualcomm and has HVX supported.
- You need a phone that disables secure boot (once enabled, cannot be reversed, so you probably can only get that type phones from manufacturers)
- You need to root your phone.
- You need to sign your phone by using testsig provided by Qualcomm. (Download Qualcomm Hexagon SDK first, plugin your phone to PC, run scripts/testsig.py)
- You need to push
Then, there you go. You can run Mace on Hexagon DSP.