You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Nov 17, 2023. It is now read-only.
* Add mx.context.gpu_memory_info() to python api for flexible tests.
* Add test_gluon_gpu.py:test_large_models to show cudnnFind headroom issue.
* Output model sizes tried by test_gluon_gpu.py:test_large_models.
* Fix perl interface to MXGetGPUMemoryInformation.
* Increase difficulty of test_gluon_gpu.py:test_large_models.
* Forgot a file in fix for perl.
* Modify test to pass on no-cudnn CI runner.
* Mutex algo reg updates, serialize cudnnFind calls.
* Fix for cudnnFind memory headroom issue.
* Fix cpplint.
* Respond to reviewers comments.
* Guard against improper MXNET_GPU_MEM_LARGE_ALLOC_ROUND_SIZE values.
* Fix potentially unassigned var.
- The cutoff threshold that decides the rounding strategy. Let's denote the threshold as T. If the memory size is smaller than `2 ** T` (by default, it's 2 ** 24 = 16MB), it rounds to the smallest `2 ** n` that is larger than the requested memory size; if the memory size is larger than `2 ** T`, it rounds to the next k * 2 ** T.
70
+
* MXNET_GPU_MEM_LARGE_ALLOC_ROUND_SIZE
71
+
- Values: Int ```(default=2097152)```
72
+
- When using the naive pool type, memory allocations larger than this threshhold are rounded up to a multiple of this value.
73
+
- The default was chosen to minimize global memory fragmentation within the GPU driver. Set this to 1 to disable.
0 commit comments