Differences
This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision | ||
gpu_resources [2017/05/05 15:25] – [nvidia-smi flags used] csteel | gpu_resources [2024/03/26 13:52] (current) – external edit 127.0.0.1 | ||
---|---|---|---|
Line 28: | Line 28: | ||
===== Preventing Job Clobbering ===== | ===== Preventing Job Clobbering ===== | ||
- | Today I was training a model and inadvertently kicked Konrad' | + | There are currently 3 GPU' |
+ | |||
+ | < | ||
+ | export CUDA_VISIBLE_DEVICES=X | ||
+ | </ | ||
+ | |||
+ | This will only take effect when you log in, so log out and back in and try the following to ensure | ||
+ | |||
+ | < | ||
+ | echo $CUDA_VISIBLE_DEVICES | ||
+ | </ | ||
+ | |||
+ | If it outputs the ID that you selected then you're ready to use the GPU. | ||
+ | |||
+ | ==== Sharing a single GPU ==== | ||
+ | To configure TensorFlow to not pre-allocate all GPU memory you can use the following Python code: | ||
< | < | ||
Line 38: | Line 53: | ||
</ | </ | ||
- | We should develop some kind of policy | + | This has been found to work only to a certain extent, and when there are several |
===== GPU Info ===== | ===== GPU Info ===== | ||
Line 70: | Line 84: | ||
nsight | nsight | ||
</ | </ | ||
+ | |||
+ | Nvidia Visual Profiler (https:// | ||
+ | < | ||
+ | / | ||
+ | </ | ||
+ | |||
===== GPU Accounting ===== | ===== GPU Accounting ===== |