What GPU model do we have? On what hosts?
The Analytics team added a GPU to stat1005 and one to stat1008. The model is AMD Radeon Pro WX 9100 16GB. The choice fell to AMD since they are currently the only ones releasing their software stack open source: https://rocm.github.io/ROCmInstall.html
Use the Debian packages
Use the GPU on the host
All users in
analytics-privatedata-users are automatically granted access to the GPUs, otherwise a user need to be in the
gpu-testers POSIX group in operations/puppet. This is a workaround to force the users in that group to be in the
render POSIX group (available on Debian), that grants access to the GPU. Please keep in mind few things:
- Be careful in launching multiple parallel jobs on the same GPU, see https://phabricator.wikimedia.org/T248574
- Miriam Redi is currently the main point of contact to decide what the schedule of GPU usage should be. In case of doubt, before starting any test or heavy job, please follow up with her.
The easiest solution is to create a Python 3 virtual environment on stat1005 or stat1008 and then pip3 install https://pypi.org/project/tensorflow-rocm/. Please remember that every version of the package is linked against a specific version of ROCm, so it may be possible that newer versions of tensorflow-rocm don't run on our hosts since we don't have an up to date version of ROCm deployed yet.
Upstream suggested to follow https://github.com/ROCmSoftwarePlatform/tensorflow-upstream/blob/develop-upstream/RELEASE.md and check every time what combination of tensorflow-rocm and ROCm is supported.
With the current version of ROCm, 3.8, only tensorflow-rocm 2.3.1 is supported.
virtualenv -p python3 test_tf source test_tf/bin/activate pip3 install tensorflow-rocm==2.3.1
Configure your Tensorflow script
By default, Tensorflow tasks take all available resources (both from the CPU and the GPU). In resource sharing settings, this might cause resources to saturate quickly and some process to block before execution. When using Tensorflow scripts on our GPU machines, please make sure you add to your code the following snippet:
For Tensorflow version 2.0 and 2.1:
import tensorflow as tf gpu_devices = tf.config.experimental.list_physical_devices('GPU') tf.config.experimental.set_memory_growth(gpu_devices, True)
import tensorflow as tf tf.config.gpu.set_per_process_memory_growth(True)
For prior versions:
import tensorflow as tf tf_config=tf.ConfigProto() tf_config.gpu_options.allow_growth=True sess = tf.Session(config=tf_config)
Also, a good practice is to limit the number of threads used by your tensorflow code.
For Tensorflow version 2.0 and 2.1:
import tensorflow as tf tf.config.threading.set_intra_op_parallelism_threads(10) #or lower values tf.config.threading.set_inter_op_parallelism_threads(10) #or lower values
For prior versions:
import tensorflow as tf tf_config=tf.ConfigProto(intra_op_parallelism_threads=10,inter_op_parallelism_threads=10) sess = tf.Session(config=tf_config)
Check the version of ROCm deployed on a host
elukey@stat1005:~/test$ dpkg -l rocm-dev Desired=Unknown/Install/Remove/Purge/Hold | Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad) ||/ Name Version Architecture Description +++-==============-============-============-================================================= ii rocm-dev 2.7.22 amd64 Radeon Open Compute (ROCm) Runtime software stack
elukey@stat1008:~$ dpkg -l rocm-dev Desired=Unknown/Install/Remove/Purge/Hold | Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad) ||/ Name Version Architecture Description +++-==============-============-============-================================================= ii rocm-dev 3.3.0-19 amd64 Radeon Open Compute (ROCm) Runtime software stack
Please note: due to https://phabricator.wikimedia.org/T247082 on stat1008 we are currently testing a new version of ROCm!
Check usage of the GPU
On the host (limited to analytics-privatedata-users):
elukey@stat1005:~$ sudo radeontop
https://github.com/ROCmSoftwarePlatform/tensorflow-upstream/issues/559 GPUs are not correctly handling multi-tasking - https://phabricator.wikimedia.org/T248574
Reset the GPU state
If the GPU gets stuck for some reason (unclean job completion, etc..) the following may happen:
- radeontop shows steady RAM usage (90%+ for example).
- tensorflow gets stuck when trying to execute jobs.
Usually rebooting the host works, but the following procedure might help as well:
sudo /opt/rocm/bin/rocm-smiand get the id of the GPU (usually 1)
sudo /opt/rocm/bin/rocm-smi --gpureset -d X(with X equals to the id of the GPU)
Upgrade the Debian packages
We import the Debian packages released by AMD for Ubuntu Xenial to the amd-rocm component in wikimedia-buster. Up to now (July 2019) there is one Debian package released by AMD that is not open source,
hsa-ext-rocr-dev. It contains binary libraries to have a better image support in OpenCL, and we don't use it for obvious reasons. The package is sadly required by other packages, and upstream still hasn't made it optional (https://github.com/RadeonOpenCompute/ROCm/issues/761).
The solution found in https://phabricator.wikimedia.org/T224723 was to create a dummy package via Debian equiv to satisfy dependencies and please the apt install process. This means that every time a new ROCm release is out, the following procedure needs to be done:
Before starting, please check https://github.com/RadeonOpenCompute/ROCm, there is a changelog for every version. Pay attention for breaking changes and OS supported.
1) Check http://repo.radeon.com/rocm/apt/ and see if a new version is out. If so, create a new component like:
Name: amd-rocmXX Method: http://repo.radeon.com/rocm/apt/X.X/ Suite: xenial Components: main>thirdparty/amd-rocmXX UDebComponents: Architectures: amd64 VerifyRelease: 9386B48A1A693C5C ListShellHook: grep-dctrl -e -S '^([..cut..])$' || [ $? -eq 1 ]
Replace the XX wildcards with the version number of course.
2) ssh to apt1001, run puppet and check for updates (remember to replace the XX wildcards):
root@apt1001:/srv/wikimedia# reprepro --noskipold --ignore=forbiddenchar --component thirdparty/amd-rocmXX checkupdate buster-wikimedia Calculating packages to get... Updates needed for 'buster-wikimedia|thirdparty/amd-rocm|amd64': [..] 'hsa-rocr-dev': newly installed as '1.1.9-87-g1566fdd' (from 'amd-rocm'): files needed: pool/thirdparty/amd-rocm/h/hsa-rocr-dev/hsa-rocr-dev_1.1.9-87-g1566fdd_amd64.deb [..]
3) find the new version of
hsa-rocr-dev, since it is the only package in ROCm that requires a precise version of the
hsa-ext-rocr-dev package (namely its version).
3) create a control file like the following on boron:
### Commented entries have reasonable defaults. ### Uncomment to edit them. # Source: <source package name; defaults to package name> Section: devel Priority: optional # Homepage: <enter URL here; no default> Standards-Version: 3.9.2 Package: hsa-ext-rocr-dev Version: 1.1.9-87-g1566fdd Maintainer: Luca Toscano <firstname.lastname@example.org> # Pre-Depends: <comma-separated list of packages> # Depends: <comma-separated list of packages> # Recommends: <comma-separated list of packages> # Suggests: <comma-separated list of packages> # Provides: <comma-separated list of packages> # Replaces: <comma-separated list of packages> Architecture: amd64 # Multi-Arch: <one of: foreign|same|allowed> # Copyright: <copyright file; defaults to GPL2> # Changelog: <changelog file; defaults to a generic changelog> # Readme: <README.Debian file; defaults to a generic one> # Extra-Files: <comma-separated list of additional files for the doc directory> # Files: <pair of space-separated paths; First is file to include, second is destination> # <more pairs, if there's more than one file to include. Notice the starting space> Description: dummy package to satisfy dependencies for hsa-rocr-dev hsa-rocr-dev-ext contains binary only and non open-source libraries .
Make sure the Version is the new one of
hsa-rocr-dev and save.
4) build the package with
5) upload the package to reprepro (remember to replace the XX wildcards):
reprepro -C thirdparty/amd-rocmXX includedeb buster-wikimedia /home/elukey/hsa-ext-rocr-dev_1.1.9-87-g1566fdd_amd64.deb
6) Update the
thirdparty/rocmXX component (remember to replace the XX wildcards):
reprepro --noskipold --ignore=forbiddenchar --component thirdparty/amd-rocmXX update buster-wikimedia
7) Update the versions supported by the amd_rocm module in operations/puppet. 8) On the host that you want to upgrade:
sudo apt autoremove -y rocm-smi miopengemm rocminfo hsakmt-roct hsa-rocr-dev rocm-cmake hsa-ext-rocr-dev rocm-device-libs hip_base hip_samples
And then run puppet to install the new packages. Some quick tests to see if the GPU is properly recognized:
elukey@stat1005:~$ /opt/rocm/bin/rocminfo [..] elukey@stat1005:~$ /opt/rocm/opencl/bin/x86_64/clinfo [..] elukey@stat1005:~$ virtualenv -p python3 test elukey@stat1005:~$ source test/bin/activate (test) elukey@stat1005:~$ pip3 install tensorflow-rocm (test) elukey@stat1005:~$ cat test_gpu_usage.py import tensorflow as tf # Creates a graph. with tf.device('/device:GPU:0'): a = tf.constant([1.0, 2.0, 3.0, 4.0, 5.0, 6.0], shape=[2, 3], name='a') b = tf.constant([1.0, 2.0, 3.0, 4.0, 5.0, 6.0], shape=[3, 2], name='b') c = tf.matmul(a, b) # Creates a session with log_device_placement set to True. sess = tf.Session(config=tf.ConfigProto(log_device_placement=True)) # Runs the op. print(sess.run(c)) (test) elukey@stat1005:~$ python3 test_gpu_usage.py [..] [[22. 28.] [49. 64.]] elukey@stat1008:~$ cat gpu_test.py import tensorflow as tf # Creates a graph. with tf.device('/device:GPU:0'): a = tf.constant([1.0, 2.0, 3.0, 4.0, 5.0, 6.0], shape=[2, 3], name='a') b = tf.constant([1.0, 2.0, 3.0, 4.0, 5.0, 6.0], shape=[3, 2], name='b') c = tf.matmul(a, b) # Runs the op. print(c) (test) elukey@stat1008:~$ python gpu_test.py tf.Tensor( [[22. 28.] [49. 64.]], shape=(2, 2), dtype=float32)