This page contains information regarding the services that are provided internally to Toolforge nodes.
Deployment components and architecture
Information on how the setup is deployed, and the different components.
Usually a VM with a cinder volume to store repository data.
Addressing, DNS and proxy
There is an horizon web proxy called deb-tools.wmcloud.org that should point to TCP/80 on the server. This allows to build docker images using toolforge internal packages.
Other than that, servers don't have any special DNS or adressing. The don't have floating IPs.
Worth noting that these servers in the tools cloudvps project may offer services for the toolsbeta project as well.
The main role in use is role::wmcs::toolforge::services.
It gets a list of tools (accounts starting with "tools."), reads their
toolinfo.json files and adds it to the
tools table in the
toollabs_p database. Maintainer information is retrieved by getting all users that belong to the tool's group and using
getpwnam() to retrieve user information, which then gets added to the
This script runs, as a service, from the active
tools-services-* server, and wakes up every 120 seconds to populate the tables with new data.
The database in use is
tools.labsdb which is
One of the main purposes of this service is to host Debian packages for other servers by means of aptly.
Repositories are declared in puppet, but packages should be added to the aptly repository by hand.
We usually have one repository per operating system and project, i.e:
Quick example of packages being stored here are:
The repository data, located at /srv/packages is stored in a mounted cinder volume.
Information on maintenance and administration of this setup.
managing aptly repo
Is managed as a standard aptly repo.
Some interesting bits to check if you want to know the status/health of the server.
- aptly repos are present, and they contain packages, i.e:
sudo aptly repo listand
sudo aptly repo show --with-packages=true stretch-tools
- disk is not filled, i.e:
df -h /
We don't have a specific failover mechanism rather than building a new VM and re-attach the cinder volume.
Care should be taken to don't loss aptly repo data, since generating it from scratch can take some time.
This was heavily remodeled when migrating the grid to SGE and to Stretch. Previous to the migration, the services nodes used to store Bigbrother (deprecated), and webservicemonitor (moved to cron servers).
Again, when migrating from Stretch to Buster, the 2 VM approach was dropped in favor of storing the data in a cinder volume, see https://phabricator.wikimedia.org/T278354.