-
Notifications
You must be signed in to change notification settings - Fork 5
/
Copy pathslurm.yml
118 lines (109 loc) · 4.13 KB
/
slurm.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
slurm_cgroup_config:
CgroupMountpoint: "/sys/fs/cgroup"
CgroupAutomount: yes
ConstrainCores: yes
TaskAffinity: no
ConstrainRAMSpace: no
ConstrainSwapSpace: no
ConstrainDevices: no
AllowedRamSpace: 99.5
AllowedSwapSpace: 0
MaxRAMPercent: 99.5
MaxSwapPercent: 100
MinRAMSpace: 30
slurm_config:
ClusterName: usegalaxy
SchedulerType: sched/backfill
TaskPlugin: task/cgroup #,task/affinity
#TaskPluginParam: Sched
SelectType: select/cons_res
ProctrackType: proctrack/cgroup
JobAcctGatherType: jobacct_gather/cgroup
AccountingStorageType: accounting_storage/slurmdbd
AccountingStorageHost: "{{ galaxy_host }}"
ReturnToService: 1
SelectTypeParameters: CR_Core_Memory
ControlMachine: "{{ galaxy_host }}"
SlurmctldLogFile: /var/log/slurm/slurmctld.log
SlurmdLogFile: /var/log/slurm/slurmd.log
SlurmdDebug: 3
DefMemPerCPU: 1000
SrunPortRange: "{{ slurm_SrunPortRange }}"
slurmdbd_config:
AuthType: auth/munge
DbdAddr: "{{ galaxy_host }}"
DbdHost: "{{ galaxy_host }}"
SlurmUser: slurm
LogFile: /var/log/slurm/slurmdbd.log
PidFile: /var/run/slurmdbd.pid
StorageType: accounting_storage/mysql
StorageHost: " {{ db_host }}"
StoragePort: 3306
StoragePass: "{{ mysql_slurm_password }}"
StorageUser: "{{ mysql_slurm_username }}"
StorageLoc: slurm_acct_db
slurm_nodes_old:
- name: "{{ slurm_host }}"
CoresPerSocket: "{{ hostvars[slurm_host].ansible_processor_cores }}"
Sockets: "{{ hostvars[slurm_host].ansible_processor_count }}"
ThreadsPerCore: "{{ hostvars[slurm_host].ansible_processor_threads_per_core }}"
CPUs: "{{ hostvars[slurm_host].ansible_processor_vcpus }}"
RealMemory: "{{ hostvars[slurm_host].ansible_memtotal_mb }}"
slurm_partitions:
- name: "usegalaxy_{{ env }}"
Default: YES
MaxTime: UNLIMITED
Nodes: "{{ groups[ 'slurm' ] | join(',') }}"
slurm_munge_key: files/munge/munge.key
slurm_node_ips: "{{ groups['slurm'] | map('extract', hostvars, ['ansible_eth0', 'ipv4', 'address']) | join(',') }}"
# Needs to be /data/part0 as it is not possible to share a drive with a mounted drive in it
# Will need to extend this when/if new drives come online
nfs_exports:
- path: "/srv/galaxy"
# export: "{{ slurm_node_ips }}(rw,async,no_root_squash)"
export: "{{ groups['slurm'] | map('extract', hostvars, ['ansible_eth0', 'ipv4', 'address']) | join('(rw,async,no_root_squash) ')}}(rw,async,no_root_squash)"
- path: "/data/part0"
# export: "{{ slurm_node_ips }}(rw,sync,no_root_squash)"
export: "{{ groups['slurm'] | map('extract', hostvars, ['ansible_eth0', 'ipv4', 'address']) | join('(rw,sync,no_root_squash) ')}}(rw,sync,no_root_squash)"
- path: "/cvmfs/data.galaxyproject.org"
export: "*(ro,sync,no_root_squash,no_subtree_check,fsid=101)"
- path: "/cvmfs/data.usegalaxy.no"
export: "*(ro,sync,no_root_squash,no_subtree_check,fsid=102)"
nfs_client_imports:
- local: "/srv/galaxy"
remote: "/srv/galaxy"
server_host: "{{ galaxy_host }}"
- local: "/data/part0"
remote: "/data/part0"
server_host: "{{ galaxy_host }}"
# - local: "/cvmfs/data.galaxyproject.org"
# remote: "/cvmfs/data.galaxyproject.org"
# server_host: "{{ galaxy_host }}"
# - local: "/cvmfs/data.usegalaxy.no"
# remote: "/cvmfs/data.usegalaxy.no"
# server_host: "{{ galaxy_host }}"
cvmfs_role: client
galaxy_cvmfs_repos_enabled: config-repo
cvmfs_server_urls:
- domain: galaxyproject.org
use_geoapi: yes
#key_dir: "/cvmfs/cvmfs-config.galaxyproject.org/etc/cvmfs/keys/galaxyproject.org/"
urls:
#- "http://invivo.hpc.uio.no/cvmfs/@fqrn@"
- "http://cvmfs1-ufr0.galaxyproject.eu/cvmfs/@fqrn@"
- "http://cvmfs1-psu0.galaxyproject.org/cvmfs/@fqrn@"
- "http://cvmfs1-iu0.galaxyproject.org/cvmfs/@fqrn@"
- "http://cvmfs1-tacc0.galaxyproject.org/cvmfs/@fqrn@"
#- "http://cvmfs1-mel0.gvl.org.au/cvmfs/@fqrn@"
- domain: "{{ galaxy_host }}"
use_geoapi: no
urls:
- "http://{{ cvmfs0_host }}/cvmfs/@fqrn@"
telegraf_plugins_extra:
services:
plugin: "exec"
config:
- commands = ["/usr/local/bin/systemd_status.py -t fail2ban slurmd autofs"]
- timeout = "15s"
- data_format = "influx"
- interval = "1m"