You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/Using_eRI/Running_Jobs/SLURM_Partitions.md
+18-40Lines changed: 18 additions & 40 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -9,18 +9,6 @@ zendesk_article_id: 360000204076
9
9
zendesk_section_id: 360000030876
10
10
---
11
11
12
-
## General Limits
13
-
14
-
- No individual job can request more than 20,000 CPU hours. This has
15
-
the consequence that a job can request more CPUs if it is shorter
16
-
(short-and-wide vs long-and-skinny).
17
-
- No user can have more than 1,000 jobs in the queue at a time.
18
-
19
-
These limits are defaults and can be altered on a per-account basis if
20
-
there is a good reason. For example we will increase the limit on queued
21
-
jobs for those who need to submit large numbers of jobs, provided that
22
-
they undertake to do so with job arrays.
23
-
24
12
## Partitions
25
13
26
14
A partition can be specified via the appropriate [sbatch option](../../Getting_Started/Cheat_Sheets/Slurm-Reference_Sheet.md),
@@ -40,7 +28,7 @@ partition then you may receive a warning, please do not ignore this.
40
28
E.g.:
41
29
42
30
```out
43
-
sbatch: `hugemem` is not the most appropriate partition for this job, which would otherwise default to `large`. If you believe this is incorrect then contact support and quote the Job ID number.
31
+
sbatch: `hugemem` is not the most appropriate partition for this job, which would otherwise default to `compute`. If you believe this is incorrect then contact support and quote the Job ID number.
44
32
```
45
33
46
34
<table><tbody>
@@ -50,7 +38,6 @@ sbatch: `hugemem` is not the most appropriate partition for this job, which woul
50
38
<th>Nodes</th>
51
39
<th>CPUs/Node</th>
52
40
<th>Available Mem/CPU</th>
53
-
<th>Available Mem/Node</th>
54
41
<th>Max CPUs/job</th>
55
42
<th>Description</th>
56
43
</tr>
@@ -59,51 +46,44 @@ sbatch: `hugemem` is not the most appropriate partition for this job, which woul
59
46
<td>14 days</td>
60
47
<td>6</td>
61
48
<td>256</td>
62
-
<td>? MB</td>
49
+
<td>3.7 GB</td>
63
50
<td>950 GB</td>
64
-
<td>?</td>
65
51
<td>Default partition.</td>
66
52
</tr>
67
53
<tr>
68
54
<td>gpu</td>
69
55
<td>14 days</td>
70
56
<td>1</td>
71
57
<td>96</td>
72
-
<td>? MB</td>
58
+
<td>4.8 GB</td>
73
59
<td>470 GB</td>
74
-
<td>?</td>
75
-
<td></td>
60
+
<td>A100.</td>
76
61
</tr>
77
62
<tr>
78
63
<td>hugemem</td>
79
64
<td>14 days</td>
80
65
<td>2</td>
81
66
<td>256</td>
82
-
<td>-</td>
67
+
<td>14.9 GB</td>
83
68
<td>3800 GB</td>
84
-
<td>-</td>
85
69
<td>Very large amounts of memory.</td>
86
70
</tr>
87
71
<tr>
88
72
<td>interactive</td>
89
73
<td>60 days</td>
90
74
<td>3<br/></td>
91
75
<td>8</td>
92
-
93
-
<td>-</td>
94
-
95
-
<td>14 GB</td>
96
-
<td>?</td>
97
-
<td></td>
76
+
<td>1.8 GB</td>
77
+
<td>14.8 GB</td>
78
+
<td>Partition for interactive jobs.</td>
98
79
</tr>
99
80
<tr>
100
81
<td>vgpu</td>
101
82
<td>60 days</td>
102
83
<td>4</td>
103
84
<td>32</td>
104
-
<td>-</td>
85
+
<td>13 GB</td>
105
86
<td>418 GB</td>
106
-
<td>-</td>
107
87
<td>Virtual GPUs.</td>
108
88
</tr>
109
89
</tbody>
@@ -118,30 +98,28 @@ its project. There are other QoSs which you can select with the
118
98
119
99
### Interactive
120
100
121
-
Specifying `--qos=interactive` will give the job very high priority, but
122
-
is subject to some limits: up to 4 jobs, 16 hours duration, 4 CPUs, 128
123
-
GB, and 1 GPU.
101
+
Specifying `--qos=interactive` will give a very high priority interactive job.
The default GPU type is P100, of which you can request 1 or 2 per node. The vgpu partition contains four virtualised compute nodes, each with a single NVIDIA A10 GPGPU, PCIe 24GB cards.
111
+
The default GPU type is A100. The vgpu partition contains four virtualised compute nodes, each with a single NVIDIA A10 GPGPU, PCIe 24GB cards.
112
+
113
+
To request for the A100 GPU:
135
114
136
115
```sl
137
-
#SBATCH --gpus-per-node=1 # or equivalently, P100:1
116
+
#SBATCH --partition gpu
117
+
#SBATCH --gpus-per-node 1 # GPU resources required per node
138
118
```
139
119
140
-
To request A100 GPUs, use instead:
120
+
To request for vGPUs, use instead:
141
121
142
122
```sl
143
-
#SBATCH --gpus-per-node=A100:1
123
+
#SBATCH --partition vgpu
124
+
#SBATCH --gpus-per-node 1
144
125
```
145
-
146
-
See [GPU use on NeSI](../../Scientific_Computing/Running_Jobs_on_Maui_and_Mahuika/GPU_use_on_NeSI.md)
0 commit comments