20%
10.09.2012
from /etc/password
and /etc/groups
. The time is computed in number of seconds since 1970-01-01 00:00:00 UTC. The script
entry records the name of the wrapper script that was used, because you might
20%
05.11.2018
# for your environment.
05 #
06 #
07 # slurm.conf file generated by configurator.html.
08 #
09 # See the slurm.conf man page for more information.
10 #
11 ClusterName=compute-cluster
12 Control
20%
13.12.2018
.conf file generated by configurator.html.
08 #
09 # See the slurm.conf man page for more information.
10 #
11 ClusterName=compute-cluster
12 ControlMachine=slurm-ctrl
13 #
14 SlurmUser=slurm
15 Slurmctld
20%
18.07.2013
rq2 ra92
08 set rq3 cdrom
09
10 attach rq0 d0.dsk
11 attach rq1 d1.dsk
12 attach rq2 d2.dsk
13
14 attach -r rq3 cdrom.iso
15
16 set rl disable
17 set ts disable
18
19 set xq mac=08-00-2B-AA-BB-CC
20
20%
16.10.2012
6), and start stream blocking (line 7), which executes the command and waits for the response. Now, write the output to a variable (lines 9-12), close the stream (line 14), and send the response
19%
31.10.2025
System, DUP: total=8.00MB, used=4.00KB
Metadata, DUP: total=1.00GB, used=79.20MB
If Btrfs only has one device available when the filesystem is created, it automatically generates a duplicate
19%
14.08.2020
and SMART values. The disks were a combination of consumer-grade drives (SATA and PATA) with speeds from 5,400 to 7,200rpm and capacities ranging from 80 to 400GB. Several drive manufacturers were represented
19%
04.12.2024
can use Kustomize or Helm to install the AWX operator with default settings, as described in the documentation [2]. Operator version 2.12.2 was used for this example.
The main playbook for the AWX 00
19%
29.09.2020
between failure and SMART values. The disks were a combination of consumer-grade drives (SATA and PATA) with speeds from 5,400 to 7,200rpm and capacities ranging from 80 to 400GB. Several drive
19%
21.08.2012
just two nodes: test1, which is the master node, and n0001, which is the first compute node):
[laytonjb@test1 ~]$ pdsh -w test1,n0001 uptime
test1: 18:57:17 up 2:40, 5 users, load average: 0.00, 0.00