19%
25.09.2013
)
8 (4)
7.2
0.9
Nehalem-EP (2009)
8 (4)
32
4
Westmere-EP (2010)
12 (6)
42
3.5
Westmere-EP (2010)
8 (4)
42
19%
05.11.2018
# for your environment.
05 #
06 #
07 # slurm.conf file generated by configurator.html.
08 #
09 # See the slurm.conf man page for more information.
10 #
11 ClusterName=compute-cluster
12 Control
19%
13.12.2018
.conf file generated by configurator.html.
08 #
09 # See the slurm.conf man page for more information.
10 #
11 ClusterName=compute-cluster
12 ControlMachine=slurm-ctrl
13 #
14 SlurmUser=slurm
15 Slurmctld
19%
21.01.2020
/proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md0 : active raid5 sdd1[5] sde1[4] sdc1[2] sdb1[1] nvme0n1p1[0](J)
20508171264 blocks super 1.2 level 5, 512k
19%
25.03.2020
] [raid6] [raid5] [raid4] [raid10]
md0 : active raid5 sdd1[5] sde1[4] sdc1[2] sdb1[1] nvme0n1p1[0](J)
20508171264 blocks super 1.2 level 5, 512k chunk, algorithm 2 [4/3] [UU
19%
30.01.2020
VPC:
04 Type: AWS::EC2::VPC
05 Properties:
06 CidrBlock:
07 Ref: VPCNet
08 Tags:
09 - Key: Name
10 Value:
11 Ref: VPCName
12
13 Forti
18%
30.01.2020
: (groupid=0, jobs=1): err= 0: pid=7055: Sat Oct 12 19:09:53 2019
write: IOPS=34.8k, BW=136MiB/s (143MB/s)(9.97GiB/75084msec); 0 zone resets
[ ... ]
Run status group 0 (all jobs):
WRITE: bw=136MiB/s (143MB
18%
30.11.2020
== 0:
08 data = { 'key1' : [10,10.1,10+11j],
09 'key2' : ('mpi4py' , 'python'),
10 'key3' : array([1, 2, 3]) }
11 else:
12 data = None
13 # end if
14
15 data = comm
18%
07.10.2014
.snap
s ntestvm1.img 5 8.0 GB 292 MB 2.4 GB 2014-03-01 11:42 982a3a 2 mar.snap
s ntestvm1.img 6 8.0 GB 128 MB 2.6 GB 2014-03-10 19:48 982a3b 2 mar2.snap
ntestvm1.img 0 8.0 GB
18%
17.10.2011
.24: 1.5-23
05 pve-kernel-2.6.24-11-pve: 2.6.24-23
06 qemu-server: 1.1-16
07 pve-firmware: 1.0-5
08 libpve-storage-perl: 1.0-13
09 vncterm: 0.9-2
10 vzctl: 3.0.23-1pve11
11 vzdump: 1.2-5
12 vzprocps: 2