45%
17.09.2013
reset_counters 0 size_mb
0 ce_noinfo_count 0 csrow2 0 csrow5 0 device 0 sdram_scrub_rate 0 ue_count
0 csrow0 0 csrow3 0 csrow6 0 mc_name 0 seconds_since_reset 0 ue_noinfo_count
Notice that this system
45%
18.07.2013
buffered disk reads: 616 MB in
3.00 seconds = 205.03 MB/sec
$ hdparm -T /dev/sda
/dev/sda:
Timing cached reads: 6292 MB in
2.00 seconds = 3153.09 MB/sec
If this were a spinning disk, you would also
45%
26.02.2014
reqs merged: 3.78/s Write reqs completed: 2.10/s
Read BW: 0.00 MB/s Write BW: 0.02 MB/s
Avg sector size issued 23.78 Avg
45%
30.11.2025
SUSE 11.3, the benchmark repository offers an up-to-date package [1]. Fio is also included with Fedora. The source code from the tarball or Git repository can be compiled with a simple make, assuming you
44%
16.03.2021
B/s), 1236KiB/s-1236KiB/s (1266kB/s-1266kB/s), io=72.5MiB (76.1MB), run=60102-60102msec
Disk stats (read/write):
md0: ios=53/18535, merge=0/0, ticks=0/0, in_queue=0, util=0.00%, aggrios=33/18732, aggrmerge
44%
13.12.2018
(USB 2), and a Samsung Fit Plus 32 (USB 3).
Disk Caches
The OS is not the only player in the caching business. Examining a Samsung 750 SATA SSD drive, you cannot but notice a 256MB RAM buffer [8
44%
11.02.2016
.40 <- < 71% idle >
0 1.00 0.00 0.37 0.00 0.00 0.06 0.00 0.00 0.00 98.57
1 100.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
2 3.77 0.00
43%
05.11.2018
it the number of cores, number of cores per socket, threads per core, and the amount of memory available (e.g., 30,000MB, or 30GB, here).
CgroupAutomount=yes
CgroupReleaseAgentDir="/etc/slurm/cgroup"
Constrain
43%
13.12.2018
socket, threads per core, and the amount of memory available (e.g., 30,000MB, or 30GB, here).
CgroupAutomount=yes
CgroupReleaseAgentDir="/etc/slurm/cgroup"
ConstrainCores=yes
Constrain
43%
21.08.2012
=/vnfs
The VNFS grew a little bit in size from the ganglia additions to 72.3MB, which is still pretty small.
Now I can boot the compute node. Once it comes up (check this by ssh
ing to the node as a user