30%
07.10.2014
(Listing 3).
Listing 3
Defining Replication Factors
01 # dog vdi list
02 Name Id Size Used Shared Creation time VDI id Copies Tag
03 one.img 0 4.0 MB 0.0 MB 0.0
30%
19.11.2019
Jobs: 1 (f=1): [w(1)][100.0%][w=654MiB/s][w=167k IOPS][eta 00m:00s]
test: (groupid=0, jobs=1): err= 0: pid=1225: Sat Oct 12 19:20:18 2019
write: IOPS=168k, BW=655MiB/s (687MB/s)(10.0GiB/15634msec); 0
30%
30.01.2020
]
test: (groupid=0, jobs=1): err= 0: pid=1225: Sat Oct 12 19:20:18 2019
write: IOPS=168k, BW=655MiB/s (687MB/s)(10.0GiB/15634msec); 0 zone resets
[ ... ]
Run status group 0 (all jobs):
WRITE: bw=655Mi
29%
28.03.2012
# Kernel: 2.6.32-220.4.1.el6.x86_64 Memory: 7540044 Swap:
# NumDisks: 2 DiskNames: sdb sda
# NumNets: 2 NetNames: lo: eth0:100
# NumSlabs: 201 Version: 2.1
# SCSI: DA:1:00:00:00 DA:2:00:00:00 CD:4
29%
30.11.2025
: AuthenticAMD Speed(MHz): 2812.629 Cores: 4 Siblings: 4 Nodes: 1
13 # Kernel: 2.6.32-220.4.1.el6.x86_64 Memory: 7540044 Swap:
14 # NumDisks: 2 DiskNames: sdb sda
15 # NumNets: 2 NetNames: lo: eth0:100
16 # Num
29%
20.02.2012
time: 11.79 secs
Data transferred: 2.47 MB
Response time: 0.22 secs
Transaction rate: 35.79 trans/sec
Throughput: 0
29%
27.08.2014
was the sequential write test using 1MB record sizes:
./iozone -i 0 -c -e -w -r 1024k -s 32g -t 2 -+n > iozone_write_1.out
To gather the block statistics, I ran ioprof in a different terminal window before I ran
29%
30.11.2025
to 100% of its capacity on a single metric. This simple, easily understood measurement is remarkably portable across machines and nearly foolproof. If used correctly, single-metric benchmarks are a quick
29%
30.11.2025
62
787
10MB < < 100MB
0
0
0
0
0
0
0
2
Table 4 from the MPI strace Analyzer report; all-zero rows for sizes >100MB
29%
11.04.2016
/s wMB/s avgrq-sz ...
sdb 0.00 28.00 1.00 259.00 0.00 119.29 939.69 ...
Parallelism
Multiple computers can access enterprise storage, and multiple threads can access