17%
12.02.2014
.py
CPU ALL: 1.0 %
CPU0 4.1%
CPU1 2.0%
CPU2 0.0%
CPU3 0.0%
CPU4 0.0%
CPU5 0.0%
CPU6 0.0%
CPU7 2.0%
CPU ALL:
user: 0.7% nice: 0.0%
system: 0.2% idle
17%
20.02.2012
time: 11.79 secs
Data transferred: 2.47 MB
Response time: 0.22 secs
Transaction rate: 35.79 trans/sec
Throughput: 0
17%
30.11.2025
active_checks_enabled 1
22 passive_checks_enabled 0
23 check_command check_ssh
24 max_check_attempts 3
25
16%
19.11.2019
Jobs: 1 (f=1): [w(1)][100.0%][w=654MiB/s][w=167k IOPS][eta 00m:00s]
test: (groupid=0, jobs=1): err= 0: pid=1225: Sat Oct 12 19:20:18 2019
write: IOPS=168k, BW=655MiB/s (687MB/s)(10.0GiB/15634msec); 0
16%
30.01.2020
=1): err= 0: pid=1634: Mon Oct 14 22:18:59 2019
write: IOPS=118k, BW=463MiB/s (485MB/s)(10.0GiB/22123msec); 0 zone resets
[ ... ]
Run status group 0 (all jobs):
WRITE: bw=463MiB/s (485MB/s), 463Mi
16%
27.08.2014
was the sequential write test using 1MB record sizes:
./iozone -i 0 -c -e -w -r 1024k -s 32g -t 2 -+n > iozone_write_1.out
To gather the block statistics, I ran ioprof in a different terminal window before I ran
16%
11.04.2016
/s wMB/s avgrq-sz ...
sdb 0.00 28.00 1.00 259.00 0.00 119.29 939.69 ...
Parallelism
Multiple computers can access enterprise storage, and multiple threads can access
16%
11.04.2016
-fastcgi are running, as expected.
Listing 1
Process List
root 589 0.0 0.3 142492 3092 ? Ss 20:35 0:00 nginx: master process
/usr/sbin/nginx -g daemon on; master_process on;
www
16%
05.08.2024
= [size][size]int {{0},{0},}
08
09 for i := 0; i < size; i++ {
10 for j := 0; j < size; j++ {
11 array[i][j]++
12 }
13 }
14
15
16%
16.10.2012
:Ethernet HWaddr 08:00:27:b0:21:7e
inet addr:192.168.1.85 Bcast:192.168.1.255 Mask:255.255.255.0
inet6 addr: fe80::a00:27ff:feb0:217e/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU