14%
30.11.2025
.network.type = veth
04 lxc.network.flags = up
05 lxc.network.link = br0
06 lxc.network.hwaddr = 08:00:12:34:56:78
07 #lxc.network.ipv4 = 0.0.0.0
08 lxc.network.ipv4 = 192.168.1.69
09 lxc.network.name = eth0
10 lxc
14%
30.01.2020
: 1 (f=1): [w(1)][100.0%][r=0KiB/s,w=1401KiB/s][r=0,w=350 IOPS][eta 00m:00s]
test: (groupid=0, jobs=1): err= 0: pid=3104: Sat Oct 12 14:39:08 2019
write: IOPS=352, BW=1410KiB/s (1444kB/s)(82.8Mi
14%
20.10.2013
database [for details use: -P show]
ATA Version is: 8
ATA Standard is: ACS-2 (unknown minor revision code: 0x001f)
Local Time is: Sun Oct 13 09:12:15 2013 EDT
SMART support is: Available - device has SMART
14%
30.11.2025
Model";
11 $work = "ebiz-tx";
12 $node1 = "WebServer";
13 $node2 = "AppServer";
14 $node3 = "DBMServer";
15 $think = 0.0 * 1e-3; # as per test system
16 $dtime = 2.2 * 1e-3; # dummy service time
17
18
14%
10.09.2012
to logfiles, and it’s pretty simple to use:
[laytonjb@test1 ~]$ logger "This is a test"
...
[root@test1 ~]# tail -n 2 /var/log/messages
Aug 22 15:54:47 test1 avahi-daemon[1398]: Invalid query packet.
Aug 22 17:00
14%
19.11.2019
, ioengine=libaio, iodepth=32
fio-3.12
Starting 1 process
Jobs: 1 (f=1): [w(1)][100.0%][w=475MiB/s][w=122k IOPS][eta 00m:00s]
test: (groupid=0, jobs=1): err= 0: pid=1634: Mon Oct 14 22:18:59 2019
write: IOPS=118k, BW=463Mi
14%
26.11.2013
(text) {
11 polycrypt.importKey("raw", hex2bin(encRawKey)).oncomplete = function(e) {
12 var key = e.target.result;
13 polycrypt.encrypt(encAlg, key, str2bin(text)).oncomplete = function(e) {
14 var
13%
16.10.2012
($ssh, '/sbin/ifconfig');
07 stream_set_blocking($stream, true);
08
09 $response = '';
10 while($buffer = fread($stream, 4096)) {
11 $response .= $buffer;
12 }
13
14 fclose($stream);
15 echo $response;
16
17
13%
30.11.2025
("------------------------------------------- ----------------------\n")
09 }
10
11 probe tcp.receive {
12 printf(" %15s %15s %5d %5d %d %d %d %d %d
13 %d\n", saddr, daddr, sport, dport, urg,
14 ack, psh, rst, syn, fin)
15
13%
09.01.2013
[12]. The Linux world also has tools for treating Amazon S3 as an rsync target. The best-known tool is probably boto_rsync [13], a Python tool that uses rsync to sync from a local directory to an S3