17%
30.11.2025
backup_active = 1
05 backup = 192.168.1.232
06 heartbeat = 1
07 heartbeat_port = 539
08 keepalive = 6
09 deadtime = 18
10 network = direct
11 debug_level = NONE
12 monitor_links = 1
13 syncdaemon = 1
14
17%
30.11.2025
Password = "YdhKKoy2Huq1CVHwIR"
05 Pid Directory = "/var/run"
06 Query File = "/usr/lib/bacula/query.sql"
07 Working Directory = "/var/lib/bacula"
08 }
09
10 Fileset {
11 Name = "Full Set"
12
17%
28.03.2012
and sequential write test using 16MB record sizes to a file that is 16GB in size (twice the physical memory).
Once the testing was complete, I grabbed the raw collectl data file and copied it into a directory
17%
04.11.2011
width;
09 uint height;
10 float *data;
11 } convolutionKernel;
12
13
14 /**
15 * Clamp value to limits of uchar.
16 */
17 uchar clampuchar(int value)
18 {
19 return (uchar) std
17%
16.05.2013
a9a6615fb5c045693 |
root@alice:~# quantum port-list
+--------------------------------------+------+-------------------+
| 0c478fa6-c12c-... | | fa:16:3e:29
17%
30.11.2025
.33
164.69
0.75
rand_mat_stat
3.37
39.34
11.64
54.54
22.07
8.12
rand_mat_mul
1.00
1.18
0.70
1.65
8
17%
13.12.2018
# and network has converged
traceroute -In 10.4.1.25
1 10.3.1.22 # backup VPN
# router site 3
2 172.16.1.7 # backup VPN
# router site 4
3 10.4.1.25
LAN Failure
17%
30.01.2020
Id: 4e90b424-95d9-4453-a2f4-8f5259f5f263 Duration: 70.72 ms Billed Duration: 100 ms Memory Size: 128 MB Max Memory Used: 55 MB Init Duration: 129.20 ms
More or Less
17%
11.06.2014
application server.
On a system in a stable state, throughput initially is not affected by file operations, but after a certain value (e.g., 16,384MB), performance collapses. As Figure 1 shows
17%
30.04.2013
of the memory:
dd if=/dev/fmem of=memory.dd bs=1MB count=512
Another tool for dumping memory is the kernel module crash
, which was developed by Red Hat. Just like fmem
, crash
creates a pseudodevice called