48%
21.08.2012
=/vnfs
The VNFS grew a little bit in size from the ganglia additions to 72.3MB, which is still pretty small.
Now I can boot the compute node. Once it comes up (check this by ssh
ing to the node as a user
48%
28.03.2012
/log/collectl -r00:00,7 -m -F60 -s+YZCD --iosize
################################################################################
# Collectl: V3.6.1-4 HiRes: 1 Options: -p localhost-20120310-133840.raw.gz -P -f
48%
14.08.2017
:31 FS_scan.csv
$ gzip -9 FS_scan.csv
$ ls -lsah FS_scan.csv.gz
268K -rw-r--r-- 1 laytonjb laytonjb 261K 2014-06-09 20:31 FS_scan.csv.gz
The original file is 3.2MB, but after using gzip with the -9
48%
11.02.2016
DestinationSizeChange 41943106 (40.0 MB)
Another view of the file statistics lists which file effected the change:
# gunzip -c /mnt/backup/rdiff-backup-data/file_statistics.\
2015-03-15T10\:44\:06+01\:00.data.gz | awk '$2
48%
25.11.2012
crash /proc/iomem
03000000-0affffff : Crash kernel
[root@rhel ~]# grep crash /proc/cmdline
ro root=/dev/mapper/VolGroup00-LogVol00 rd_NO_LUKS KEYBOARDTYPE=pc \
KEYTABLE=de-latin1-nodeadkeys rd
48%
30.09.2013
the complex handling and high costs of other tools.
Based on openSUSE 11.4 and BusyBox, StressLinux is available as a 200MB, or 225MB ISO image for 64-bit and 32-bit architectures. Alternative versions
48%
23.03.2016
_scrub_rate 0 ue_count
0 csrow0 0 csrow3 0 csrow6 0 mc_name 0 seconds_since_reset 0 ue_noinfo_count
In the /sys
filesystem, for each csrow, a number of entries contain a lot of information
48%
14.08.2020
offline Completed without error 00% 12441 -
# 3 Short offline Completed without error 00% 12441 -
The output of the extended test is reported in the Extended offline
line. It, too, was completed
48%
26.01.2025
south of 200MB on your hard drive, which is surprisingly little in view of the broad feature set. One particular advantage of the tool is that the developers behind this project always look to support
48%
07.01.2014
a great deal of power in a few lines:
rm -rf backup.3
mv backup.2 backup.3
mv backup.1 backup.2
cp -al backup.0 backup.1
rsync -a --delete source_directory/ backup.0/
To better understand the script, I