63%
23.04.2014
as a module):
[laytonjb@home4 HPC_028]$ ls -lsa /lib/modules/2.6.32-431.5.1.el6.x86_64/kernel/fs/fuse
total 168
4 drwxr-xr-x 2 root root 4096 Mar 20 20:09 ./
4 drwxr-xr-x 30 root root 4096 Mar 20 20
63%
18.03.2012
:\Users\khess> $Host
Name : ConsoleHost
Version : 2.0
InstanceId : ccbc6963-c33e-4848-a8e5-aa1ccaf7d65c
UI : System
63%
07.01.2014
[laytonjb@home4 TEST]$ rsync -a -delete /home/laytonjb/TEST/SOURCE/ backup.0/
[laytonjb@home4 TEST]$ ls -s
total 20
4 backup.0/ 4 backup.1/ 4 backup.2/ 4 backup.3/ 4 SOURCE/
[laytonjb@home4 TEST]$ du -sh
28M
63%
30.01.2020
Fn::Base64:
20 Fn::Join:
21 - ''
22 -
23 - "{\n"
24 - '"bucket"'
25 - ' : "'
26 - Ref: S3Bucketname
27
63%
10.04.2015
.6.1.2.1.2.2.1.19
# ifOutErrors
snmpwalk -v1 -Oqs -c $ROCOMMUNITY $HOST .1.3.6.1.2.1.2.2.1.20
}
# Directory for logfiles
LOGDIR="/home/falko/monitorlog"
# community string
ROCOMMUNITY="community"
# date
YEAR=$( date
63%
04.11.2011
.
Listing 1: Running Many Commands at Once
01 $ cat commands
02 date
03 hostname
04 echo foo | md5sum
05
06 $ cat commands | parallel
07 Thu Jul 8 10:37:20 EST 2010
08 desktop‑machine
09 d3b07384d113edec49
63%
14.11.2013
to the computer (Figure 1). You can double-click a VHD(X) file to this, too. After you have created the virtual hard disk, Windows displays it in Disk Management, and you can manage it like any other disk
62%
30.11.2025
X4 920 CPU (quad core)
8GB of memory (DDR2-800)
The operating system and boot drive are on an IBM DTLA-307020 (20GB drive at Ultra ATA/100)
/home is on a Seagate ST1360827AS
A 64GB Intel X
62%
22.05.2023
.HistogramOpts{
Name: "prometheus_http_request_duration_seconds",
Help: "Histogram of latencies for HTTP requests.",
- Buckets: []float64{.1, .2, .4, 1, 3, 8, 20, 60, 120},
+ NativeHistogramBucketFactor: 1
62%
20.05.2014
.2.0 include the introduction of YARN (Yet Another Resource Negotiator) as an optional replacement for MapReduce. MapReduce in Hadoop 1.x (Figure 3) is not optimal for all workloads. It achieves its best ... Hadoop 2.x and its associated tools promise to deliver big data solutions not just to the IT-heavy big players, but to anyone with unstructured data and the need for multidimensional data analysis. ... 20