21%
17.02.2015
Load
30 check_command check_nrpe!check_load!5.0!4.0!3.0!10.0!6.0!4.0
31 }
32 define service{
33 use generic-service ; Name of service
21%
20.05.2014
Viewing Server Topology
01 # numactl --hardware
available: 8 nodes (0-7)
node 0 cpus: 0 1 2 3 4 5 6 7 8 9
node 0 size: 16373 MB
node 0 free: 15837 MB
node 1 cpus: 10 11 12 13 14 15 16 17 18 19
node 1
21%
11.06.2014
_auth;
04
05 group server => "hercules", "sugar";
06
07 desc "Get the uptime of all servers";
08
09 task "uptime", group => "server", sub {
10 my $output = run "uptime";
11 say $output;
12 }
You might
21%
14.08.2017
api.raml
01 #%Raml 0.8
02 title: Contacts
03 version: 0.1
04 #baseUri: http://www.rve.com/contactshelf
05 baseUri: https://mocksvc.mulesoft.com/mocks/d4c4356f-0508-4277-9060-00ab6dd36e9b
21%
29.10.2013
ScalingGroupName/awseb-e-hxaxpp3bsa-stack-AWSEBAutoScalingGroup-1NMEE2RF9GMBP:policyName/awseb-e-hxaxpp3bsa-stack-AWSEBAutoScalingScaleUpPolicy-1Q75QYV2QAY6X
2013-05-08 20:25:18 INFO Created CloudWatch alarm named: awseb-e-hxaxpp3
21%
05.11.2018
nodes, and make sure to do this as a user and not as root.
3. To make life easier, use shared storage between the controller and the compute nodes.
4. Make sure the UIDs and GIDs are consistent
21%
13.12.2018
In previous articles, I examined some fundamental tools for HPC systems, including pdsh [1] (parallel shells), Lmod environment modules [2], and shared storage with NFS and SSHFS [3]. One remaining
21%
07.10.2014
.snap
s ntestvm1.img 5 8.0 GB 292 MB 2.4 GB 2014-03-01 11:42 982a3a 2 mar.snap
s ntestvm1.img 6 8.0 GB 128 MB 2.6 GB 2014-03-10 19:48 982a3b 2 mar2.snap
ntestvm1.img 0 8.0 GB
21%
30.01.2020
-06f4dce9c3ae2c504" # for eu-west-3 paris
07 InstanceType: t2.small
08 AvailabilityZone: !GetAtt FortiVPCFrontNet.AvailabilityZone
09 KeyName:
10 Ref: KeyName
11 Network
21%
10.04.2015
; Figure 3) [6], and Project Atomic, "a lean and mean OS built for containers" [7] looks promising, despite the beta state. If you want to dive deeper into the underlying technology, the OpenShift blog