15%
16.10.2012
to the screen (STDOUT; line 15).
Listing 1: SSH Script
01 #!/usr/bin/php
02
03
04 $ssh = ssh2_connect('192.168.1.85', 22);
05 ssh2_auth_password($ssh, 'khess', 'password');
06 $stream = ssh2_exec
15%
11.04.2016
hiawatha running
www-data 4766 0.1 0.3 118232 4016 ? Ssl 20:13 0:00 /usr/sbin/hiawatha
You can use netstat to check the bindings:
netstat -tulpn
See the output in Listing 2.
Listing 2
15%
21.01.2020
local server machine (Listing 1). In this example, the four drives sdb
to sde
in lines 12, 13, 15, and 16 will be used to create the NVMe target. Each drive is 7TB, which you can verify
15%
05.11.2018
Machine=slurm-ctrl
13 #
14 SlurmUser=slurm
15 SlurmctldPort=6817
16 SlurmdPort=6818
17 AuthType=auth/munge
18 StateSaveLocation=/var/spool/slurm/ctld
19 SlurmdSpoolDir=/var/spool/slurm/d
20 SwitchType=switch/none
21 Mpi
15%
13.12.2018
.conf file generated by configurator.html.
08 #
09 # See the slurm.conf man page for more information.
10 #
11 ClusterName=compute-cluster
12 ControlMachine=slurm-ctrl
13 #
14 SlurmUser=slurm
15 Slurmctld
15%
30.11.2025
/nsswitch.conf (Konfigurieren der Namensauflösung)
08 ...
09 hosts files dns
10 ipnodes files dns
11 ...
12 # vi /etc/hosts
13 ::1 localhost
14 127.0.0.1 localhost
15 192.168.1.200 myiscsiserver myiscsiserver.local loghost
16
15%
29.09.2020
sitting at less than 50MB (and using less than half the RAM of a standard cluster) the binary that runs K3s is a sight to behold and well worth getting your hands on. Especially when it's deemed production
15%
31.10.2025
System, DUP: total=8.00MB, used=4.00KB
Metadata, DUP: total=1.00GB, used=79.20MB
If Btrfs only has one device available when the filesystem is created, it automatically generates a duplicate
15%
25.03.2020
/server/node-token
K10fc63f5c9923fc0b5b377cac1432ca2a4daa0b8ebb2ed1df6c2b63df13b092002::server:bf7e806276f76d4bc00fdbf1b27ab921
API Token
Make sure you note the API token correctly or you might see the error
15%
09.01.2013
[12]. The Linux world also has tools for treating Amazon S3 as an rsync target. The best-known tool is probably boto_rsync [13], a Python tool that uses rsync to sync from a local directory to an S3