21%
07.06.2019
this script, please let me know, preferably in Python
06 # hanscees@AT@hanscees.com
07 # modified from https://gist.github.com/cirocosta/17ea17be7ac11594cb0f290b0a3ac0d1x
08
09 set -o errexit
10
11 main() {
12
21%
01.08.2019
whoami
ls -al /usr/share/nginx
root
total 12
drwxr-xr-x 3 root root 4096 Feb 6 08:11 .
drwxr-xr-x 1 root root 4096 Feb 18 13
21%
29.09.2020
-line operations.
To install Dockly [3], you can choose one of two routes: with npm (see the "Installation by npm" box for that route) and in a Docker container. For context, on my laptop, about 43MB of file space
20%
25.03.2021
, ioengine=libaio, iodepth=32
fio-3.12
Starting 1 process
Jobs: 1 (f=1)
test: (groupid=0, jobs=1): err= 0: pid=5956: Sat Jan 9 16:38:53 2021
read: IOPS=256k, BW=998MiB/s (1047MB/s)(2045MiB/2049msec
20%
11.04.2016
(512 MB) copied, 49.1424 s, 10.4 MB/s
If you want to empty the read and write cache for benchmark purposes, you can do so using:
sync; echo 3 > /proc/sys/vm/drop_caches
Sequential access
20%
09.01.2013
[12]. The Linux world also has tools for treating Amazon S3 as an rsync target. The best-known tool is probably boto_rsync [13], a Python tool that uses rsync to sync from a local directory to an S3 ... advantages: S3QL. ... S3QL filesystem for cloud backups
20%
25.03.2020
7 1 56008 loop1
06 7 2 56184 loop2
07 7 3 91264 loop3
08 259 0 244198584 nvme0n1
09 8 0 488386584 sda
10 8 1 1024 sda1
11
20%
21.01.2020
7 1 56008 loop1
06 7 2 56184 loop2
07 7 3 91264 loop3
08 259 0 244198584 nvme0n1
09 8 0 488386584 sda
10 8 1 1024 sda1
11
20%
07.01.2013
OS
01 name: centos
02 summary: CentOS installation with BoxGrinder
03 os:
04 name: centos
05 version: 6
06 hardware:
07 partitions:
08 "/":
09 size: 4
10 "/home":
11 size: 1
12
20%
09.10.2017
"SubnetMax": "10.99.0.0",
07 "Backend": {
08 "Type": "udp",
09 "Port": 7890
10 }
11 }
12 [...]
At first glance, this concept looks robust and simple