15%
11.04.2016
(512 MB) copied, 49.1424 s, 10.4 MB/s
If you want to empty the read and write cache for benchmark purposes, you can do so using:
sync; echo 3 > /proc/sys/vm/drop_caches
Sequential access
15%
14.09.2021
$(find /sys/devices/system/cpu -regex ".*cpu[0-9]+/topology/thread_siblings_list") | sort -n | uniq
0,32
1,33
2,34
3,35
4,36
5,37
6,38
7,39
8,40
9,41
10,42
11,43
12,44
13,45
14,46
15,47
16,48
17,49
18,50
19,51
20,52
21,53
22,54
23,55
24,56
25
15%
30.01.2020
without profiling).
Listing 3
pprofile Output
Command line: md_002.py
Total duration: 1662.48s
File: md_002.py
File duration: 1661.74s (99.96%)
Line #| Hits| Time| Time per hit
15%
05.11.2018
nodes, and make sure to do this as a user and not as root.
3. To make life easier, use shared storage between the controller and the compute nodes.
4. Make sure the UIDs and GIDs are consistent
14%
09.12.2019
in execution time by about a factor of 10 (i.e., it ran 10 times slower than without profiling).
Listing 3: pprofile
Output
Command line: md_002.py
Total duration: 1662.48s
File: md_002.py
File duration: 1661
14%
11.02.2016
/mnt/fuse/
2015-03-15T09:15:19 2015-03-15T09:50:34 2015-03-19T20:15:46
2015-03-21T09:45:05 2015-03-15T09:28:48 2015-03-15T10:44:06
2015-03-21T08:43:49
Make sure the TZ variable is set; otherwise
14%
25.03.2020
1
16 8 80 39078144 sdf
17 8 48 6836191232 sdd
18 8 49 6836189184 sdd1
19 8 32 6836191232 sdc
20 8 33 6836189184 sdc1
21 11 0 1048575 sr0
14%
21.01.2020
7 1 56008 loop1
06 7 2 56184 loop2
07 7 3 91264 loop3
08 259 0 244198584 nvme0n1
09 8 0 488386584 sda
10 8 1 1024 sda1
11
14%
14.03.2018
the solution presented in this article twice.
Within the account are Virtual Private Clouds (VPCs). Five VPCs per region and account are possible. In a VPC, the user can create up to 200 separate subnets. (An
14%
14.08.2017
:31 FS_scan.csv
$ gzip -9 FS_scan.csv
$ ls -lsah FS_scan.csv.gz
268K -rw-r--r-- 1 laytonjb laytonjb 261K 2014-06-09 20:31 FS_scan.csv.gz
The original file is 3.2MB, but after using gzip with the -9