77%
30.01.2020
Id: 4e90b424-95d9-4453-a2f4-8f5259f5f263 Duration: 70.72 ms Billed Duration: 100 ms Memory Size: 128 MB Max Memory Used: 55 MB Init Duration: 129.20 ms
More or Less
77%
11.02.2016
Use of mysqld
# ps aux | egrep 'mysqld|VSZ'
USER PID %CPU %MEM VSZ RSS START TIME COMMAND
mysql 1568 0.0 0.0 4440 744 Sep08 0:00 mysqld_safe
mysql 2337 0.1 5.9 29163040
77%
11.04.2016
RProxyExtensions = css, gif, html, jpg, js, png, txt
CacheSize = 25 #mb
CacheMaxFilesize = 128 #kb
##load tinkering
#ListenBacklog = 128 #default 16
#MaxServerLoad = 0.7 #drop on high load (eeks)
#Socket
77%
10.04.2015
Qualys SSL-Test A+
Qualys SSL-Test A
Syss GmbH
Audits [2]
TLS 1.2
Yes
Yes
Yes
Yes
HTTPS/HSTS
Yes
Yes
Yes
Yes
77%
14.03.2013
command:
# sockstat | grep "\*:[0-9]"
The description for rc.conf(5) lists the services that need to be bound to a fixed IP address via parameters. One example is the inetd service:
inetd_flags="-wW -a
77%
05.11.2018
One way to share HPC systems among several users is to use a software tool called a resource manager. Slurm, probably the most common job scheduler in use today, is open source, scalable, and easy ... # for your environment.
05 #
06 #
07 # slurm.conf file generated by configurator.html.
08 #
09 # See the slurm.conf man page for more information.
10 #
11 ClusterName=compute-cluster
12 Control ...
One way to share HPC systems among several users is to use a software tool called a resource manager. Slurm, probably the most common job scheduler in use today, is open source, scalable, and easy
76%
13.12.2018
.conf file generated by configurator.html.
08 #
09 # See the slurm.conf man page for more information.
10 #
11 ClusterName=compute-cluster
12 ControlMachine=slurm-ctrl
13 #
14 SlurmUser=slurm
15 Slurmctld
76%
30.11.2025
the Buffer
01 [global]
02 ioengine=libaio
03 direct=1
04 filename=testfile
05 size=2g
06 bs=4m
07
08 refill_buffers=1
09
10 [write]
11 rw=write
12 write_bw_log
13
14 [read]
15 stonewall
16 rw=read
17 write
76%
14.11.2013
. This translates to Google's experiencing about 25,000-75,000 correctable errors (CE) per billion device hours per megabit, which translates to 2,000-6,000 CE/GB-yr (or about 250-750 CE/Gb-yr). This is much higher
76%
30.11.2025
_world[]="Hello World\n";
08 static dev_t hello_dev_number;
09 static struct cdev *driver_object;
10 static struct class *hello_class;
11 static struct device *hello_dev;
12
13 static ssize_t driver_read( struct file