85%
18.04.2022
You might assume Razer would have released their new Tensorbook as a Windows-specific device designed for machine-learning tasks that require powerful hardware to get the job done. That new
85%
17.02.2015
Combined Search
# ausearch -k hosts-file -ts today -ul 1000
----
time->Tue Jul 1 16:36:00 2014
type=PATH msg=audit(1404225360.791:124138): item=0 name="/etc/hosts"
inode=2755966 dev=fd:02 mode=0100644
85%
23.03.2016
of csrows can be found by examining the /sys
entries for a memory controller. For example:
$ ls -s /sys/devices/system/edac/mc/mc0
The number of elements, labeled csrowX
(where X
is an integer
84%
19.11.2019
, ioengine=libaio, iodepth=32
fio-3.1
Starting 1 process
Jobs: 1 (f=1): [w(1)][100.0%][r=0KiB/s,w=1401KiB/s][r=0,w=350 IOPS][eta 00m:00s]
test: (groupid=0, jobs=1): err= 0: pid=3104: Sat Oct 12 14:39:08 2019
write: IOPS=352
84%
30.01.2020
: 1 (f=1): [w(1)][100.0%][r=0KiB/s,w=1401KiB/s][r=0,w=350 IOPS][eta 00m:00s]
test: (groupid=0, jobs=1): err= 0: pid=3104: Sat Oct 12 14:39:08 2019
write: IOPS=352, BW=1410KiB/s (1444kB/s)(82.8Mi
84%
05.11.2018
One way to share HPC systems among several users is to use a software tool called a resource manager. Slurm, probably the most common job scheduler in use today, is open source, scalable, and easy ...
29 SlurmctldTimeout=300
30 SlurmdTimeout=300
31 InactiveLimit=0
32 MinJobAge=300
33 KillWait=30
34 Waittime=0
35 #
36 # SCHEDULING
37 SchedulerType=sched/backfill
38 SelectType=select/cons_res
39 ...
One way to share HPC systems among several users is to use a software tool called a resource manager. Slurm, probably the most common job scheduler in use today, is open source, scalable, and easy ... Slurm Job Scheduling System
84%
13.12.2018
Timeout=300
31 InactiveLimit=0
32 MinJobAge=300
33 KillWait=30
34 Waittime=0
35 #
36 # SCHEDULING
37 SchedulerType=sched/backfill
38 SelectType=select/cons_res
39 SelectType ... One way to share HPC systems among several users is to use a software tool called a resource manager. Slurm, probably the most common job scheduler in use today, is open source, scalable, and easy ... Slurm Job Scheduling System
83%
21.11.2012
) ! Delta (x and y directions)
101
102 ALLOCATE ( unew(1:m,1:m), u(1:m,1:m) ) ! mem for unew, u
103
104 !
105 ! Boundary Conditions
106 ! ===================
107 !
108 pi = DACOS(0.0
83%
12.03.2013
Previously we talked about using iostat to monitor local storage on your server or compute nodes, but what if you use NFS in your compute nodes to run jobs? The nfsiostat tool can help you ... MB_nor/s rMB_dir/s wMB_dir/s rMB_svr/s wMB_svr/s ops/s rops/s wops/s
192.168.1.250:/home
1230649.19 1843536.81 0.00 0.00 1229407.77 1843781 ...
Previously we talked about using iostat to monitor local storage on your server or compute nodes, but what if you use NFS in your compute nodes to run jobs? The nfsiostat tool can help you
82%
05.12.2019
TIME CMD
Root 1 0 0 19:05 ? 00:00:00 sleep 1000
Listing 2
Process on the Host
$ ps -ef|grep sleep
Cherf 30328 29757 0 20:44 ? 00:00:00 sleep 1000
Cherf 30396 3353