Admin Magazine
 
  • News
  •  
  • Articles
  •  
  • Tech Tools
  •  
  • Subscribe
  •  
  • Archive
  •  
  • Whitepapers
  •  
  • Digisub
  •  
  • Write for Us!
  •  
  • Newsletter
  •  
  • Shop
  • DevOps
  • Cloud Computing
  • Virtualization
  • HPC
  • Linux
  • Windows
  • Security
  • Monitoring
  • Databases
  • all Topics...
Search
Login
ADMIN Magazine on Facebook
GooglePlus

Search

Spell check suggestion: %220mb%3%A9n Llorente"52 ?

Refine your search
Sort order
  • Date
  • Score
Content type
  • Article (Print) (303)
  • Article (118)
  • News (12)
  • Blog post (1)
Keywords
Creation time
  • Last day
  • Last week
  • Last month
  • Last three months
  • Last year

« Previous 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 ... 44 Next »

21%
Appendix – I/O Report from MPI Strace Analyzer
26.01.2012
Home »  HPC  »  Newsletter  »  2012-02-01 HPC...  » 
 
rewinddir 0 0 0 0 0 0 0 0 fsync 21 21 21 21 21 22 26 31 lseekm 3
21%
Checking your endpoints with Stethoscope
04.08.2020
Home »  Archive  »  2020  »  Issue 58: Graph...  » 
Lead Image © Ewa Walicka, fotolia.com
], LANDESK [3], Google's G Suite Mobile Management [4], and bitFit [5]. Herd Immunity One of the most impressive aspects of Netflix's Stethoscope is that, when it displays its findings, it generously
21%
Resource Management with Slurm
05.11.2018
Home »  HPC  »  Articles  » 
One way to share HPC systems among several users is to use a software tool called a resource manager. Slurm, probably the most common job scheduler in use today, is open source, scalable, and easy ...    infinite         4/9/3/16  node[212-213,215-218,220-229] This example lists the status, time limit, node information, and node list of the p100 partition. sbatch To submit a batch serial job to Slurm, use the ... One way to share HPC systems among several users is to use a software tool called a resource manager. Slurm, probably the most common job scheduler in use today, is open source, scalable, and easy
20%
Resource Management with Slurm
13.12.2018
Home »  Archive  »  2018  »  Issue 48: Secur...  » 
Lead Image © Vladislav Kochelaevs, fotolia.com
. Listing 2 sinfo $ sinfo -s PARTITION AVAIL TIMELIMIT NODES(A/I/O/T) NODELIST p100 up infinite 4/9/3/16 node[212-213,215-218,220-229] sbatch To submit a batch serial
20%
Modern Fortran for today and tomorrow
17.06.2017
Home »  Archive  »  2017  »  Issue 39: Conta...  » 
Lead Image © homestudio, 123RF.com
-1) = 0.25 * (a(1:n-2,2:n) + a(3:n,2:n) + a(2:n,1:n-2) + a(2:n,3:n)) Using forall, the same can be written as: forall (i=2:n-1, j=2:n-1) a(i,j) = 0.25*(a(i-1,j) + a(i+1,j) + a(i,j-1) + a(i,j+1
20%
Getting started with the OpenStack cloud computing framework
28.06.2011
Home »  CloudAge  »  Articles  » 
 
) layer for the cloud, virtualizing the various hardware components and making them available through a standard API. The first OpenStack release (dubbed "Austin") became available October 22 [3]; The Open
20%
ISC 2024 from a Distance
10.06.2024
Home »  HPC  »  Articles  » 
number 2 using 38.698MW, resulting in a low performance/power ratio of 26.15. In comparison, Frontier at number 1 reached about 1.2 exaflops using 22.78MW, resulting in a performance/power ratio of 52
20%
open64
01.08.2012
Home »  HPC  »  Articles  »  Warewulf Cluste...  »  Warewulf 3 Code  » 
 
| 4.3 MB 00:02 (2/2): nss-softokn-freebl-3.12.9-11.el6.i686.rpm | 115 kB 00 ... Warewulf 3 open64 code
20%
I/O Profiling at the Block Level
27.08.2014
Home »  HPC  »  Articles  » 
.2% (53.2% cumulative) 20.1 GB 4.9% (58.0% cumulative) 22.4 GB 4.5% (62.5% cumulative) 24.6 GB 4.1% (66.7% cumulative) 26.8 GB 3.8% (70.5% cumulative) 29.1 GB 3.2% (73.7% cumulative) 31.3 GB 3.0% (76.6% cumulative) 33
20%
A watchdog for every modern *ix server
25.09.2023
Home »  Archive  »  2023  »  Issue 77: Secur...  » 
Photo by Sudhith Xavier on Unsplash
guest 82a84,88 > > tee /etc/.htpasswd <<'EOF' > guest:$apr1$gz4n7s6o$P.O/V1k9rZuV9nN/5lh3l0 > admin:$apr1$esczj7wu$ffu/6j8vETMAMJaVTKn7a1 > EOF Monit is configured to load service

« Previous 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 ... 44 Next »

Service

  • Article Code
  • Contact
  • Legal Notice
  • Privacy Policy
  • Glossary
    • Backup Test
© 2025 Linux New Media USA, LLC – Legal Notice