31%
07.04.2022
of 1KiB (in case really small payload sizes have some exceptional performance); 4KiB, 32KiB, or 64KiB; and maybe even 128KiB, 256KiB, or 1MiB. The reason I like to see a range of payload sizes
31%
26.03.2025
into a fixed character string by a hash function. This character string, also known as a hash, represents the original data but cannot be easily converted back. One common example of hashing is the SHA-256
31%
09.01.2013
a 256-bit AES key. An additional SHA-256 HMAC checksum protects the data from manipulation.
Compression: S3QL compresses the data before storing, using either LZMA, bzip2, or gzip. This compression
31%
22.05.2023
the sequential read performance with the blockdev command. For example, to set a read-ahead value of 2048 (1MB) for the /dev/sdb1 device, use:
blockdev --setra 2048 /dev/sdb1
For kernel parameters, you can
31%
05.11.2013
the number of cores. On average, a card from the 5100 series only has 35MB of RAM for each thread, compared with several hundred megabytes for each thread on current server systems. Because of the limited
31%
03.09.2013
should not use legacy 10Mb technology, which is often encountered in the form of small switches that are still running, even if everything else has been modernized. Each hardware bottleneck leads to poor
31%
01.04.2014
bootstrapping of its own local cloud, which does not require access to a public cloud provider and thus imposes very few requirements.
The Cloudify tarball is a heavyweight at 160MB, but the user initially only
31%
21.01.2013
earlier.
As a client for the Admin Console, Red Hat recommends a Windows 7 system. This just leaves the mini-hypervisor, RHEV-H, for the virtualization host. This can be downloaded as a 172MB evaluation
31%
11.06.2013
available devices, including any LVM groups or RAID arrays.
In the example, Anaconda automatically creates the LVM group for the Fedora partition during the partitioning, and the same goes for the 500MB boot
31%
25.09.2013
the code states should be good enough for caches up to 20MB. The Stream FAQ recommends you use a problem size such that each array is four times the sum of the caches (L1, L2, and L3). You can either change