Software →Lime Tech unRAID Server OS 6.6.6
Published by: mitsumi on 27-12-2018, 05:29 | 0
Lime Tech unRAID Server OS 6.6.6 | 260 Mb
Where many competitors have focused on creating an appliance offering, unRAID has taken a hardware-agnostic approach to network-attached storage. The benefit is unRAID's ability to boot on nearly any x86 64-bit capable system and manage an array of disks that vary in size, speed, brand, and protocol. In addition, by eliminating the use of traditional RAID-based technologies, we can scale on-demand by adding more disks and without needing to rebalance data.
unRAID's storage capabilities are broken down into three components: the array, the cache, and the user share file system.
unRAID Array
The primary concept behind an unRAID array is its ability to manage an aggregate of disk devices (JBOD) that is protected by a dedicated parity device. A parity device provides a way for you to reconstruct data from a failed disk onto a new one. While it seems mind boggling that one drive can possibly back up other drives that have way more storage capacity than the parity, it is able to reconstruct the missing data from a failed drive using binary logic called XOR (eXclusive-OR). Since hard drives store data as zeroes and ones, when a drive fails the parity compares the binary data on all the surviving drives and can deduce the missing data to rebuild.
unRAID Cache
The cache drive feature of unRAID provides faster data capture. Generally speaking, by using a cache alongside an array of 3 or more devices, you can achieve up to 3x write performance. When data is written to a user share that has been configured to use the cache device, all of that data is initially written directly to the dedicated cache device. Because this device is not a part of the array, the write speed is unimpeded by parity calculations. Subsequently an unRAID process called "the mover" copies the data from the cache to the array at time and frequency of your choosing. Once the data has been successfully copied, the space on the cache drive is once again freed up to front-end other write operations to cache-enabled user shares.
Cache Pool
With a single cache device, data captured there is at risk, as a parity device doesn't protect it. However, you can configure a cache with multiple devices both to increase your cache capacity as well as to add protection for that data. The grouping of multiple devices in a cache is referred to as building a cache pool. The unRAID cache pool is created through a unique twist on traditional RAID 1. Here are just some of the benefits:
Improved Data Protection
With a single cache device, there's a possibility that you can lose your data if the device fails before the data gets moved to the array. With a cache pool, however, all write operations are replicated across two separate disks to ensure that the loss of any one drive in the pool does not result in data loss.
Increased System Uptime
If a cache pool device fails, the system will continue to operate as normal. No need to drop everything to deal with a system outage. You can simply change the device when convenient.
Better Scalability
Add more devices of different sizes to your pool as you need to and grow on-demand.
Optimized for SSDs
unRAID now has native support for TRIM, which can substantially reduce the number of write operations when used as a cache device. Benefits of SSDs vs. HDDs:
- They don't require time to 'spin up' or consume a lot of power to operate (they are fast and efficient);
- They are also smaller, so you can fit more of them into a smaller space for highly compact, crazy fast storage.
- When used for storing large quantities of smaller files (e.g., metadata), SSDs can provide a faster response time for these files to the application compared to spinning hard disks; and
- SSDs are most ideal for supporting virtual machines. VM performance benefits on an SSD are comparable to what a user would experience with them on a desktop PC vs. a spinning disk.
Optimized for Virtualization
Virtual machines and applications can have their data reside on the cache pool permanently for overall improved performance, while keeping mass-storage content on the array still accessible to those virtual instances using VirtFS (for KVM Virtual Machines) and Docker (for Containers). Given the desire for "fast-as-you" responsiveness in application and machine performance, using the cache pool for virtual machine/application storage is a no-brainer. Use of SSDs in a cache pool extends this benefit even further.
unRAID Shares
Unlike most RAID systems, unRAID saves data to individual drives. To simplify manageability, users can create shares that allow files written to them to be spread across multiple drives. Each share can be thought of as a top-level folder on a drive. When browsing through a share, all data from all drives that participate in that share will be displayed together. Users do not need to know which disk a file is on in order to access it under a share. Shares can be tuned to include/exclude specific disks and to utilize various methods for determining how files are allocated across those disks.
In addition to controlling how data is distributed across drives, users can also control what network protocols the share is visible through as well as define user-level security policy. When accessing your unRAID server over a network protocol, all shares exported through that protocol will be visible, but you can toggle protocols for both individual shares as well as at a global setting level. Should you have private data on your system that you wish to protect from anonymous access, user accounts can be created and policies defined to limit access to only trusted individuals. In addition to this concept of user shares, entire individual disks can be shared individually as well for advanced management capabilities. The simple concept with shares is that they enable you to organize your data any way you want.
Hardware Requirements:
Contains detailed information about the hardware requirements for each of the three primary functions of unRAID Server OS 6.
USB Flash Device
unRAID Server runs entirely from RAM, minimizing writes to the Flash device; however, it is also used to store vital system configuration data and must remain installed in your server during operation. In addition, to obtain a registration key for unRAID Server OS, your flash device must contain a unique GUID (Globally Unique Identifier) and needs to be at least 1GB in size, but no larger than 32GB. A few of the brands that are known for this include:
- Lexar Firefly or JumpDrive
- SanDisk Cruzer Fit
- Sony Micro Vault Tiny
There are a number of USB Flash devices that do not have a valid serial number and can not be registered:
- Lacie MosKeyto
- Delock USB2.0 Nano Memory stick
- Maxell ONYX
- There may be others.
Store and Protect Data
In order to install and begin taking advantage of the basic NAS capabilities unRAID Server OS 6 has to offer, you will need to meet the following minimum hardware requirements:
- A 64-bit capable processor (1.0 GHz or better)
- At least 1GB of RAM
- Linux hardware driver support (for storage, Ethernet, and USB controllers)
- At least one hard disk drive (two if you wish to protect your data with a parity disk)
Homepage -
https://lime-technology.com
Download link:
uploadgig_com:
https://uploadgig.com/file/download/107021b17723A378/ig2gz.Lime.Tech.unRAID.Server.OS.6.6.6.rar
rapidgator_net:
https://rapidgator.net/file/789a4389df0ea7f249df9a1f446363bd/ig2gz.Lime.Tech.unRAID.Server.OS.6.6.6.rar.html
nitroflare_com:
http://nitroflare.com/view/DB565D793699C68/ig2gz.Lime.Tech.unRAID.Server.OS.6.6.6.rar
https://uploadgig.com/file/download/107021b17723A378/ig2gz.Lime.Tech.unRAID.Server.OS.6.6.6.rar
rapidgator_net:
https://rapidgator.net/file/789a4389df0ea7f249df9a1f446363bd/ig2gz.Lime.Tech.unRAID.Server.OS.6.6.6.rar.html
nitroflare_com:
http://nitroflare.com/view/DB565D793699C68/ig2gz.Lime.Tech.unRAID.Server.OS.6.6.6.rar
Links are Interchangeable - No Password - Single Extraction
Related News
-
{related-news}
Comments (0)
Information
Users of Guests are not allowed to comment this publication.