subject: Benefits Of Data Deduplication [print this page] There are a number of different ways that you can achieve the results of data Deduplication:
You will be able to improve the performance of across your network with the savings of your disk purchases.
You can use the application savings feature which will allow you to back up your email and other data as well as manage other applications on your computer.
You might not know how data duplication works; if this is the case then you can continue to read on to learn more about the process of data deduplication.
Basically how data deduplication works is it breaks down data objects into parts to determine the duplicated data. For each part or chunk that is calculated the identifier is calculated. This occurs with the use of a hash function such as SHA, and MD5. The index of identifiers is compared for each chunk that is read. The data of the chunks have to be managed properly in the system. The good news is that the majority of the data systems have an underlying system where the files of the data are stored. Each one of the chunks is associated with a metadata associate which includes information about how often the information is accessed. Some of the things that are tracked include the file name and the disk blocks.
The ratios of deduplication come in a large variety of settings and handlings. It all depends upon the date stream that is being used and the data volumes. The Metadata size is something that would need to be taken into consideration. The reasons for this is because the size of the metadata file will have a huge impact on the deduplication ratio ranges. The ranges can go anywhere from 4:1-200:1+.
Something else that should be taken into consideration is the performance. This needs to be taken into consideration for the purpose of calculating the hash. It should also be well sorted out in the case of updating the metadata and for storing and location certain data. So all in all the entire capability of the scale of the system should be taken into consideration when planning out this process so that you will know what to expect with the outcome. You need to know how long it will take for the system to reconstruct the data.
When systems are already backed up you can expect the next full back up to take up about 3-5 times the amount of compression that it did the first time. If the files are incremental then they would probably take compressions of about 5-8 times. For subsequent full back ups the compression ratios are between 50-60 times higher. Aggregated compressions can be expected to be 20+ times higher. Traditional storage of disks and tapes are compared to these gains.
There are different points within the system that data deduplications can be completed. They can be completed at the client, server, or the block storage array.
Whether you have a need to improve the investment return on your systems or address some of the environmental issues; data deduplication can play a significant role in helping you to get that change.