Data domain total compression factor
Apr 21, 2021DDOS uses "compression ratio" to measure the effectiveness of its data compression. Generally, it is the ratio of the total user data sizeĀ
What is DDoS compression ratio?
DDOS uses " compression ratio " to measure the effectiveness of its data compression.
Generally, it is the ratio of the total user data size to the total size of compressed data or the used physical space size.
Data Domain file system is a "log-structured" dedupe file system.
,
What is the global compression ratio & dedupe ratio?
The global compression ratio depends on the percentage of new data within the incremental backup.
The dedupe ratio of a full backup (the non-initial ones), can also be low in a number of scenarios.
,
What is total compression?
Total compression is the total amount of compression that the Data Domain system performed with the data it received.
The first backup, which had entirely unique data, had the lowest total compression factor.
The second and third backups, which had similar amounts of unique data, had total compression factors of 7.7 and 9.2 respectively.
Compression: System Overall Effect
1. Fastcopy. When a fastcopy is done from a file in the active namespac…
2 Compression: Inline Statistics
1. The length of each write, referred to as raw_bytes;
2 Mysteries Brought by Dedupe
1. Do not be surprised when the initial backups only achieve small system …
2 What is DDoS compression ratio?
DDOS uses " compression ratio " to measure the effectiveness of its data compression
Generally, it is the ratio of the total user data size to the total size of compressed data or the used physical space size
Data Domain file system is a "log-structured" dedupe file system
What is the global compression ratio & dedupe ratio?
The global compression ratio depends on the percentage of new data within the incremental backup
The dedupe ratio of a full backup (the non-initial ones), can also be low in a number of scenarios
The figure below shows the local compression factor savings based on the default algorithm (maximized throughput) on the Data Domain system. A relationship exists between the amount of unique data and the local compression factor: The greater the amount of unique data, the more opportunity for compression, and the higher the compression factor.Looking at the total space written to the Data Domain System divided by the total space used will give an absolute compression ratio. This is accurate at the time it is run, but it does not differentiate the Global and Local Compression rates.