-
- Enthusiast
- Posts: 37
- Liked: 2 times
- Joined: Oct 30, 2009 2:43 am
- Full Name: Felix Buenemann
- Contact:
Veeam Backup on ZFS (NexentaStor)
Hey,
I'm currently working on a project that will use NexentaStor / ZFS for primary VM and Backup storage.
The Backup storage will consist of 12x 2TB 7.2k SATA III drives in RAID10 (or ZFS load shared mirrors to be exact) and be attached to the ESXi 4.1 servers via 10GBe links. The appliance itself will be powered by 8x 2.4 GHz (+HT) and 54 GB RAM. The ESXi servers will each have 12x 2.66 GHz (+HT) and 96 GB RAM.
We already orderer Veeam Enterprise licenses, so we can leverage Veeam Backups full potential.
Now the interesting thing:
ZFS allows for Dedup and Compression inside the filesystem, for both block devices (called ZVOLS, used by iSCSI) or regular shares (NFS, SMB/CIFS).
Given that similar technologies are implemented in Veeam Backup itself, has anyone compared performance when using either of these features in Veeam vs. on the storage appliance?
Best Regards,
Felix Buenemann
I'm currently working on a project that will use NexentaStor / ZFS for primary VM and Backup storage.
The Backup storage will consist of 12x 2TB 7.2k SATA III drives in RAID10 (or ZFS load shared mirrors to be exact) and be attached to the ESXi 4.1 servers via 10GBe links. The appliance itself will be powered by 8x 2.4 GHz (+HT) and 54 GB RAM. The ESXi servers will each have 12x 2.66 GHz (+HT) and 96 GB RAM.
We already orderer Veeam Enterprise licenses, so we can leverage Veeam Backups full potential.
Now the interesting thing:
ZFS allows for Dedup and Compression inside the filesystem, for both block devices (called ZVOLS, used by iSCSI) or regular shares (NFS, SMB/CIFS).
Given that similar technologies are implemented in Veeam Backup itself, has anyone compared performance when using either of these features in Veeam vs. on the storage appliance?
Best Regards,
Felix Buenemann
-
- Chief Product Officer
- Posts: 31835
- Liked: 7325 times
- Joined: Jan 01, 2006 1:01 am
- Location: Baar, Switzerland
- Contact:
Re: Veeam Backup on ZFS (NexentaStor)
Hi Felix, I've seen a tweet about this recently from another user. He found ZFS cannot really dedupe/compress beyond what we already do. However, theoretically ZFS should be able to dedupe between different backup files (if you are using multiple jobs). In any case, I would not recommend disabling Veeam dedupe and compression, because it is done on the source side, and thus positively affects processing performance and backup window. Thanks!
-
- Enthusiast
- Posts: 37
- Liked: 2 times
- Joined: Oct 30, 2009 2:43 am
- Full Name: Felix Buenemann
- Contact:
Re: Veeam Backup on ZFS (NexentaStor)
Thanks for the tips.
We'll also be using ZFS replication to do block level replication of the primary backup storage to a remote site.
In this scenarios, will it be helpful to use true incrementals instead of Veeam Backups default reverse-inrementals? Which scenario would cause less data changes on disk, so less replication traffic?
We'll also be using ZFS replication to do block level replication of the primary backup storage to a remote site.
In this scenarios, will it be helpful to use true incrementals instead of Veeam Backups default reverse-inrementals? Which scenario would cause less data changes on disk, so less replication traffic?
-
- Chief Product Officer
- Posts: 31835
- Liked: 7325 times
- Joined: Jan 01, 2006 1:01 am
- Location: Baar, Switzerland
- Contact:
Re: Veeam Backup on ZFS (NexentaStor)
Default are regular (true) incrementals. For block level replication, least traffic can be achieved by using reversed incremental mode and 5.0.1 HF1 with special registry key that prevents renaming VBK file after each incremental pass. All of this can be obtained through support.
Regular incremental backup is second best option, it will produce less traffic on business days than above option (due to only having to shoot incremental file, versus reversed incremental file plus updates to VBK), but on the other hand it will produce more traffic during weekend because you will need to transfer new full backup file in its entirity. For some customers, this might be a better option (smaller traffic required on business days - which is when it really matters, at the price of more traffic on weekend - when it does not really matter).
Hope this helps.
Regular incremental backup is second best option, it will produce less traffic on business days than above option (due to only having to shoot incremental file, versus reversed incremental file plus updates to VBK), but on the other hand it will produce more traffic during weekend because you will need to transfer new full backup file in its entirity. For some customers, this might be a better option (smaller traffic required on business days - which is when it really matters, at the price of more traffic on weekend - when it does not really matter).
Hope this helps.
-
- Novice
- Posts: 5
- Liked: never
- Joined: Mar 13, 2011 6:45 pm
- Full Name: Steve Neruda
- Contact:
Re: Veeam Backup on ZFS (NexentaStor)
Make sure you have alot of memory on the ZFS server if you attempt deduplication. If the dedupe table dosen't stay in memory you will see severe performance degradation.
-
- Chief Product Officer
- Posts: 31835
- Liked: 7325 times
- Joined: Jan 01, 2006 1:01 am
- Location: Baar, Switzerland
- Contact:
Re: Veeam Backup on ZFS (NexentaStor)
Hi Steve, how much is "a lot"? What would you recommend? Thanks.
-
- Enthusiast
- Posts: 26
- Liked: never
- Joined: Sep 04, 2010 11:06 pm
- Full Name: Richard Yamauchi
- Contact:
Re: Veeam Backup on ZFS (NexentaStor)
Taken from this link:
http://hub.opensolaris.org/bin/view/Com ... +zfs/dedup
3. Make sure your system has enough memory to support dedup. Determine the memory requirements for deduplicating your data as follows:
A. Use the zdb -S ouput to determine the in-core dedup table requirements:
Each in-core dedup table entry is approximately 250 bytes
Multiply the number of allocated blocks times 250. For example:
in-core DDT size = 3.75M x 250 = 937.50M
B. Additional memory considerations from Roch's excellent blog:
20 TB of unique data stored in 128K records or more than 1TB of unique data in 8K records would require about 32 GB of physical memory. If you need to store more unique data than what these ratios provide, strongly consider allocating some large read optimized SSD to hold the deduplication table (DDT). The DDT lookups are small random I/Os that are well handled by current generation SSDs.
/////////////////////////////////////////////////
Personally I've done a little bit of testing with The Community edition of Nexenta. I have definitely not been running it in an optimized environment: it's run on Xen fully virtualized of a desktop with only 4GB or RAM. I'm using a Quantum Dxi4510 to backup to, but I'm trying to work the Nexenta as a Secondary and Archive system.
http://hub.opensolaris.org/bin/view/Com ... +zfs/dedup
3. Make sure your system has enough memory to support dedup. Determine the memory requirements for deduplicating your data as follows:
A. Use the zdb -S ouput to determine the in-core dedup table requirements:
Each in-core dedup table entry is approximately 250 bytes
Multiply the number of allocated blocks times 250. For example:
in-core DDT size = 3.75M x 250 = 937.50M
B. Additional memory considerations from Roch's excellent blog:
20 TB of unique data stored in 128K records or more than 1TB of unique data in 8K records would require about 32 GB of physical memory. If you need to store more unique data than what these ratios provide, strongly consider allocating some large read optimized SSD to hold the deduplication table (DDT). The DDT lookups are small random I/Os that are well handled by current generation SSDs.
/////////////////////////////////////////////////
Personally I've done a little bit of testing with The Community edition of Nexenta. I have definitely not been running it in an optimized environment: it's run on Xen fully virtualized of a desktop with only 4GB or RAM. I'm using a Quantum Dxi4510 to backup to, but I'm trying to work the Nexenta as a Secondary and Archive system.
-
- Enthusiast
- Posts: 94
- Liked: 3 times
- Joined: Jun 10, 2010 6:32 pm
- Contact:
Re: Veeam Backup on ZFS (NexentaStor)
Funny, I have been experimenting with an Openindiana box (OpenSolaris derivitive), which support ZFS de-dupe.
Transferring the same full VM backup up 5 times in a row, in which the data should roughly be the same, I'm seeing 1.01 dedupe ratio.
It looks like trying to dedupe a Veeam backup is a non-starter. Which is OK, since the file is already deduped.
Transferring the same full VM backup up 5 times in a row, in which the data should roughly be the same, I'm seeing 1.01 dedupe ratio.
It looks like trying to dedupe a Veeam backup is a non-starter. Which is OK, since the file is already deduped.
-
- Enthusiast
- Posts: 26
- Liked: never
- Joined: Sep 04, 2010 11:06 pm
- Full Name: Richard Yamauchi
- Contact:
Re: Veeam Backup on ZFS (NexentaStor)
What you can do is experiment by turning off the Dedupe in Veeam and then manually calculating the difference in size that you save. Sometimes the ZFS statistics are not always updated. Turn off all compression from Veeam as well because that will almost always prevent ZFS from being able to deduplicate the data.digitlman wrote:Funny, I have been experimenting with an Openindiana box (OpenSolaris derivitive), which support ZFS de-dupe.
Transferring the same full VM backup up 5 times in a row, in which the data should roughly be the same, I'm seeing 1.01 dedupe ratio.
It looks like trying to dedupe a Veeam backup is a non-starter. Which is OK, since the file is already deduped.
http://constantin.glez.de/blog/2010/03/ ... -need-know
-
- Enthusiast
- Posts: 94
- Liked: 3 times
- Joined: Jun 10, 2010 6:32 pm
- Contact:
Re: Veeam Backup on ZFS (NexentaStor)
Should the deduplication ratio be around the same, whether the dedupe is being done through Veeam or ZFS?
-
- Chief Product Officer
- Posts: 31835
- Liked: 7325 times
- Joined: Jan 01, 2006 1:01 am
- Location: Baar, Switzerland
- Contact:
Re: Veeam Backup on ZFS (NexentaStor)
No, not really.
1. Different block sizes (smaller block size = better deduplication).
2. Veeam also does good compression of deduped content (ZFS does not, or very basic - not sure).
We did a lot of testing on this matter, and found that dedupe with (relatively) large block size plus compression is a winning combination for an image-level backup application comparing to only compression, or only dedupe (with small block size). Results are vastly superior in terms of performance to compression ratio.
1. Different block sizes (smaller block size = better deduplication).
2. Veeam also does good compression of deduped content (ZFS does not, or very basic - not sure).
We did a lot of testing on this matter, and found that dedupe with (relatively) large block size plus compression is a winning combination for an image-level backup application comparing to only compression, or only dedupe (with small block size). Results are vastly superior in terms of performance to compression ratio.
Who is online
Users browsing this forum: Bing [Bot], Google [Bot] and 51 guests