Comprehensive data protection for all workloads
Post Reply
BobbyHood
Influencer
Posts: 18
Liked: 1 time
Joined: May 04, 2021 2:46 pm
Full Name: Bobby Hood
Contact:

Why is the backup size larger than the data size?

Post by BobbyHood »

Can anyone help me explain this? I keep finding backup jobs where the "backup size" is larger than the "data size" (see highlighted numbers at bottom).

Image

I found this thread indicating it could be a cluster size issue. It's hard to imagine because the size difference seems pretty extreme. But the repository is a Data Domain so I wouldn't be surprised at the weirdness.
vmware-vsphere-f24/backup-size-larger-t ... 83945.html
Mildur
Product Manager
Posts: 8735
Liked: 2296 times
Joined: May 13, 2017 4:51 pm
Full Name: Fabian K.
Location: Switzerland
Contact:

Re: Why is the backup size larger than the data size?

Post by Mildur »

Hello Bobby

Is this a forever incremental backup chain? Do you have scheduled regular full backups?
In case it's a forever incremental backup chain, the Full Backup File won't get smaller. It may have blocks stored inside that file which already doesn't exist anymore. Maybe you had once connected some huge disks to it?

You can schedule a backup file compact operation and see if it's help:
https://helpcenter.veeam.com/docs/backu ... ml?ver=120

If not, please open a support case with our customer support. Please let me know the case number.

Best,
Fabian
Product Management Analyst @ Veeam Software
BobbyHood
Influencer
Posts: 18
Liked: 1 time
Joined: May 04, 2021 2:46 pm
Full Name: Bobby Hood
Contact:

Re: Why is the backup size larger than the data size?

Post by BobbyHood »

Ok that was simple, thanks. They are forever forward and needed compacted. That leads to a question, what is best practice for a dedup appliance like a Data Domain? Should I do periodic synthetic fulls, active fulls, or schedule a compact operation? Because as soon as I turned on the compact operation and ran the jobs again we lost about half of the DD's reduction. Is that just temporary or should I do fulls instead?
MarkBoothmaa
Veeam Legend
Posts: 181
Liked: 49 times
Joined: Mar 22, 2017 11:10 am
Full Name: Mark Boothman
Location: Darlington, United Kingdom
Contact:

Re: Why is the backup size larger than the data size?

Post by MarkBoothmaa »

With a Data Domain (or any dedup appliance) you should be doing at least weekly synth fulls.
depending on the size of the VM's you could do an active full then on a monthly basis.
Mildur
Product Manager
Posts: 8735
Liked: 2296 times
Joined: May 13, 2017 4:51 pm
Full Name: Fabian K.
Location: Switzerland
Contact:

Re: Why is the backup size larger than the data size?

Post by Mildur »

Mark is correct.
Please schedule weekly synthetic full backups if you use the Data Domain over DDBoost.
Because as soon as I turned on the compact operation and ran the jobs again we lost about half of the DD's reduction. Is that just temporary or should I do fulls instead?
DD reduction is calculated by pre-compression and post-compression size of your backups.
You had a file with a lot of zeroed data. 17 Terabyte. 17 Terabyte of empty blocks can be stored as a single block on the Data Domain. I assume your high reduction comes from that.

Now with the empty blocks gone from the backup file, the compression rate is a lot smaller.

Best,
Fabian
Product Management Analyst @ Veeam Software
Post Reply

Who is online

Users browsing this forum: No registered users and 26 guests