The local storage in our R740XD2 is 14x4TB 7.2K in RAID10.
Using DiskSpd.exe we can get close to 700MB/s.
Code: Select all
Command Line: C:\IT\DiskSpd-2.0.21a\x86\diskspd.exe -c1G -b512K -w0 -r4K -Sh -d600 E:\test.vbk
Input parameters:
timespan: 1
-------------
duration: 600s
warm up time: 5s
cool down time: 0s
random seed: 0
path: 'E:\test.vbk'
think time: 0ms
burst size: 0
software cache disabled
hardware write cache disabled, writethrough on
performing read test
block size: 524288
using random I/O (alignment: 4096)
number of outstanding I/O operations: 2
thread stride size: 0
threads per file: 1
using I/O Completion Ports
IO priority: normal
System information:
computer name: VAN-BACKUP01
start time: 2020/08/24 17:15:13 UTC
Results for timespan 1:
*******************************************************************************
actual test time: 600.00s
thread count: 1
proc count: 32
Total IO
thread | bytes | I/Os | MiB/s | I/O per s | file
------------------------------------------------------------------------------
0 | 437503655936 | 834472 | 695.39 | 1390.78 | E:\test.vbk (1024MiB)
------------------------------------------------------------------------------
total: 437503655936 | 834472 | 695.39 | 1390.78
Read IO
thread | bytes | I/Os | MiB/s | I/O per s | file
------------------------------------------------------------------------------
0 | 437503655936 | 834472 | 695.39 | 1390.78 | E:\test.vbk (1024MiB)
------------------------------------------------------------------------------
total: 437503655936 | 834472 | 695.39 | 1390.78
Write IO
thread | bytes | I/Os | MiB/s | I/O per s | file
------------------------------------------------------------------------------
0 | 0 | 0 | 0.00 | 0.00 | E:\test.vbk (1024MiB)
------------------------------------------------------------------------------
total: 0 | 0 | 0.00 | 0.00
Code: Select all
IBM Tape Diagnostic Tool Standard Edition - Full Write
Host Bus ID LUN Model Serial Fware Changer
+----+----+----+----+--------------+------------+------+------------+
| 5 | 0 | 0 | 0 | ULTRIUM-TD7 | F3A300A000 | J4D0 | |
+----+----+----+----+--------------+------------+------+------------+
Compres- Transfer Data Size Elapsed Data Rate
sible Size (KB) (MB) Time (s) (MB/s) Remaining (MB)
+------+---------+---------+--------+----------+ +---------------+
| No | 256 | 346181 | 1207.4 | 286.717 | | 5948068 |
+------+---------+---------+--------+----------+ +---------------+
Status:
+----------------------+
| FULL WRITE |
+----------------------+
Progress:
+------------------------------------------------------------------+
|### |
+------------------------------------------------------------------+
Based on tape-f29/slow-tape-job-performance-t61054.html post with similar issue there were a few hints dropped that are similar in our environment, backup files being dehydrated while sitting in ReFS
It goes on to quote Veeam TS coming back with...
Our B2D2T is reverse incremental using CBT, SAN-MODE and FAST-CLONE on REFS; we can't change out of reverse incremental because every tape job has to be a full backup.The situation is unfortunately quite expected.
As you know, ReFS (when using Block Cloning) allows several files to share the same physical data. Instead of high-cost copy of real block, it just copies metadata and sets up references to physical regions. However to read the such file, it is required to read the address with the link and only then read the data from are where data is stored. More links are used, more effort is required to read the file.
So degradation of read performance seems to be an expected cost of fast merge operations. Backup to tape includes reading data from source, and, as your BTT logs say, source is always the most time-consuming operation. This is unfortunately well-known ReFS limitation which hardly could be overcome.
How can I confirm 100% that ReFS dehydration is the root cause? Is there anything else that we can do to reduce the ReFS metadata that is potentially slowing down the B2T?
I guess one could delete all backups and start over and it would be fast for the first X amount of runs.. Switch back to NTFS... Swap spinning disks for SSDs...
Is there anything we can do with Veeam software to reduce the metadata overhead? Can we re-hydrate the reverse incremental backups so that the B2T does not take such a read I/O hit?
After banging my head against this for 2 days I came to this ReFS dehydration conclusion and looking for some help/options that don't involve going back to NTFS and loosing out on Fast-Clone (30min B2D jobs are nice).
Support Case #04351586