I don't think you can copy backup repo to another volume and still keep block cloning.alesovodvojce wrote:After week of tests, and after countless ReFS horror days personally lived here, we have copied our Refs 4k repo to different filesystems to see the size differences. Here it is.
In actual numbers
ReFS 4k: 21 TB (source repo)
ReFS 64k 31 TB at least - we had to stop the file copy as the underlying disks runs out of free space
NTFS: 31 TB at least - same reason to stop. Finally we have shrinked source repo to 13 TB by deleting it files. After that, the target NTFS partition copied that size to 24 TB. So 13 TB Refs 4k made 24 TB NTFS).
Generalized
Refs 4k - best space saver. But lot of troubles (as in this thread)
Refs 64k - not a win in space saving, whilst still lot of refs benefits. but, the troubles will theoretically start as well, they are just postponed for later (when the repo size grows over unsaid limit)
NTFS - not win in space saving, no special benefits. Main benefit is stable filesystem = backups secured
We migrated first repo to NTFS now, enjoying stable backups. Second repo remais in Refs 4k for now for experiments
-
- Influencer
- Posts: 21
- Liked: 4 times
- Joined: Apr 14, 2017 5:25 pm
- Full Name: xudaiqing
- Contact:
Re: REFS 4k horror story
-
- Service Provider
- Posts: 13
- Liked: 2 times
- Joined: Apr 05, 2017 10:48 am
- Full Name: Richard Kraal
- Contact:
Re: REFS 4k horror story
already did that, MS did not want to support Veeam.Gostev wrote: No, we really want everyone experiencing the issue to open a ticket with Microsoft instead, to help raise the priority of this issue on their side.
So we tested with multiple Windows Backup jobs and had no problems, the destination FS was reachable all the time.
At this point the MS support stopped and is pointing at a Veeam software issue.
MS Case [REG:117040515557419]
-
- Influencer
- Posts: 21
- Liked: 4 times
- Joined: Apr 14, 2017 5:25 pm
- Full Name: xudaiqing
- Contact:
Re: REFS 4k horror story
The way veeam use refs seem to create large number of metadata and windows try to load that into memory which caused the problem.
In my test it seems the biggest spike appears when i delete some thing from the volume, my throught is windows loads the entire metadata into memory in order to figure out which block is not in use.
64K will difinitely help here as they reduces the metadata needed.
Base on that i think synthetic full will make things worse as they will need own sets of metadata.
I current have a reverse increasement repo of 2T used and the metadata spike on delete is still below 4G.
In my test it seems the biggest spike appears when i delete some thing from the volume, my throught is windows loads the entire metadata into memory in order to figure out which block is not in use.
64K will difinitely help here as they reduces the metadata needed.
Base on that i think synthetic full will make things worse as they will need own sets of metadata.
I current have a reverse increasement repo of 2T used and the metadata spike on delete is still below 4G.
-
- Enthusiast
- Posts: 26
- Liked: 12 times
- Joined: Jan 30, 2017 7:42 pm
- Full Name: Sam
- Contact:
Re: REFS 4k horror story
I've been having this issue since we got going back in January. And I'm able to pretty reliably encounter the problem. I have one particular disk that just doesn't want to cooperate.Gostev wrote:Yes, we've been in touch with the ReFS development team on this issue for a while now. Right now, they are working with one of our customers who has the issue reproducing most consistently. Internally, we do not have a lab that replicates the issue (and based on our support statistics, it does not seems to be very common in general ).
Let me know if I can be of any help!
-
- Chief Product Officer
- Posts: 31812
- Liked: 7302 times
- Joined: Jan 01, 2006 1:01 am
- Location: Baar, Switzerland
- Contact:
Re: REFS 4k horror story
If you open a support case with Microsoft, please PM me the case ID.
-
- Chief Product Officer
- Posts: 31812
- Liked: 7302 times
- Joined: Jan 01, 2006 1:01 am
- Location: Baar, Switzerland
- Contact:
Re: REFS 4k horror story
Correct, this test by @alesovodvojce was totally invalid, as simply copying files inflates them. ReFS space savings can be enormous, in the same league with deduplicating storage appliances. Also, the real difference in used space consumption between 4KB and 64KB ReFS volumes is only 10%.xudaiqing wrote:I don't think you can copy backup repo to another volume and still keep block cloning.
-
- Chief Product Officer
- Posts: 31812
- Liked: 7302 times
- Joined: Jan 01, 2006 1:01 am
- Location: Baar, Switzerland
- Contact:
Re: REFS 4k horror story
Nice, so it's like comparing Windows Calculator to Microsoft Excel. Windows backup does not integrate with ReFS at all (does not use block cloning or file integrity streams), so what's the point of using one for troubleshooting the issue clearly associated with said technologies? This makes zero sense.richardkraal wrote:So we tested with multiple Windows Backup jobs and had no problems, the destination FS was reachable all the time.
At this point the MS support stopped and is pointing at a Veeam software issue.
I suggest you ask the case to be escalated to the support management, otherwise this issue will never get to R&D in sufficient volume. Needless to say, we will be more than happy to help them reproduce the issue without Veeam in the picture if so desired - all we do is basic public API calls against regular files on the ReFS volume. That is, IF they are actually interested in finding and fixing the issue with their next-gen file system, so that all of their customers could start leveraging one (tell this to whoever will review your escalation).
-
- Enthusiast
- Posts: 26
- Liked: 12 times
- Joined: Jan 30, 2017 7:42 pm
- Full Name: Sam
- Contact:
Re: REFS 4k horror story
I'll talk to my management and see if they're willing to pay up for a support ticket with MS. More likely I'm just going to go through a process to rebuild/migrate everything.Gostev wrote:If you open a support case with Microsoft, please PM me the case ID.
-
- Chief Product Officer
- Posts: 31812
- Liked: 7302 times
- Joined: Jan 01, 2006 1:01 am
- Location: Baar, Switzerland
- Contact:
Re: REFS 4k horror story
Oh, I did not realize you'd have to pay for one. I will keep you in mind then. For example, Windows dedupe team wanted the affected users to work with them so they provided a way to open the support case at no charge for everyone. They only needed the case open so that all their processes work correctly though, most importantly they were actually interested in understanding and solving the issue - and needed customers with reliable repro.
-
- Enthusiast
- Posts: 63
- Liked: 9 times
- Joined: Nov 29, 2016 10:09 pm
- Contact:
Re: REFS 4k horror story
agree - the test design was bad. Thanks for pointing that.Gostev wrote: Correct, this test by @alesovodvojce was totally invalid, as simply copying files inflates them. ReFS space savings can be enormous, in the same league with deduplicating storage appliances. Also, the real difference in used space consumption between 4KB and 64KB ReFS volumes is only 10%.
Refs is promising filesystem. I have green light from boss to pay for MS ticket and raise the count. Will do that tomorrow
-
- Expert
- Posts: 127
- Liked: 29 times
- Joined: Oct 10, 2014 2:06 pm
- Contact:
Re: REFS 4k horror story
Note that when raising a ticket with Prosupport, you'll have to pay in advance since march 2016 if I remember correctly, could be 2015 as well. If it's identified as a bug, at the end of the ticket you'll get a refund. So have a company creditcard at hand when raising it. It used to be the other way around, you'd get an incoive when a ticket was closed which was not identified as a bug. Guess who gets the interest on all that cash now (although that might not be the primary reason )
-
- Enthusiast
- Posts: 26
- Liked: 12 times
- Joined: Jan 30, 2017 7:42 pm
- Full Name: Sam
- Contact:
Re: REFS 4k horror story
Sounds good. After poking through the thread I tried turning off synthetic fulls for all my jobs and what do you know, I got through a night without it locking up!Gostev wrote:Oh, I did not realize you'd have to pay for one. I will keep you in mind then. For example, Windows dedupe team wanted the affected users to work with them so they provided a way to open the support case at no charge for everyone. They only needed the case open so that all their processes work correctly though, most importantly they were actually interested in understanding and solving the issue - and needed customers with reliable repro.
I'll run this for a while and see if locks up again. If not, I'll probably monitor this thread and see if there's ever a permanent fix. We'd definitely like to be able to take synthetic fulls.
-
- Chief Product Officer
- Posts: 31812
- Liked: 7302 times
- Joined: Jan 01, 2006 1:01 am
- Location: Baar, Switzerland
- Contact:
Re: REFS 4k horror story
Just checked and the ReFS dev team is on it.
Re: REFS 4k horror story
Any news?Gostev wrote:Just checked and the ReFS dev team is on it.
Any casenumber at MS I can refer to in my ticket?
//Anders
-
- Enthusiast
- Posts: 59
- Liked: 20 times
- Joined: Dec 14, 2016 1:56 pm
- Contact:
Re: REFS 4k horror story
I got three memory dumps submitted to MS the other day. Gostev has my case ID, but if there's some reference to give MS I could submit that to the tech dealing with my case as well, to be sure there's no internal communication breakdown.
-
- Novice
- Posts: 3
- Liked: 4 times
- Joined: Jun 20, 2016 8:30 am
- Full Name: Jesper Sorensen
- Contact:
Re: REFS 4k horror story
Had to kill 2 different ReFS 64k repositories this week as both servers were completed locked due to the ReFS issue. I will be very hesitant to switch back to ReFS again, even if MS comes out with another "fix".
-
- Chief Product Officer
- Posts: 31812
- Liked: 7302 times
- Joined: Jan 01, 2006 1:01 am
- Location: Baar, Switzerland
- Contact:
Re: REFS 4k horror story
One theory I have that would explain why some users have issues and other don't is the difference in the amount of concurrent tasks. So one troubleshooting step for those experiencing lockups would be to reduce the amount of concurrent tasks on the repository by half and see if that makes any difference to stability. Perhaps even change it to 1 task if you have a lab environment where this issue reproduces. Thanks!
-
- Veeam Legend
- Posts: 1203
- Liked: 417 times
- Joined: Dec 17, 2015 7:17 am
- Contact:
Re: REFS 4k horror story
@gostev: We use no concurrent task limit at all right now and all runs well. But our REFS storage now has 96 Disks and not 24 as before. But even with a limit of 4 tasks REFS locked up with the 24 disk storage...
So in our case the much faster storage seems to have fixed the issue for now...
So in our case the much faster storage seems to have fixed the issue for now...
-
- Enthusiast
- Posts: 59
- Liked: 20 times
- Joined: Dec 14, 2016 1:56 pm
- Contact:
Re: REFS 4k horror story
I thought so at first too. One of the last series of lockups I got, though, was actually just trying to physically delete some Veeam backup images on disk, with Veeam completely disabled. The background "cleanup" process that slowly frees up available disk space after deletion spiked memory steadily up to 100% and nuked the server...you would get a few minutes to scramble around on each system boot before the ReFS kernel driver wrecked the machine. So, while concurrent tasks might be a factor in some cases, even just deleting some files (albeit, veeam block-cloned files) from disk without Veeam initiating any tasks at all can cause the issue. This is the series of memory dumps that Microsoft now has their hands on with my case.Gostev wrote:One theory I have that would explain why some users have issues and other don't is the difference in a difference in the amount of concurrent tasks. So one troubleshooting step for those experiencing lockups would be to reduce the amount of concurrent tasks on the repository by half and see if that makes any difference to stability. Perhaps even change it to 1 task if you have a lab environment where this issue reproduces. Thanks!
I can only imagine that Veeam itself requesting that a repo remove some old block cloned files just like I manually did could result in the same thing happening.
I'm not sure if deleting similarly-sized files on 2016 ReFS which were *not* block cloned would result in the same sort of disaster or not. I'm just trying not to breathe too hard around any of our ReFS servers.
Meanwhile, as insurance, I've put another server in place offsite with ZFS, to which I'm running file-based copies and doing daily ZFS snapshotting of important servers, and covering other SQL/etc servers with another backup product also targeting the ZFS storage. I'd encourage everyone using ReFS on 2016 to do something similar (getting a copy to some other technology than ReFS), because there's serious danger of losing an entire repo. We barely got it to recover after endless attempts on our copy repo after manually deleting those files caused it to blow up. For a while I thought the entire volume was a loss. And if my primary repo, also running ReFS, had died at the same time, or during a re-seed? I don't want to think about that.
-
- VP, Product Management
- Posts: 6035
- Liked: 2860 times
- Joined: Jun 05, 2009 12:57 pm
- Full Name: Tom Sightler
- Contact:
Re: REFS 4k horror story
Deleting files definitely seems to be one of the big triggers. In my testing that was always the point where Windows seemed to get crazy, either when Veeam was deleting lots of files or even if I just started deleting lots of block cloned files manually. I've almost wondered if it would be worthwhile to throttle file deletions on ReFS until Microsoft gets to the root of this problem.
It's been very difficult to find any correlation between users that are having problems vs those that are not, but one pattern that I think I'm starting to notice is that customers that have the most problems are using periodic synthetic fulls vs forever forward or reverse incremental. My theory is that this is because periodic synthetic fulls lead to much bigger file deletion events when a new synthetic full is created and prior full/incremental chains are deleted as a group. It may also be related to people that have much bigger jobs, so many files are merged/deleted near the same time since all of the synthetic processing/deletions happen at the end of the job, but it's really hard to tell because of how disparate the hardware and configurations are.
Note that I'm not saying people running forever forward or reverse couldn't have the problem, but it just seems to trigger more for those with periodic synthetic fulls, but admittedly the sample size is small. The majority of people I talk to that are using ReFS aren't having any issues.
It's been very difficult to find any correlation between users that are having problems vs those that are not, but one pattern that I think I'm starting to notice is that customers that have the most problems are using periodic synthetic fulls vs forever forward or reverse incremental. My theory is that this is because periodic synthetic fulls lead to much bigger file deletion events when a new synthetic full is created and prior full/incremental chains are deleted as a group. It may also be related to people that have much bigger jobs, so many files are merged/deleted near the same time since all of the synthetic processing/deletions happen at the end of the job, but it's really hard to tell because of how disparate the hardware and configurations are.
Note that I'm not saying people running forever forward or reverse couldn't have the problem, but it just seems to trigger more for those with periodic synthetic fulls, but admittedly the sample size is small. The majority of people I talk to that are using ReFS aren't having any issues.
-
- Enthusiast
- Posts: 59
- Liked: 20 times
- Joined: Dec 14, 2016 1:56 pm
- Contact:
Re: REFS 4k horror story
Wouldn't that be everyone using 2016+ReFS+Veeam?tsightler wrote:customers that have the most problems are using periodic synthetic fulls vs forever forward or reverse incremental
I'd love to just use forever forward/reverse, but my understanding is that the only way to to have space-efficient backups with long-term grandfather-father-son points in time is by using 2016-ReFS+Synthetic Fulls? (using the "Keep the following restore points as full backups for archival purposes) ... I'm new to Veeam, so correct me if I'm wrong.
EDIT: Oh, nm, I see you're probably talking about the primary backup job options to the primary repo instead of copy jobs.
-
- Veteran
- Posts: 391
- Liked: 56 times
- Joined: Feb 03, 2017 2:34 pm
- Full Name: MikeO
- Contact:
Re: REFS 4k horror story
Just an update from me. Got a Private message from one of the board members here. I've been stable for almost 2 months no daily reboots by doing the following. Not installing latest KB4013429 that addresses the ReFS issues. Reformatting my array using a smaller block size on the controller(P841) not block size in windows, that is still set to 64k. At the same time I did move the raid controller into a different HP server as well. These changes contributed to my 100% uptime so far. No more failed backups or BSODs. One last thing I did also disable the Data Integrity Scans & The Data Integrity scan for Crash recovery
Code: Select all
Windows PowerShell
Copyright (C) 2016 Microsoft Corporation. All rights reserved.
PS C:\Users\administrator> get-hotfix
Source Description HotFixID InstalledBy InstalledOn
------ ----------- -------- ----------- -----------
KDNAP-UTIL2 Update KB3192137 NT AUTHORITY\SYSTEM 9/12/2016 12:00:00 AM
KDNAP-UTIL2 Update KB3211320 \adm... 3/23/2017 12:00:00 AM
KDNAP-UTIL2 Security Update KB3213986 NT AUTHORITY\SYSTEM 3/23/2017 12:00:00 AM
-
- VP, Product Management
- Posts: 6035
- Liked: 2860 times
- Joined: Jun 05, 2009 12:57 pm
- Full Name: Tom Sightler
- Contact:
Re: REFS 4k horror story
Yes, I was specifically referring to using forever forward/reverse on the primary, or simple retention on the backup copy job target. Believe it or not, the vast majority of customers I work with use simple retention of either 14 or 30 days on their Veeam repositories. Most have a dedupe appliance or some other means (tape) if they need longer retention, because those boxes provide global dedupe which is really useful for longer term retention, especially since the majority of customers I work with have >1000 VMs.graham8 wrote:EDIT: Oh, nm, I see you're probably talking about the primary backup job options to the primary repo instead of copy jobs.
What block size did you use? The general Veeam recommendation has been to recommend larger blocks sizes for RAID, typically 64-128K, but I can see how using something smaller might be useful with ReFS + block clone if the underlying device is RAID5/6 as large block sizes could lead to lots of stripe re-writes.kubimike wrote:Reformatting my array using a smaller block size on the controller(P841) not block size in windows, that is still set to 64k.
-
- Veteran
- Posts: 391
- Liked: 56 times
- Joined: Feb 03, 2017 2:34 pm
- Full Name: MikeO
- Contact:
Re: REFS 4k horror story
@tsightler I am RAID 6+0, per HPs crappy recommendation I was at 512k. I've since blown away the array left the settings default for strip size (128k) issues are gone. I was getting a warning in my controllers log about memory being maxed out all the time when I was 512k. My controller has 4GB on board, found it amazing that it would be full. Those errors are gone too.
-
- Novice
- Posts: 9
- Liked: 1 time
- Joined: Nov 23, 2016 7:41 pm
- Full Name: Brad Jackson
- Contact:
Re: REFS 4k horror story
This is not a Veeam specific issue. I've had it for 6 months now. I'm running a Server 2016 cluster with S2D and Hyper-V. The S2D virtual disks are formatted with ReFS 4k. I've had different hosts hang(black screen, no response to keyboard or mouse, requires hard power reset) in the cluster. Initially, I thought it was a hardware problem. Worked with the hardware vendor for a few months. No problems found by them. Issue "seemed" to happen when one of the nodes in the cluster was over 50% RAM utilization. Only found this thread yesterday(the MS forums are not very helpful) because I didn't think it was a Veeam problem(we run 9.5 to backup all of our VMs, but not to ReFS). The Data Integrity Scan does not correlate to any times that I've had the system hang. For sure the problem is with ReFS. Not sure if it's only Storage Spaces users or anyone running ReFS. My problem I have not figured out how to reproduce consistently(anywhere from 2-90 days between crashes).
Gostev, is there anything that I should reference when opening my Microsoft ticket to help the ReFS team? All of the issues I'm seeing are identical to what other users in this forum are seeing.
For those of you who haven't opened a Microsoft ticket because of the cost, does anyone in your organization have an MSDN account? I learned yesterday that you can open 2 tickets a year for free with EACH MSDN account. I work in a .NET software dev shop, so I will be opening plenty of tickets with Microsoft now.
System Specs for each of the six nodes in our cluster:
-2x Intel Xeon E5-2640
-512GB RAM
-128GB SSD for OS
-2x 2TB Samsung SM863 for Journal in S2D
-10x 8TB Hitachi Ultrastar He8 for triple mirror in S2D
-virtual disks in S2D
--1x 5TB ReFS
--3x 10TB ReFS
Gostev, is there anything that I should reference when opening my Microsoft ticket to help the ReFS team? All of the issues I'm seeing are identical to what other users in this forum are seeing.
For those of you who haven't opened a Microsoft ticket because of the cost, does anyone in your organization have an MSDN account? I learned yesterday that you can open 2 tickets a year for free with EACH MSDN account. I work in a .NET software dev shop, so I will be opening plenty of tickets with Microsoft now.
System Specs for each of the six nodes in our cluster:
-2x Intel Xeon E5-2640
-512GB RAM
-128GB SSD for OS
-2x 2TB Samsung SM863 for Journal in S2D
-10x 8TB Hitachi Ultrastar He8 for triple mirror in S2D
-virtual disks in S2D
--1x 5TB ReFS
--3x 10TB ReFS
-
- Chief Product Officer
- Posts: 31812
- Liked: 7302 times
- Joined: Jan 01, 2006 1:01 am
- Location: Baar, Switzerland
- Contact:
Re: REFS 4k horror story
Brad, thanks for sharing you have the same issue with Hyper-V on S2D. This should certainly raise the priority of the issue in Microsoft's eyes, if that is the case (Hyper-V on S2D is the reference architecture for Hyper-V 2016 anyway). Also, this cancels one of my theories (that our usage of file integrity streams is potentially causing this).
As far as the support case - I offered the ReFS team more joint customers to work with if needed, but I am guessing they probably have enough to collect info from already. I suggest you mention in your support case that you are likely having the same issue as the ReFS team is currently troubleshooting with Veeam customers, this should help them to connect the dots faster once your ticket gets to the dev team through escalations.
Thanks!
As far as the support case - I offered the ReFS team more joint customers to work with if needed, but I am guessing they probably have enough to collect info from already. I suggest you mention in your support case that you are likely having the same issue as the ReFS team is currently troubleshooting with Veeam customers, this should help them to connect the dots faster once your ticket gets to the dev team through escalations.
Thanks!
-
- Service Provider
- Posts: 13
- Liked: 2 times
- Joined: Apr 05, 2017 10:48 am
- Full Name: Richard Kraal
- Contact:
Re: REFS 4k horror story
I've upgraded my dedicated backend (SMB share server hosting the ReFS volume) from 64GB to 384GB and backups are running fine now.
The gateway server is running on a different dedicated machine.
Also the perfmon logs seems to be fine now, no gaps in de logs, also the system does not lockup anymore. RamMap shows a Metafile usage of 50GB (!).
let's see what's going to happen te next weeks.
fingers crossed
used hardware
DL380 Gen9, dual cpu, 384gb ram, 12x 8TB, P841/4GB (64k stripe, Raid6). Win2016 ReFS 64k
The gateway server is running on a different dedicated machine.
Also the perfmon logs seems to be fine now, no gaps in de logs, also the system does not lockup anymore. RamMap shows a Metafile usage of 50GB (!).
let's see what's going to happen te next weeks.
fingers crossed
used hardware
DL380 Gen9, dual cpu, 384gb ram, 12x 8TB, P841/4GB (64k stripe, Raid6). Win2016 ReFS 64k
-
- Veteran
- Posts: 391
- Liked: 56 times
- Joined: Feb 03, 2017 2:34 pm
- Full Name: MikeO
- Contact:
Re: REFS 4k horror story
@richardkraal You running 4.58 on that P841 ?
-
- Service Provider
- Posts: 13
- Liked: 2 times
- Joined: Apr 05, 2017 10:48 am
- Full Name: Richard Kraal
- Contact:
Re: REFS 4k horror story
sorry, it is the P840 (non external version)
we use firmware 4.52
just saw a new firmware for the controller -> 5.04(21 Apr 2017)
will postpone this for a few weeks to see if the memory upgrade works
we use firmware 4.52
just saw a new firmware for the controller -> 5.04(21 Apr 2017)
will postpone this for a few weeks to see if the memory upgrade works
-
- Veteran
- Posts: 391
- Liked: 56 times
- Joined: Feb 03, 2017 2:34 pm
- Full Name: MikeO
- Contact:
Re: REFS 4k horror story
@richardkraal Oh really? I just googled quickly didn't see it. Have the URL handy ?
** edit
found it.
http://h20564.www2.hpe.com/hpsc/swd/pub ... nvOid=4064
Even includes fixes for storage spaces direct.
** edit
found it.
http://h20564.www2.hpe.com/hpsc/swd/pub ... nvOid=4064
Even includes fixes for storage spaces direct.
Who is online
Users browsing this forum: Bing [Bot], Google [Bot] and 112 guests