Replies: 1 comment
-
We did fix the situation by keeping the pool but sending a fresh copy of the dataset. The corruption has not resurfaced on the new dataset. I see several closed issues about send/receive corruption with 1M datasets. It was hopefully one of those we hit and the bug is already fixed. Unfortunately, I cannot know for sure since the replication history of the original dataset is not fully documented. As this discussion has not received any response I am now going to destroy the faulty dataset and further investigation will not be possible. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
We have experienced silent data corruption on a replication target. I first thought the problem was related to #6224 but this happened on TrueNAS CORE 12.0-U5 which should already have a fix for that issue.
Most files are okay, some are holes (i.e. zeroed) and other files look like this:
That looks like 1MB of data in a 128k record? (The dataset currently has a 1M recordsize but this particular file has 128k records on the replication source)
I can
zfs send
the bad dataset, with a resulting stream like this:but the kernel panics on
zfs receive
:Is there some known bug/gotcha that can cause this situation?
Can I fix the situation by destroying the affected dataset or should I rebuild the entire pool?
Beta Was this translation helpful? Give feedback.
All reactions