AIUI neither LVM nor ext4 have data and metadata checksum and correction features. But, it should be possible to achieve such by including dm-integrity (for checksumming) and some form of RAID (for correction)
in the storage stack. I need to explore that possibility further.
AIUI neither LVM nor ext4 have data and metadata checksum and correction features. But, it should be possible to achieve such by including dm-integrity (for checksumming) and some form of RAID (for correction)
in the storage stack. I need to explore that possibility further.
The most obvious alternative to ZFS on Debian would be Btrfs. Does anyone >> have any comments or suggestions regarding Btrfs and data corruption bugs, >> concurrency, CMM level, PSP, etc.?
If you're worried about such things, I'd think "the most obvious
alternative" is LVM+ext4. Both Btrfs and ZFS share the same underlying problem: more features => more code => more bugs.
Stefan
The most obvious alternative to ZFS on Debian would be Btrfs. Does anyone have any comments or suggestions regarding Btrfs and data corruption bugs, concurrency, CMM level, PSP, etc.?
i intend to create a huge backup server from some oldish hardware.
Hardware has been partly refurbished and offers 1 SSD + 8 HDD on a
6core Intel with 64 GB RAM. ... the [Debian] installer ... aborts.
A friend of mine just let me use an external CD-Drive with the netboot image. ... all is well.
Does anyone have any comments or suggestions regarding how to use magnetic hard disk drives, commodity x86 computers, and Debian for long-term data storage with ensured integrity?
For offline storage:
On Tue, Apr 02, 2024 at 05:53:15AM -0700, David Christensen wrote:
Does anyone have any comments or suggestions regarding how to use magnetic >> hard disk drives, commodity x86 computers, and Debian for long-term data
storage with ensured integrity?
I use LVM on ext4, and I add a MD5SUMS file at the root.
I then power up the drives at least once a year and check the MD5SUMS.
A simple CRC could also work, obviously.
So far, I have not detected MORE corruption with this method than the
drive ECC itself (current drives & buses are much better than they
used to be). When I have errors detected, I replace the file with
another copy (I usually have multiple off-site copies, and sometimes
even on-site online copies, but not always). When the errors add
up, it is time to buy another drive, usually after 5+ years or
even sometimes 10+ years.
So, just re-reading the content might be enough, once a year or so.
This is for HDD (for SDD I have no offline storage experience, it
could be shorter).
So, an ext4 file system on an LVM logical volume?
Why LVM? Are you implementing redundancy (RAID)? Is your data larger than
a single disk (concatenation/ JBOD)? Something else?
Hello,
On Mon, Apr 08, 2024 at 11:28:04AM -0700, David Christensen wrote:
So, an ext4 file system on an LVM logical volume?
Why LVM? Are you implementing redundancy (RAID)? Is your data larger than >> a single disk (concatenation/ JBOD)? Something else?
For off-site long-term offline archiving, no, I am not using RAID.
No, it's not LVM+md, just plain LVM for flexibility.
Typically I use 16 TB hard drives, and I tend to use one LV per data
source, the LV name being the data source and the date of the copy.
Or sometimes I just copy a raw volume (ext4 or something else)
to a LV.
With smaller drives (4 TB) I tend to not use LVM, just plain ext4 on the
raw disk.
I almost never use partitionning.
However, I tend to use luks encryption (per ext4 filesystem) when the
drives are stored off-site. So it's either LVM -> LV -> LUKS -> ext4
or raw disk -> LUKS -> ext4.
You can find some of the scripts I use to automate this off-site
long-term archiving here:
https://git.alphanet.ch/gitweb/?p=various;a=tree;f=offsite-archival/LVM-LUKS
Does anyone have any comments or suggestions regarding how to use
magnetic hard disk drives, commodity x86 computers, and Debian for
long-term data storage with ensured integrity?
[1] https://github.com/openzfs/zfs/issues/15526
[2] https://github.com/openzfs/zfs/issues/15933
On 02/04/2024 13:53, David Christensen wrote:
Does anyone have any comments or suggestions regarding how to use
magnetic hard disk drives, commodity x86 computers, and Debian for
long-term data storage with ensured integrity?
I use Btrfs, on all my systems, including some servers, with soft Raid1
and Raid10 modes (because these modes are considered stable and
production ready). I decided on Btrfs not ZFS, because Btrfs allows to migrate drives on the fly while partition is live and heavily used,
replace them with different sizes and types, mixed capacities, change
Raid levels, change amount of drives too. I could go from single drive
to Raid10 on 4 drives and back while my data is 100% available at all
times.
It saved my bacon many times, including hard checksum corruption on NVMe drive which otherwise I would never know about. Thanks to Btrfs I
located the corrupted files, fixed them, got hardware replaced under warranty.
Also helped with corrupted RAM: Btrfs just refused to save file because
saved copy couldn't match read checksum from the source due to RAM bit
flips. Diagnosed, then replaced memory, all good.
I like a lot when one of the drives get ATA reset for whatever reason,
and all other drives continue to read and write, I can continue using
the system for hours, if I even notice. Not possible in normal
circumstances without Raid. Once the problematic drive is back, or after reboot if it's more serious, then I do "scrub" command and everything is resynced again. If I don't do that, then Btrfs dynamically correct
checksum errors on the fly anyway.
And list goes on - I've been using Btrfs for last 5 years, not a single problem to date, it survived hard resets, power losses, drive failures, countless migrations.
[1] https://github.com/openzfs/zfs/issues/15526
[2] https://github.com/openzfs/zfs/issues/15933
Problems reported here are from Linux kernel 6.5 and 6.7 on Gentoo
system. Does this even affects Debian Stable with 6.1 LTS?
--
With kindest regards, Piotr.
⢀⣴⠾⠻⢶⣦⠀
⣾⠁⢠⠒⠀⣿⡁ Debian - The universal operating system ⢿⡄⠘⠷⠚⠋⠀ https://www.debian.org/
⠈⠳⣄⠀⠀⠀⠀
I use Btrfs, on all my systems, including some servers, with soft Raid1
and Raid10 modes (because these modes are considered stable and
production ready). I decided on Btrfs not ZFS, because Btrfs allows to
migrate drives on the fly while partition is live and heavily used,
replace them with different sizes and types, mixed capacities, change
Raid levels, change amount of drives too. I could go from single drive
to Raid10 on 4 drives and back while my data is 100% available at all
times.
It saved my bacon many times, including hard checksum corruption on NVMe
drive which otherwise I would never know about. Thanks to Btrfs I
located the corrupted files, fixed them, got hardware replaced under
warranty.
Also helped with corrupted RAM: Btrfs just refused to save file because
saved copy couldn't match read checksum from the source due to RAM bit
flips. Diagnosed, then replaced memory, all good.
I like a lot when one of the drives get ATA reset for whatever reason,
and all other drives continue to read and write, I can continue using
the system for hours, if I even notice. Not possible in normal
circumstances without Raid. Once the problematic drive is back, or after
reboot if it's more serious, then I do "scrub" command and everything is
resynced again. If I don't do that, then Btrfs dynamically correct
checksum errors on the fly anyway.
And list goes on - I've been using Btrfs for last 5 years, not a single
problem to date, it survived hard resets, power losses, drive failures,
countless migrations.
Those sound like some compelling features.
On 4/9/24 17:08, piorunz wrote:
On 02/04/2024 13:53, David Christensen wrote:
Does anyone have any comments or suggestions regarding how to use
magnetic hard disk drives, commodity x86 computers, and Debian for
long-term data storage with ensured integrity?
I use Btrfs, on all my systems, including some servers, with soft Raid1
and Raid10 modes (because these modes are considered stable and
production ready). I decided on Btrfs not ZFS, because Btrfs allows to
migrate drives on the fly while partition is live and heavily used,
replace them with different sizes and types, mixed capacities, change
Raid levels, change amount of drives too. I could go from single drive
to Raid10 on 4 drives and back while my data is 100% available at all
times.
It saved my bacon many times, including hard checksum corruption on NVMe
drive which otherwise I would never know about. Thanks to Btrfs I
located the corrupted files, fixed them, got hardware replaced under
warranty.
Also helped with corrupted RAM: Btrfs just refused to save file because
saved copy couldn't match read checksum from the source due to RAM bit
flips. Diagnosed, then replaced memory, all good.
I like a lot when one of the drives get ATA reset for whatever reason,
and all other drives continue to read and write, I can continue using
the system for hours, if I even notice. Not possible in normal
circumstances without Raid. Once the problematic drive is back, or after
reboot if it's more serious, then I do "scrub" command and everything is
resynced again. If I don't do that, then Btrfs dynamically correct
checksum errors on the fly anyway.
And list goes on - I've been using Btrfs for last 5 years, not a single
problem to date, it survived hard resets, power losses, drive failures,
countless migrations.
Those sound like some compelling features.
I believe the last time I tried Btrfs was Debian 9 (?). I ran into
problems because I did not do the required manual maintenance (rebalancing). Does the Btrfs in Debian 11 or Debian 12 still require manual maintenance? If so, what and how often?
Am 10.04.2024 um 13:10 schrieb David Christensen:
Does the Btrfs in Debian 11 or Debian 12 still require
manual maintenance? If so, what and how often?
Scrub and balance are actions which have been recommended. I am using btrfsmaintenance scripts [1][2] to automate this. I am doing a weekly
balance and a monthly scrub. After some reading today, I am getting
unsure if this is approach is correct, especially if balance is
necessary anymore (it usually doesn't find anything to do anyway), so
please take these periods with caution. My main message is that such operations can be automated using the linked scripts.
Best regards,
Paul
[1] https://packages.debian.org/bookworm/btrfsmaintenance
[2] https://github.com/kdave/btrfsmaintenance
Those sound like some compelling features.
I believe the last time I tried Btrfs was Debian 9 (?). I ran into
problems because I did not do the required manual maintenance (rebalancing). Does the Btrfs in Debian 11 or Debian 12 still require manual maintenance? If so, what and how often?
On 10/04/2024 12:10, David Christensen wrote:
Those sound like some compelling features.
I believe the last time I tried Btrfs was Debian 9 (?). I ran into
problems because I did not do the required manual maintenance
(rebalancing). Does the Btrfs in Debian 11 or Debian 12 still require
manual maintenance? If so, what and how often?
I don't do balance at all, it's not required.
Scrub is recommended, because it will detect any bit-rot due to hardware errors on HDD media. It scans the entire surface of allocated sectors on
all drives. I do scrub usually monthly.
For off-site long-term offline archiving, no, I am not using RAID.
https://github.com/t13a/dm-integrity-benchmarks
Contenders are btrfs, zfs, and notably ext4+dm-integrity+dm-raid
On Mon, Apr 08, 2024 at 10:04:01PM +0200, Marc SCHAEFER wrote:
For off-site long-term offline archiving, no, I am not using RAID.
Now, as I had to think a bit about ONLINE integrity, I found this
comparison:
https://github.com/t13a/dm-integrity-benchmarks
Contenders are btrfs, zfs, and notably ext4+dm-integrity+dm-raid
I tend to have a biais favoring UNIX layered solutions against
"all-into-one" solutions, and it seems that performance-wise,
it's also quite good.
I wrote this script to convince myself of auto-correction
of the ext4+dm-integrity+dm-raid layered approach.
Thank you for devising a benchmark and posting some data. :-)
FreeBSD also offers a layered solution. From the top down:
Sysop: | Keyop |
---|---|
Location: | Huddersfield, West Yorkshire, UK |
Users: | 546 |
Nodes: | 16 (2 / 14) |
Uptime: | 04:53:02 |
Calls: | 10,386 |
Calls today: | 1 |
Files: | 14,058 |
Messages: | 6,416,627 |