ZFS

Adding in a stand-by spare drive for ZFS on FreeBSD

This is a follow up to Adding in a hot-spare for zfs on FreeBSD from two months ago. The replacement for the returned drive has arrived and after sitting for many weeks on my coffee table, it is installed in r730-03. Here it is, as found in /var/log/messages – this host is a Dell R730 which has drive cages allowing me to insert the drive without powering off the host. Jun 30 16:17:15

Adding in a stand-by spare drive for ZFS on FreeBSD Read More »

What drive did I just remove from the system?

So there I was… ready to remove the drive from the system. This was the drive which was giving errors and which had already been replaced. In this post: FreeBSD 14.2 Let’s look at the drive I just wiped … I was doing this command: [20:11 r730-03 dvl ~] % sudo dd if=/dev/zero of=/dev/gpt/HGST_8CJVT8YE bs=4M Let’s run it again and see which drive LED lights up. Yep, there it is. CTL-C, LED goes

What drive did I just remove from the system? Read More »

Adding in a hot-spare for zfs on FreeBSD

But first, there’s more Right after asking “Anyone running zfsd? Did you do anything in particular to configure it? I just added my first hot-spare to a zpool.”, ivy told me “noooooooo don’t use hot spares!! Keep a cold spare or at least an online device not attached to a pool. otherwise your zpool will randomly decide to attach its hot spare due to a temporary cabling issue or something like that. the

Adding in a hot-spare for zfs on FreeBSD Read More »

Is deleting empty snapshots faster?

During the 2025-01-22 OpenZFS Production User Call, ‘atomic operations’ was mentioned with respect to ~/tmp/deleting zfs list -r -t snapshot data01/snapshots/deleting > ~/tmp/deleting 2.54s user 48.47s system 99% cpu 51.042 total 50 seconds. That’s OK. 60,000 deletes starting on the 23rd I started the delete. Actually, it’s not 60,000 deletes. It’s one destroy, of 60,000 snapshots. [12:52 r730-03 dvl ~] % time sudo zfs destroy data01/snapshots/deleting@1%60000 After starting the above command, I started

Is deleting empty snapshots faster? Read More »

Way too many snapshots

In this post In this post, we have: FreeBSD 14.1-RELEASE-p5 r730-03 Lots of boring repetitive sections, so skip over that to find what you need This article was written over a couple of days. The zpool in question is 3 pairs of 12TB HDD: [13:44 r730-03 dvl ~/tmp] % zpool status data01 pool: data01 state: ONLINE scan: scrub in progress since Fri Jan 17 05:29:21 2025 22.1T / 26.1T scanned at 200M/s, 19.2T

Way too many snapshots Read More »

Why does the same command appear on two different ports with different times?

While I was destroying some snapshots, I found this interesting / disturbing. Why is the same destroy appearing twice, on different TTYs? [16:11 r730-03 dvl ~/tmp] % w | grep air dvl pts/2 air01.startpoint.vpn.unixathome.org 1:04PM 2:18 sudo zfs destroy data01/snapshots/homeassistant-r730-01@autosnap_2024-10-19_22:15:08_frequently dvl pts/1 air01.startpoint.vpn.unixathome.org Fri11PM – w dvl pts/4 air01.startpoint.vpn.unixathome.org 4:11PM – sudo zfs destroy data01/snapshots/homeassistant-r730-01@autosnap_2024-10-19_22:15:08_frequently dvl pts/5 air01.startpoint.vpn.unixathome.org 1:07PM 33 -zsh (zsh) What’s up with that? And why different times? It is not

Why does the same command appear on two different ports with different times? Read More »

Renaming and replacing zroot filesystems using mfsBSD

This might have interesting tidbits. It’s a draft from June 2020 where I was renaming filesystems when booting from a live thumbdrive. I am replacing one zroot with another because of missing binaries and suspect layout. The existing zroot is an old system predating current zroot layouts. I have already copied over some filesystems via zfs send | zfs recv over ssh. I will now try renaming the filesystems on the new drives

Renaming and replacing zroot filesystems using mfsBSD Read More »

The backup disk image … sparsebundle could not be accessed (error 19)

I wrote this back when and left it in drafts. Time to publishh. I was getting this message when one of my MacBooks was trying to back via TimeMachine to my FreeBSD ZFS server. “The backup disk image … sparsebundle could not be accessed (error 19)” imeMachine also told me the last backup was on Oct 7. Checking my snapshots: [20:01 r730-03 dvl ~] % zfs list -r -t snapshot data01/timemachine/dvl-air01 NAME USED

The backup disk image … sparsebundle could not be accessed (error 19) Read More »

zpool replace

Yesterday, I added a new drive into the system. Today, I will replace the failing drive with that one. In this post: FreeBSD 14.1 Before [18:23 r730-03 dvl ~] % zpool list NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT data01 32.7T 26.0T 6.76T – – 32% 79% 1.00x ONLINE – zroot 412G 18.5G 393G – – 18% 4% 1.00x ONLINE – [18:24 r730-03 dvl ~] % zpool status data01

zpool replace Read More »

Scroll to Top