ZFS

Is deleting empty snapshots faster?

During the 2025-01-22 OpenZFS Production User Call, ‘atomic operations’ was mentioned with respect to ~/tmp/deleting zfs list -r -t snapshot data01/snapshots/deleting > ~/tmp/deleting 2.54s user 48.47s system 99% cpu 51.042 total 50 seconds. That’s OK. 60,000 deletes starting on the 23rd I started the delete. Actually, it’s not 60,000 deletes. It’s one destroy, of 60,000 snapshots. [12:52 r730-03 dvl ~] % time sudo zfs destroy data01/snapshots/deleting@1%60000 After starting the above command, I started […]

Is deleting empty snapshots faster? Read More »

Way too many snapshots

In this post In this post, we have: FreeBSD 14.1-RELEASE-p5 r730-03 Lots of boring repetitive sections, so skip over that to find what you need This article was written over a couple of days. The zpool in question is 3 pairs of 12TB HDD: [13:44 r730-03 dvl ~/tmp] % zpool status data01 pool: data01 state: ONLINE scan: scrub in progress since Fri Jan 17 05:29:21 2025 22.1T / 26.1T scanned at 200M/s, 19.2T

Way too many snapshots Read More »

Why does the same command appear on two different ports with different times?

While I was destroying some snapshots, I found this interesting / disturbing. Why is the same destroy appearing twice, on different TTYs? [16:11 r730-03 dvl ~/tmp] % w | grep air dvl pts/2 air01.startpoint.vpn.unixathome.org 1:04PM 2:18 sudo zfs destroy data01/snapshots/homeassistant-r730-01@autosnap_2024-10-19_22:15:08_frequently dvl pts/1 air01.startpoint.vpn.unixathome.org Fri11PM – w dvl pts/4 air01.startpoint.vpn.unixathome.org 4:11PM – sudo zfs destroy data01/snapshots/homeassistant-r730-01@autosnap_2024-10-19_22:15:08_frequently dvl pts/5 air01.startpoint.vpn.unixathome.org 1:07PM 33 -zsh (zsh) What’s up with that? And why different times? It is not

Why does the same command appear on two different ports with different times? Read More »

Renaming and replacing zroot filesystems using mfsBSD

This might have interesting tidbits. It’s a draft from June 2020 where I was renaming filesystems when booting from a live thumbdrive. I am replacing one zroot with another because of missing binaries and suspect layout. The existing zroot is an old system predating current zroot layouts. I have already copied over some filesystems via zfs send | zfs recv over ssh. I will now try renaming the filesystems on the new drives

Renaming and replacing zroot filesystems using mfsBSD Read More »

The backup disk image … sparsebundle could not be accessed (error 19)

I wrote this back when and left it in drafts. Time to publishh. I was getting this message when one of my MacBooks was trying to back via TimeMachine to my FreeBSD ZFS server. “The backup disk image … sparsebundle could not be accessed (error 19)” imeMachine also told me the last backup was on Oct 7. Checking my snapshots: [20:01 r730-03 dvl ~] % zfs list -r -t snapshot data01/timemachine/dvl-air01 NAME USED

The backup disk image … sparsebundle could not be accessed (error 19) Read More »

zpool replace

Yesterday, I added a new drive into the system. Today, I will replace the failing drive with that one. In this post: FreeBSD 14.1 Before [18:23 r730-03 dvl ~] % zpool list NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT data01 32.7T 26.0T 6.76T – – 32% 79% 1.00x ONLINE – zroot 412G 18.5G 393G – – 18% 4% 1.00x ONLINE – [18:24 r730-03 dvl ~] % zpool status data01

zpool replace Read More »

kernel: swap_pager: cannot allocate bio

This morning, I saw this in my logs: In this post: FreeBSD 14.1 r730-03 Looking in he logs, there wasn’t much useful: Searching online, I found swap_pager: cannot allocate bio so I decided to limit ARC size. I’ve done this before, and recently too.. Here is what I added to /boot/loader.conf: I just picked the same value I was using for another host. The values in the file can using units, such as

kernel: swap_pager: cannot allocate bio Read More »

Loading up an old copy of a PostgreSQL database, zfs snapshots, and sanoid snapshot management

Sometimes you want to go back to a known good state. And you want to do it quickly, without waiting around. In my case, that’s going to be a zfs snapshot. Today I am preparing to test some code changes so that FreshPorts properly creates a new category. See these recent blog posts for details: Welcome to the new category: filesystems Fixing the category creation code In this post: FreeBSD 14.1-RELEASE-p5 postgresql16-server-16.4 dbclone

Loading up an old copy of a PostgreSQL database, zfs snapshots, and sanoid snapshot management Read More »

zfs hold / release

Let’s talk about zfs hold. Based on man zfs-hold: “If a hold exists on a snapshot, attempts to destroy that snapshot by using the zfs destroy command return EBUSY.” Let’s apply that to some FreshPorts snapshots which I don’t want going away: This is your typical FreshPorts website: [15:03 dvl-nginx01 dvl ~] % zfs list NAME USED AVAIL REFER MOUNTPOINT data02 1.13T 565G 96K none data02/freshports 671G 565G 88K none data02/freshports/jailed 326G 565G

zfs hold / release Read More »

Transferring a VM from one provider to another

Yesterday, I mentioned (in more than one place) that I planned to move a 2017 Digital Ocean droplet over to Azure. As I sit here, with coffee, on the balcony, overlooking lot of green trees, at 7:45 AM, I’m trying to put into words the plan I came up with about 30 minutes ago. In this post: Digital Ocean Microsoft Azure FreeBSD 14.1 Why move? There is no technical issue or dissatisfaction involved

Transferring a VM from one provider to another Read More »

Scroll to Top