FreeBSD

zpool degraded – one drive missing from system

I rebooted knew yesterday for upgrades. When it came back, the main storage zpool was degraded: Is the drive alive? The drive is not listed at all in /var/run/dmesg.boot. I keep a list of the expected drives in /etc/periodic.conf, for use by a Nagios check: [dan@knew:~] $ /usr/sbin/sysrc -nf /etc/periodic.conf daily_status_smart_devices /dev/da22 /dev/da21 /dev/da20 /dev/da19 /dev/da18 /dev/da17 /dev/da16 /dev/da15 /dev/da14 /dev/da13 /dev/da12 /dev/da11 /dev/da10 /dev/da9 /dev/da8 /dev/da7 /dev/da6 /dev/da5 /dev/da4 /dev/da3 /dev/da2 /dev/da1 […]

zpool degraded – one drive missing from system Read More »

Creating a drive-bay map

When the time comes to replace a drive, it is very nice to know which drives is missing. I created this drive map to help me figure out which drive disappeared. I created this drive-bay map using a combination of: zpool status sesutil map lsblk camcontrol /var/run/dmesg.boot I have not included /var/run/dmesg.boot here. If you click on this image, you’ll see a larger version which is easier to read. I went through sesutil

Creating a drive-bay map Read More »

Migrating a Dell TL4000 to a new FreeBSD server and attaching it to a jail

I recently migrated a bunch of jails from one server to another. Today I attached the Dell TL4000 tape library. A jail on this server copied Bacula backups from disk to tape. In this post: FreeBSD 12.0 Bacula 9.4.3 Dell R720 Investigation As anticipated, I needed to update the server configuration to cope with changed device numbers. The symptoms: when running the update slots command from within bconsole, I was getting this response.

Migrating a Dell TL4000 to a new FreeBSD server and attaching it to a jail Read More »

Oct 4 09:01:24 slocum kernel: pid 1409 (check_bacula), uid 181: exited on signal 11

I’ve been getting these messages in /var/log/messages on slocum for as long as I can remember. Today I found out why those errors are occurring. They are logged on the FreeBSD jail host for a Nagios instance I run. Nagios runs in a jail on that host. I’ve just been ignoring the messages, but today it got me. I was having trouble getting a new FreeBSD port to work, so I was easily

Oct 4 09:01:24 slocum kernel: pid 1409 (check_bacula), uid 181: exited on signal 11 Read More »

How allow.mount.zfs affects mountpoints for ZFS

I noticed this the other day, and thought it was interesting. When the jailed property is set on a ZFS fileset, it affects the mountpoint within the jail. If your jail uses allow.mount.zfs (known as allow_mount_zfs when using iocage), the mountpoints become relative to the jail. For example, in my poudriere jail, this is what I see the main dataset: [dan@pkg01:~] $ zfs list tank_fast/poudriere NAME USED AVAIL REFER MOUNTPOINT tank_fast/poudriere 51.1G 452G

How allow.mount.zfs affects mountpoints for ZFS Read More »

Migrating all your iocage jails to a new host

In this post, I’m not exporting an iocage jail to another host. I am moving the entire iocage instance to another host. This is accomplished by doing a zfs export on the zpool, moving the drives to a new system, and doing a zfs import. The drive migration is covered in a previous post. In this post: FreeBSD iocage old host: R710 – r710-01 new host: R720 – r720-01 The first power The

Migrating all your iocage jails to a new host Read More »

Moving poudriere from the host into a jail

This post is all about moving poudriere from the host into a jail, but you could probably use it for creating a new jail and running poudriere in it. NOTE: If you’re looking for a jail configuration for poudriere, please refer to Configuration for running poudriere in a jail on FreeBSD 14 – it contains new stuff, like mlock (thanks to feld for pointing this out). This also assumes use of iocage, but

Moving poudriere from the host into a jail Read More »

My plan for moving the R710 into the R720

Today the drive caddies arrived for the R720. I refer to the services provided by the R710, not the server itself. I will list those services later and outline how I want to move them. I could do all this over this coming weekend but I have already allocated that time to some errands I have to catch up on. Physical things This section discusses the physical things which must move. Drives There

My plan for moving the R710 into the R720 Read More »

Going from partition to label in zpool status

In my recent post you saw this: [dan@r720-01 ~]$ zpool status zroot pool: zroot state: ONLINE scan: resilvered 818M in 0 days 00:00:09 with 0 errors on Tue Oct 15 21:32:12 2019 config: NAME STATE READ WRITE CKSUM zroot ONLINE 0 0 0 mirror-0 ONLINE 0 0 0 ada0p3 ONLINE 0 0 0 gpt/zfs1 ONLINE 0 0 0 errors: No known data errors [dan@r720-01 ~]$ I want to change ada0p3 to gpt/zfs0, as

Going from partition to label in zpool status Read More »

Creating a mirror from your zroot

I planned to create a zroot mirror when I installed FreeBSD on the R720 but the supplier shipped only one of the two purchased SSDs. Today the second drive arrived. Let’s create a zroot mirror. In this post: FreeBSD 12 man 8 zpool Posts I looked at: creating a mirrored ZFS Pool out of a single-disk rpool – not FreeBSD Convert Single disk ZFS-On-Root to Mirror – used the GUID via zbd; I

Creating a mirror from your zroot Read More »

Scroll to Top