hardware

stressdisk

This was written a few weeks ago and I just noticed today that it had not been published. Following on from I bought a new nvme drive – or did I?, I ran stressdisk on it. [dvl@test ~]$ sudo gpart add -a 4k -t freebsd-zfs nvd0 nvd0p1 added [dvl@test ~]$ gpart show nvd0 => 34 500118125 nvd0 GPT (238G) 34 6 – free – (3.0K) 40 500118112 1 freebsd-zfs (238G) 500118152 7 – […]

stressdisk Read More »

I bought a new nvme drive – or did I?

Bought a new NVMe drive. Installed it using cheap PCIe adaptor bought off Amazon. From /var/run/dmesg.boot nvd0: NVMe namespace nvd0: 244198MB (500118192 512 byte sectors) Surprisingly, the device is already partitioned. [dvl@test ~]$ gpart show nvd0 => 34 500118125 nvd0 GPT (238G) 34 500118125 – free – (238G) [dvl@test ~]$ Eh, is is this a used device? The static bag it arrived in was sealed. Let’s see. [dvl@test ~]$ sudo smartctl -a /dev/nvd0

I bought a new nvme drive – or did I? Read More »

Accessing a Dell iDRAC 7 via Nginx reverse proxy

HEADS UP: this is still under review. Until this notice is removed, I recommend not using this. I have a Dell R720 at an ISP. The iDRAC connection is filtered, by the ISP, and is accessible only from a small number of IP addresses. Sure, this talks about a web proxy. It works. So does the ssh forwarding shown at the bottom of this post. NOTE: I should start monitoring that IP addr

Accessing a Dell iDRAC 7 via Nginx reverse proxy Read More »

Eaton EMA114-10 ePDU – initial configuration

On Friday, the latest addition to the #homelab turned up. An Eaton EMA114-10 PDU. A PDU is a Power Distribution Unit. Think of it as a step up from a power strip. This is a 24 outlet, single phase, managed PDU. With it, individual outlets can be monitored for power usage and they can be switched on or off. Access methods include ssh, telnet, serial port, http, and https. This PDU is zero-U,

Eaton EMA114-10 ePDU – initial configuration Read More »

Eaton ePDU Auth Algorithm & Crypto Algorithm for SNMPv3 Configuration

I recently acquired an Eaton EMA114-10 PDU. When adding it to LibreNMS, I did not know what Authorization Algorithm or Crypto Algorithm to use when configuring it for SNMPv3. Searching did not find anything. IRC did. dagb on the LibreNMS IRC channel knew what was used for older EATON devices. Turns out, that hasn’t changed. These are the answers: Auth Algorithm – MD5 Crypto Algorithm – AES I hope this saves you time.

Eaton ePDU Auth Algorithm & Crypto Algorithm for SNMPv3 Configuration Read More »

writing random data via geli

You might recall that suspect drive from the zpool replace on the weekend. Thomas Hurst suggested: Might be worth overwriting the drive, try to encourage it to actually reallocate the sectors now the data on them is no longer needed. I, being one to take advice from people on the internet, and Michael W Lucas, decided to try his suggestion. The drive in question. [dan@knew:~] $ tail /var/log/messages Dec 14 00:00:00 knew newsyslog[88570]:

writing random data via geli Read More »

Replacing a failing drive in a ZFS zpool

In this post I will replace a working, but suspect, drive with another drive. No down time. The server is knew. In this post: FreeBSD 12.2 ZFS TOSHIBA MD04ACA500 5TB drive – the suspect drive: da17 TOSHIBA HDWE150 5TB drive – the replacement: da22 None of these drives are under warranty What drives are in this server? I have had good luck with Toshiba DT01ACA300 3TB drives (presenting as Hitachi HDS723030BLE640), starting a

Replacing a failing drive in a ZFS zpool Read More »

Duplicating a zpool

I have these boot drives, I want to copy them to another. They both happen to be zroot and bootable. I booted from mfsBSD and imported both new and old zroot, using the pool ID shown by zpool import -N. root@mfsbsd:~ # zpool import -N pool: data01 id: 2668514456528412656 state: ONLINE action: The pool can be imported using its name or numeric identifier. config: data01 ONLINE mirror-0 ONLINE gpt/S59VNS0N809087J_S00 ONLINE gpt/S59VNJ0N631973D_S01 ONLINE mirror-1

Duplicating a zpool Read More »

Creating a 4 drive ZFS zpool: stripe over mirrors

In this post, I am creating a 4 drive ZFS pool consisting of two mirror. In this post: I’m booting from mfsBSD 12.1 ZFS Creating partitions On FreeBSD, it is convention to create partitions. It is not necessary, but I don’t use all the space in case a future replacement drive is slightly smaller than these drives. NOTE: I am not, but have not confirmed that this is no longer a problem in

Creating a 4 drive ZFS zpool: stripe over mirrors Read More »

Scroll to Top