Search Results for: bacula

Copying to tape, missed something….

I recently wrote about copying jobs with a very good backup tool. My goal was to keep copies of the backups on both disk and tape. Since then, I started playing around with a little webpage to keep tabs on what’s in my tape libary. Today I noticed a little problem. I have backups from 75 days ago that are waiting to be copied to tape.

I’ll just let you think about why this happened. Yesterday, this queue held jobs that were about 16 hours old. How did this old job get into the queue?

Copying to tape, missed something…. Read More »

zfs destroy

Here I go, removing an old ZFS dataset that is no longer required. But first, some stats (which have been edited to remove non-zfs output): [dan@kraken:~] $ df -h Filesystem Size Used Avail Capacity Mounted on storage 1.4T 15G 1.4T 1% /storage storage/bacula.DO.NOT.USE 5.8T 4.4T 1.4T 76% /storage/bacula.DO.NOT.USE storage/compressed 1.4T 49K 1.4T 0% /storage/compressed storage/compressed/bacula

zfs destroy Read More »

How much are those volumes in the library? (redone)

I have an improvement on the original query. This one outputs ‘nil’ if the volume is empty. bacula=# SELECT mediaid, volumename, volstatus, slot, inchanger, pool.name, CASE WHEN volbytes < = 64512 THEN 'nil' else pg_size_pretty(volbytes) END FROM media, pool WHERE media.poolid = pool.poolid AND volumename LIKE 'ETU0%' AND inchanger != 0 ORDER BY slot; mediaid

How much are those volumes in the library? (redone) Read More »

How much are those volumes in the library?

I pretty much ignore my tape library. It asks for tapes when it runs out. However, I think I should be looking ahead and anticipating requests. Thus, I came up with this SQL query: [dan@ngaio:~] $ psql bacula psql (8.4.5) Type “help” for help. bacula=# SELECT mediaid, volumename, volstatus, slot, inchanger, pool.name, volbytes FROM media,

How much are those volumes in the library? Read More »

Why not use the tape that’s there?

*run A job name must be specified. The defined Job resources are: 1: BackupCatalog 2: RestoreFiles 3: MigrateDiskToTape 4: CopyMegaFileToTape 5: laptop 6: laptop-vpn 7: laptop-wifi-5.3 8: laptop-verify 9: laptop-websites 10: laptop-xp 11: wocker 12: bast basic 13: bast home 14: dbclone basic 15: dbclone home 16: polo basic 17: polo 18: kraken basic 19:

Why not use the tape that’s there? Read More »

zfs send | zfs receive

For the record, I was doing a copy within a system recently: # time zfs send storage/bacula@transfer | mbuffer | zfs receive storage/compressed/bacula-buffer in @ 0.0 kB/s, out @ 0.0 kB/s, 4394 GB total, buffer 100% fullll summary: 4394 GByte in 35 h 11 min 35.5 MB/s real 2111m18.302s user 11m44.831s sys 258m59.151s I was

zfs send | zfs receive Read More »

ntp wasn’t running but Nagios didn’t notice

Earlier today, I noticed the following output from a Bacula job: 24-Sep 14:14 bacula-dir JobId 38548: Start Backup JobId 38548, Job=latens_home.2010-09-24_14.12.38_31 24-Sep 14:14 bacula-dir JobId 38548: Using Device “MegaFile-latens” 24-Sep 14:09 latens-fd JobId 38548: DIR and FD clocks differ by -307 seconds, FD automatically compensating. That’s 5 minutes. It shouldn’t be varying by that much.

ntp wasn’t running but Nagios didn’t notice Read More »

Graphing

I have about 14,000 data points collected from backups. I started in Oct 2007, but have not done anything with the data. I’d like your help. I have put the data into a PostgreSQL database and dumped it. The dump is here: http://langille.org/tmp/dlt_stats.sql.gz The table looks like: $ psql dan psql (8.4.2) Type “help” for

Graphing Read More »

Scroll to Top