And, here's the logs from OOM killer that killed bacula-dir, in case it means anything - the system has 4 G of RAM.
oom-killer: gfp_mask=0x1d2 Mem-info: Node 1 DMA per-cpu: empty Node 1 Normal per-cpu: cpu 0 hot: low 32, high 96, batch 16 cpu 0 hot: low 32, high 96, batch 16 cpu 0 cold: low 0, high 32, batch 16 cpu 1 hot: low 32, high 96, batch 16 cpu 1 cold: low 0, high 32, batch 16 Node 1 HighMem per-cpu: empty Node 0 DMA per-cpu: cpu 0 hot: low 2, high 6, batch 1 cpu 0 cold: low 0, high 2, batch 1 cpu 1 hot: low 2, high 6, batch 1 cpu 1 cold: low 0, high 2, batch 1 Node 0 Normal per-cpu: cpu 0 hot: low 32, high 96, batch 16 cpu 0 cold: low 0, high 32, batch 16 cpu 1 hot: low 32, high 96, batch 16 cpu 1 cold: low 0, high 32, batch 16 Node 0 HighMem per-cpu: empty Free pages: 14732kB (0kB HighMem) Active:816179 inactive:181901 dirty:0 writeback:0 unstable:0 free:3683 slab:3323 mapped:998006 pagetables:5135 Node 1 DMA free:0kB min:0kB low:0kB high:0kB active:0kB inactive:0kB present:0kB pages_scanned:0 all_unreclaimable? no protections[]: 0 0 0 Node 1 Normal free:1352kB min:1408kB low:2816kB high:4224kB active:1810780kB inactive:157412kB present:2031040kB pages_scanned: 4198207 all_unreclaimable? yes protections[]: 0 0 0 Node 1 HighMem free:0kB min:128kB low:256kB high:384kB active:0kB inactive:0kB present:0kB pages_scanned:0 all_unreclaimable? n o protections[]: 0 0 0 Node 0 DMA free:11940kB min:8kB low:16kB high:24kB active:0kB inactive:0kB present:16384kB pages_scanned:3653 all_unreclaimable ? yes protections[]: 0 0 0 Node 0 Normal free:1440kB min:1440kB low:2880kB high:4320kB active:1453880kB inactive:570192kB present:2080764kB pages_scanned: 3227326 all_unreclaimable? yes protections[]: 0 0 0 Node 0 HighMem free:0kB min:128kB low:256kB high:384kB active:0kB inactive:0kB present:0kB pages_scanned:0 all_unreclaimable? n o protections[]: 0 0 0 Node 1 DMA: empty Node 1 Normal: 0*4kB 1*8kB 0*16kB 2*32kB 0*64kB 0*128kB 1*256kB 0*512kB 1*1024kB 0*2048kB 0*4096kB = 1352kB Node 1 HighMem: empty Node 0 DMA: 7*4kB 3*8kB 1*16kB 3*32kB 2*64kB 3*128kB 2*256kB 1*512kB 0*1024kB 1*2048kB 2*4096kB = 11940kB Node 0 Normal: 0*4kB 8*8kB 2*16kB 2*32kB 0*64kB 0*128kB 1*256kB 0*512kB 1*1024kB 0*2048kB 0*4096kB = 1440kB Node 0 HighMem: empty Swap cache: add 1049916, delete 1049914, find 3010/3221, race 0+0 Free swap: 0kB 1032047 pages of RAM 17166 reserved pages 1602 pages shared 2 pages swap cached On Wed, Jun 21, 2006 at 09:25:51AM -0500, Mark Bober wrote: > > I'm attempting to move up from 1.38.0 to the latest (1.38.10) again (we've > had no luck with any version past 1.38.0 on our dual 64 bit Centos 4 box, > what with random crashes and memory becoming exhausted) > > Tonight, I had a tape error, and it was asking for a new tape, happened > around midnight. > > Errored out about 3:45 AM, with: > > > > > ------------- > 21-Jun 03:43 duct-sd: Please mount Volume "CIVER-Base-0002" on Storage Device > "HPSDLT" (/dev/nst1) for Job dagny.raid.2006-06-20_16.17.07 > 21-Jun 03:44 duct-sd: Please mount Volume "CIVE-Base-0003" on Storage Device > "HPSDLT" (/dev/nst1) for Job dagny.raid.2006-06-20_16.17.07 > 21-Jun 03:45 duct-sd: dagny.raid.2006-06-20_16.17.07 Warning: Director wanted > Volume "CIVE-Base-0003" for device "HPSDLT" (/dev/nst1). > Current Volume "CIVER-Base-0002" not acceptable because: > 1998 Volume "CIVER-Base-0002" status is Full, but should be Append, > Purged or Recycle. > 21-Jun 03:45 duct-sd: Please mount Volume "CIVER-Base-0002" on Storage Device > "HPSDLT" (/dev/nst1) for Job dagny.raid.2006-06-20_16.17.07 > 21-Jun 03:46 duct-sd: Please mount Volume "CIVE-Base-0003" on Storage Device > "HPSDLT" (/dev/nst1) for Job dagny.raid.2006-06-20_16.17.07 > 21-Jun 03:47 duct-sd: dagny.raid.2006-06-20_16.17.07 Fatal error: Too many > tries: Wrong Volume mounted on device "HPSDLT" (/dev/nst1): Wanted > CIVE-Base-0003 have CIVER-Base-0002 > 21-Jun 03:47 duct-sd: dagny.raid.2006-06-20_16.17.07 Fatal error: > append.c:207 Fatal append error on device "HPSDLT" (/dev/nst1): > ERR=dev.c:1542 ioctl MTWEOF error on "HPSDLT" (/dev/nst1). ERR=Input/output > error > ------------ > > > Now, it's asking both for the volume it just errored out, *and* the volume it > created for itself in the catalog. It sees the volume it errored, tries to > mount it, remembers that it errored it, and then restarts. That's odd in and > of itself. > > > I'd really rather it not stop trying, or, perhaps, try less often - it was > spitting out 'Please Mount' messages almost every minute for 4 hours, didn't > start any scheduled jobs, and crashed the director. > > I have no "Volume Poll Interval" set, and I haven't found what other > directive would apply in this situation. > > Thanks! > > Mark > > > _______________________________________________ Bacula-users mailing list Bacula-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/bacula-users