On Tue, Feb 07, 2006 at 03:16:57PM -0500, Matt Kettler wrote: > My current training ratio is about 7:1 spam:nonspam, but in the past it's been > as bad as 20:1. Both of those are very far off from equal amounts, but the > imbalance has never caused me any problems. > > From my sa-learn --dump magic output as of today: > 0.000 0 995764 0 non-token data: nspam > 0.000 0 145377 0 non-token data: nham
Interesting... it appears I actually need to do a better job of training spam! sa-learn --dump magic|grep am 0.000 0 98757 0 non-token data: nspam 0.000 0 255134 0 non-token data: nham I just changed bayes_auto_learn_threshold_spam to 5.0, we'll see what that does... -- Jim C. Nasby, Database Architect [EMAIL PROTECTED] Give your computer some brain candy! www.distributed.net Team #1828 Windows: "Where do you want to go today?" Linux: "Where do you want to go tomorrow?" FreeBSD: "Are you guys coming, or what?"