On Tue, Feb 07, 2006 at 03:16:57PM -0500, Matt Kettler wrote:
> My current training ratio is about 7:1 spam:nonspam, but in the past it's been
> as bad as 20:1. Both of those are very far off from equal amounts, but the
> imbalance has never caused me any problems.
> 
> From my sa-learn --dump magic output as of today:
> 0.000          0     995764          0  non-token data: nspam
> 0.000          0     145377          0  non-token data: nham

Interesting... it appears I actually need to do a better job of training
spam!
sa-learn --dump magic|grep am
0.000          0      98757          0  non-token data: nspam
0.000          0     255134          0  non-token data: nham

I just changed bayes_auto_learn_threshold_spam to 5.0, we'll see what
that does...
-- 
Jim C. Nasby, Database Architect                [EMAIL PROTECTED] 
Give your computer some brain candy! www.distributed.net Team #1828

Windows: "Where do you want to go today?"
Linux: "Where do you want to go tomorrow?"
FreeBSD: "Are you guys coming, or what?"

Reply via email to