Ryan,

> Configuration:  Spam Filter Store and Forward Gateway (non authenticated)

You may want to add clamd to the mix.

> I want to make sure this is as optimized as possible to provide a fair
> performance picture versus SpamAssassin and Barracuda.
>...
> I also have read a lot where people are improving accuracy by increasing
> the scoring of the Bayesian database (which needs training).
>...
> SpamAssassin is typically represented as a magic dance of tweaking rules. 
> Are the default rule thresholds good values to start at?  How can I
> adequately decide which rules to tweak and how much to tweak them by?  In
> other words, how do you manage your adjustments without users noticing wide
> spam classifying variations?

For a fair comparision to other products, it would look good to do as
little tweaking to SA as reasonable. The default SA 3.1.1 scores are
quite well adjusted. The BAYES_99 used to be too low, but I believe
with SA 3.1 it now defaults to 3.5, which is just fine. Have a default
set of SARE rules (or hand-picked, if you prefer), and most if not all
network tests enabled.

After evaluation period is over, you might want to tweak a score or two,
but that will come from your experience.

> I could pretty much trust a small subset of users to be fairly regular in
> their training.  There is a somewhat larger portion of users who would
> train here and there.  Lastly, the largest portion of users may never
> train.  We also do not know which user belongs to which group (yet).  With
> this scenario it seems that I will have to use some kind of common
> database.  In the default configuration SA uses one Bayesian database for
> all users.  Is there a reason to change this?

For a company (a rather homogeneous user group), a common Bayesian database 
works fairly well, even left to itself for automatic learning only, the 
provided set of static (+SARE) and network tests alone does a remarkable job
of training Bayes, which after a while starts to contribute back what it
has learned. For an ISP with a diverse user population, that would't work
so well. Manual training does't hurt, but there is no need to rely heavily
on it, uneducated users may even do more damage than good when reporting
what they consider spam.


> It also seems that there is a falling out between pyzor, dcc, razor, and
> the community.  Is it simply a licensing issue (with legal implications),
> or are these systems flawed otherwise.

If license permits, use them all.
It might be interesting to tell that of all these, the pyzor is
the largest CPU consumer. Use it if you can afford, but it can be
the first to drop if host resources are scarce. Don't let network
latencies bother you (e.g. Razor or DNS checks can be slow at times),
the wait time can be easily utilized by other parallel processes.

> Do I even need this functionality?  Has anyone seen a detriment to
> SpamAssassin's performance without DCC, Pyzor, or Razor.

Yes, you need them, they often contribute the ultimate score to tip
the scale. Without networks tests (and static rules) you end up
in a dspam league.

> What about an initial corpus to train the Bayesian database?  Will this
> hurt my accuracy in the long term?  What corpuses are being used?  Am I
> better off letting the Bayesian autolearn gradually perform this function?

Let it auto-learn for a week, then start evaluation test.

If you do hand training, and you are comparing to dpam,
try to train both systems approximately the same.

  Mark

Reply via email to