Matt (but not just to Matt), I don't understand your reply (though I am deeply in your dept for the work you do for this community). The sample emails that Nigel posted are identical in content, including obfuscation. I've noted the same situation. Yet, the scoring is really different. On the low scoring ones, DCC and RAZOR2 didn't hit, and the BAYES score is different. The main differences are in the headers' different forged From and To addresses. I thought these samples were worthy of deeper analysis.

Sincerely,
Andy Figueroa

Matt Kettler wrote:
Nigel Frankcom wrote:
Hi All,

Does anyone have any idea why there are such scoring disparities
between these two emails? I've been seeing a few of these creep
through lately.

http://dev.blue-canoe.net/spam/spam01.txt
http://dev.blue-canoe.net/spam/spam02.txt
http://dev.blue-canoe.net/spam/spam03.txt
http://dev.blue-canoe.net/spam/spam04.txt

More to the point with these is why are they not hitting any of the
drugs rules?

There's a few million obfuscation methods, and the rules can't always
cover em all.

The examples you posted are using "duplicated letters", as well as
inserted underscores.

The old Antidrug rules (part of xx_drugs.cf now) that I wrote will deal
with the underscores, and a wide range of character substitutions, but
only a few special-cases of insertions.

It's taken the spammers a long time to figure that out, but it appears
they finally have.

I used to have to update the set constantly, but lately I've been a bit
too busy with real life.

Reply via email to