>...
>List Mail User wrote:
>> Huh?  (Lookup "strawman" in a dictionary, please.)
>That's my understanding of what you were claiming happened. Yes, it
>looks like an absurdly weak argument. However, it's the argument you
>presented, as best I can make sense of your posts.
>
>Or are you admitting that you made those arguments intentionally as a
>straw man to confuse the issues?
>
>> Scenario 3:)
>>      A hardware store posts a *direct* link to a rebate form (or had a
>> pad of them on or near a shelf);  Customer prints (or takes a copy of) the
>> rebate form (with no opportunity to every see the Scotts' web page telling
>> him that any email he provides will be used for marketing);  Customer buys
>> a Scotts' product at the store and mails in the printed form,  Customer then
>> receives spam.  QED
>>
>>      Please read the rebate form, then read the Scotts' privacy policy.
>>   
>
>As I read the form, and Scotts privacy policy, Scotts cannot send you
>additional marketing information by email just because you entered this
>contest. As I read it they can only send you requested mail, but it
>depends on how you interpret the sentence structure.
>
>Regardless, even if we accept your theory that their privacy policy
>allows it, it doesn't prove, or even suggest, that they did.
>
>I find your willingness to accept a twisted reading of a privacy policy
>as satisfactory proof of spamming activity rather disturbing.
>

        Hold on here - there are two different pages/forms involved.
One is for a contest - a copy still exists at:

http://www.winterizewithscotts.com/index.tbapp?page=intro

The other is a rebate form, with a copy at:

http://www.winterizewithscotts.com/index.tbapp?page=rebate_page

To get to the rebate form navigating on Scotts' own pages you must go through
a page which contains a link to a privacy policy which states that you do
consent to receive promotional materials.  But the rebate form itself does
not say any such thing and if a third party (Lowes, Home Depot, etc.) just
provided a link to the rebate form, or a consumer used a "rebate site",
they would never see the policy.  In all likelyhood the "promotional material"
is simply the mailings which you already get by choice, so you've never seen
anything "extra".

        If you look, you'll see as I did that the contest page *does* have
a link to the "privacy policy", and as such I consider any one who signed
up to have made "informed consent" (even if they are Joe Average, and didn't
know that was what they were doing).

        So once again, I'll say:  Based on the REBATE form, I believe it
is more likely than not that at least some consumers did receive unsolicited
commercial email from Scotts - i.e. spam.

--------------------------------------------------------------------------------

        Now lets leave all the arguing about specific cases behind, because
I believe you have a point regarding the interactions of the various RBLs
used in SA, not just for URIs but also for DNS_* and RCVD_* rules.

        From my reading of the FAQ and the code used in the perceptron and
to create meta-rules, there is a strong bias against creating meta-rules
after the fact (i.e. after the perceptron run) and an extremely strong
bias against negaptive scoing meta-rules (a separate issue, but related to
the exponent used in weighting the construction of such rules).

        I also believe that a better result would occur with more URI RBLs
in use (currently the SBL is the only IP based RBL used for URIs, and hence
for NS lookups).  A naive construction of all possible meta-rules will
quickly expand exponentially a huge number cases with simple "and" clauses
(and counting clause meta-rules and other type of constructions would add
even more cases).  Also, the exist several RHS and IP URIs, which have a
much higher FP rate than either any of the SURBLS (ignoring the recent 7%
[ws] report) or URIBL, but which have proven very useful to me.  Some of
these include using the AHBL for URIs (out performs the RCVD rule for me
in both spam hit and with a lower FP rate), the RFCI lists as URI rules,
(armored clothing on) the SORBS spamtrap list, and the completewhois list;
The last two having the advantage of being IP based so they catch nameservers
on IPs just as the SBL does.

        The evaluation of meta-rules is *much* less expensive than that of
the primary rules (simple expressions only to evaluate, no additional DNS
lookups or server queries required);  If we consider the 5 SURBLs (not
counting XS), 2 URIBL lists (black and grey, ignoring red), 5 RFCI lists,
completewhois, SORBS' spamtrap list, the SBL and the RAZOR "CF"/e8 test(s),
we have at least 16 rules which take URIs into account; All ot the possible
meta-rules "and"'d cases would give us millions of meta-rules, far too many.
But by guessing which ones are reasonably related we could simply contruct
the 31 rules created by the set of RFCI lists (1 choose 5 plus 2 choose 5...)
the 323 rules constructed from the SURBL, URIBL and AHBL lists and the 7 rules
created by combinations of the SBL, completewhois and SORBS rules;  If they
were initially used as inputs to the perceptron run (instead of possibly being
derived after the fact by the the evolve_metarule code - which by my reading
the documentation would never find or even examine most of these cases), you
might indeed find that the cases for multiple hits would result in much lower
scores.  Also, very likely is that many of these combinations would not have
a high enough hit rate to warrant inclusion in the set of SA rules - So I am
not proposing adding 150 new rules, but merely testing that number and then
selecting the (expected) few dozen that actually have value.  By including them
in the original perceptron test inputs, we would immediately see if the old
individual URI rules were lowered in value and if the "new" meta-rules raised
those values all the way back up.

        Add in the counting cases and we (assuming a similar grouping) add
only a few dozen more test rules to these three sets of URI rule groups and
we would only have another 15 rules (5 + 7 + 3).  Using both "equal" and the
case for "greater than or equal" give 12 more cases (you can't be greater
than the maximum).  If we also form the possible meta-rules from the cross
product of all the different group "counting" rules, we get 1050 more rules
to test (10 * 15 * 7 - the zero cases should also be included).  All told now
we have ~1500 new rules to test, but only expect a few dozen to have useful
values.

        Also, performing a similar set of meta-rule constructions for the
DNS_* and RCVD_* rules would give a coule of thousand more test cases and most
likely result in another hundred or so rules of significant value.

        Also, contructing "cross" type rules after the fact in the evolve code
now becomes possible and maybe even likely (e.g. URI in > 2 RHS RBLs, and DNS_
in > 1 IP RBL and RCVD_'d in > 2 RBLs - or similar cases);  These cases might
also result is a few dozen more useful rules.

        Yet another class of potentially profitable meta-rules could be
constructed by the intersection of RCVD_* rules and digest rules (if
it comes from a DUL, the XBL, DSBL, the AHBL or SpamCop and is bulk...).

        Of course this means that the scoring runs will take far more time
and resources than they do today, but the actual run time of operating SA
would barely change at all.

        I strongly suspect this would lead overall to close to a few hundred
new meta-rules, but almost all (the meta-rules and the existing "direct" ones)
having a much lower score than most of the current existing net tests are
assigned *and* your concerns about overlapping FPs would be directly addressed.

        Clearly this is not hard - all the cases I have listed could easily
be generated exhaustively be a simple script (yes, I'm volunteering if any
developer who can run the tests on a large corpus asks).

        I strongly believe that this would result in a much higher number
of rules overall, with a much lower average score and a much lower FP rate.

        As always, net tests do almost nothing for those who get spam at the
start of a spam run, and the "classic" rules and the work of SARE continues
to be the only way to stop much of that, but adding URI tests for both SORBS
spamtraps and completewhois *will* stop some cases of new spam by catching
IPs which have previously been used, but are currently being spammed with new
domains or are using known "evil" nameserver IPs (as it is, only the SBL can
catch those cases in the default SA distribution today).

        Paul Shupak
        [EMAIL PROTECTED]

Reply via email to