On Tue, 18 Nov 2003, Larry Gilson wrote:
> Most of you guys can get over my head quickly.
Read, learn, and exceed us. :-)
> This sounds a lot like the squidGuard blacklist implementation. You start
> with a base text file - one each for domains, urls, and regex. It is up to
> each the a
> -Original Message-
> From: Chris Santerre
--snip--
> >
> > One of the big advantages of using a DB type system is that it
> > can be updated 'hot' on a running system. A system based upon
> > parsing a config file and creating an in-memory hash table would
> > requirerestarting sp
> > would be looked up as "www.stearns.org" or "stearns.org".)
>
> The parser in the Bayes routine (tokenize_line in Bayes.pm)
> creates 'UD:'
> lookup tokens for each component of the domain name. So for the above
> example, it would create:
> UD:www.stearns.org
> UD:stearns.org
>
On Mon, 17 Nov 2003, Justin Mason wrote:
> BTW, given that a URI DB cannot use regular expressions, or patterns,
> would this really be useful?
>
> Basically with a DB you only gain efficiency when looking up exact
> strings. So for this to be useful against URIs, you'd have to pick out
> *just*
Oooh I got goosebumps!
*snip out everything. it was all good*
I still think that that the best way is to use the AWL code and convert it
to auto-blacklist code. The experation stuff and DB stuff is already there.
We would just have to put in code to parse out the URIs. This way, the more
the spam
On Sat, 15 Nov 2003, Carl R. Friend wrote:
>On Sat, 15 Nov 2003, David B Funk wrote:
> >
> > I've been thinking about that exact topic. The Bayes engine
> > already parses and tokenizes hostnames from URIs (the UD: tokens).
> > If there were a hash DB made with the spam-site hostname as key an
On Sat, 15 Nov 2003, David B Funk wrote:
> On Fri, 14 Nov 2003, Carl R. Friend wrote:
>
> >For the assembled group -- is it possible to do a DB lookup,
> > either in an eval() or some other mechanism, in a "uri" rule?
> > If we could do a DB lookup on URIs (or, more properly, the
> > domai