Arthur Dent wrote:
Learned tokens from 8 message(s) (3165 message(s) examined)
Learned tokens from 4628 message(s) (8703 message(s) examined)
Learned tokens from 3890 message(s) (8634 message(s) examined)
Learned tokens from 2264 message(s) (8671 message(s) examined)
Learned tokens from 2303 message(s) (8620 message(s) examined)
"Odds 2,000,127 against one... and counting..."
Notice that although the amount of tokens being learned seems to be coming
down gradually, the total far exceeds the total amount of ham mails in the
corpus.
The number of *messages* learned is decreasing, not the number of tokens.
Could it be that something deletes the temp folder before sa-learn has
finished, so it gets distracted and starts flying away carrying a suitcase?
Or do you receive >8600 messages each day? Some of them might have been
autolearned on the incoming SMTP channel, BTW.
IMHO it is not necessary to train so extensively the Bayes DB. If you want the
process to complete in a decent amount of time, feed it fewer messages at a time.
Paolo
PS: who knows who "Arthud Dent" is/was, will understand the oddities in this
reply. All others: get a copy of the HHGTTG. :-)