Hi,

Same problem here with a 1.3G bayes_seen file.

No CPU load linked to this but a too big file is never good...

Can someone help to deal with this ? As long as I remember this problem =
were
discussed a lot of time here but I never saw a trick for this

-----Message d'origine-----
De : Richard Smits [mailto:R.Smits@tudelft.nl]=20
Envoy=E9 : mardi 12 juin 2007 09:30
=C0 : users@spamassassin.apache.org
Objet : How to decrease the bayes database size

Hello,

We realy need some help here. It has come to our attention that our =
bayes
database is 2.4 GB big. It is really slowing down our servers and they =
have
a big cpu load.

Now we have tried the trick with the sa-learn --force-expire , and it
deletes a lot of entrys, but the file is not getting any smaller.

79K Jun 12 09:26 bayes_journal
20M Jun 12 09:26 bayes_toks
2.5G Jun 12 09:26 bayes_seen*

Does anyone has some tricks to help us out ?

Greetings... Richard Smits

----
0.000 0 3 0 non-token data: bayes db version
0.000 0 14201082 0 non-token data: nspam
0.000 0 7760360 0 non-token data: nham
0.000 0 916962 0 non-token data: ntokens
0.000 0 1181559955 0 non-token data: oldest atime
0.000 0 1181633069 0 non-token data: newest atime
0.000 0 1181633115 0 non-token data: last journal=20
sync atime
0.000 0 1181604237 0 non-token data: last expiry =
atime
0.000 0 43200 0 non-token data: last expire=20
atime delta
0.000 0 360013 0 non-token data: last expire=20
reduction count

----------------------