2023-01-30 16:27:06 +03:00
// Package junk implements a bayesian spam filter.
//
// A message can be parsed into words. Words (or pairs or triplets) can be used
// to train the filter or to classify the message as ham or spam. Training
// records the words in the database as ham/spam. Classifying consists of
// calculating the ham/spam probability by combining the words in the message
// with their ham/spam status.
package junk
// todo: look at inverse chi-square function? see https://www.linuxjournal.com/article/6467
// todo: perhaps: whether anchor text in links in html are different from the url
import (
2023-05-22 15:40:36 +03:00
"context"
2023-01-30 16:27:06 +03:00
"errors"
"fmt"
"io"
2024-02-08 16:49:01 +03:00
"log/slog"
2023-01-30 16:27:06 +03:00
"math"
"os"
make mox compile on windows, without "mox serve" but with working "mox localserve"
getting mox to compile required changing code in only a few places where
package "syscall" was used: for accessing file access times and for umask
handling. an open problem is how to start a process as an unprivileged user on
windows. that's why "mox serve" isn't implemented yet. and just finding a way
to implement it now may not be good enough in the near future: we may want to
starting using a more complete privilege separation approach, with a process
handling sensitive tasks (handling private keys, authentication), where we may
want to pass file descriptors between processes. how would that work on
windows?
anyway, getting mox to compile for windows doesn't mean it works properly on
windows. the largest issue: mox would normally open a file, rename or remove
it, and finally close it. this happens during message delivery. that doesn't
work on windows, the rename/remove would fail because the file is still open.
so this commit swaps many "remove" and "close" calls. renames are a longer
story: message delivery had two ways to deliver: with "consuming" the
(temporary) message file (which would rename it to its final destination), and
without consuming (by hardlinking the file, falling back to copying). the last
delivery to a recipient of a message (and the only one in the common case of a
single recipient) would consume the message, and the earlier recipients would
not. during delivery, the already open message file was used, to parse the
message. we still want to use that open message file, and the caller now stays
responsible for closing it, but we no longer try to rename (consume) the file.
we always hardlink (or copy) during delivery (this works on windows), and the
caller is responsible for closing and removing (in that order) the original
temporary file. this does cost one syscall more. but it makes the delivery code
(responsibilities) a bit simpler.
there is one more obvious issue: the file system path separator. mox already
used the "filepath" package to join paths in many places, but not everywhere.
and it still used strings with slashes for local file access. with this commit,
the code now uses filepath.FromSlash for path strings with slashes, uses
"filepath" in a few more places where it previously didn't. also switches from
"filepath" to regular "path" package when handling mailbox names in a few
places, because those always use forward slashes, regardless of local file
system conventions. windows can handle forward slashes when opening files, so
test code that passes path strings with forward slashes straight to go stdlib
file i/o functions are left unchanged to reduce code churn. the regular
non-test code, or test code that uses path strings in places other than
standard i/o functions, does have the paths converted for consistent paths
(otherwise we would end up with paths with mixed forward/backward slashes in
log messages).
windows cannot dup a listening socket. for "mox localserve", it isn't
important, and we can work around the issue. the current approach for "mox
serve" (forking a process and passing file descriptors of listening sockets on
"privileged" ports) won't work on windows. perhaps it isn't needed on windows,
and any user can listen on "privileged" ports? that would be welcome.
on windows, os.Open cannot open a directory, so we cannot call Sync on it after
message delivery. a cursory internet search indicates that directories cannot
be synced on windows. the story is probably much more nuanced than that, with
long deep technical details/discussions/disagreement/confusion, like on unix.
for "mox localserve" we can get away with making syncdir a no-op.
2023-10-14 11:54:07 +03:00
"path/filepath"
2023-01-30 16:27:06 +03:00
"sort"
"time"
"github.com/mjl-/bstore"
"github.com/mjl-/mox/message"
"github.com/mjl-/mox/mlog"
)
var (
2023-07-24 14:55:36 +03:00
// errBadContentType = errors.New("bad content-type") // sure sign of spam, todo: use this error
errClosed = errors . New ( "filter is closed" )
2023-01-30 16:27:06 +03:00
)
type word struct {
Ham uint32
Spam uint32
}
type wordscore struct {
Word string
Ham uint32
Spam uint32
}
// Params holds parameters for the filter. Most are at test-time. The first are
// used during parsing and training.
type Params struct {
Onegrams bool ` sconf:"optional" sconf-doc:"Track ham/spam ranking for single words." `
Twograms bool ` sconf:"optional" sconf-doc:"Track ham/spam ranking for each two consecutive words." `
Threegrams bool ` sconf:"optional" sconf-doc:"Track ham/spam ranking for each three consecutive words." `
MaxPower float64 ` sconf-doc:"Maximum power a word (combination) can have. If spaminess is 0.99, and max power is 0.1, spaminess of the word will be set to 0.9. Similar for ham words." `
TopWords int ` sconf-doc:"Number of most spammy/hammy words to use for calculating probability. E.g. 10." `
IgnoreWords float64 ` sconf:"optional" sconf-doc:"Ignore words that are this much away from 0.5 haminess/spaminess. E.g. 0.1, causing word (combinations) of 0.4 to 0.6 to be ignored." `
RareWords int ` sconf:"optional" sconf-doc:"Occurrences in word database until a word is considered rare and its influence in calculating probability reduced. E.g. 1 or 2." `
}
add a "backup" subcommand to make consistent backups, and a "verifydata" subcommand to verify a backup before restoring, and add tests for future upgrades
the backup command will make consistent snapshots of all the database files. i
had been copying the db files before, and it usually works. but if the file is
modified during the backup, it is inconsistent and is likely to generate errors
when reading (can be at any moment in the future, when reading some db page).
"mox backup" opens the database file and writes out a copy in a transaction.
it also duplicates the message files.
before doing a restore, you could run "mox verifydata" on the to-be-restored
"data" directory. it check the database files, and compares the message files
with the database.
the new "gentestdata" subcommand generates a basic "data" directory, with a
queue and a few accounts. we will use it in the future along with "verifydata"
to test upgrades from old version to the latest version. both when going to the
next version, and when skipping several versions. the script test-upgrades.sh
executes these tests and doesn't do anything at the moment, because no releases
have this subcommand yet.
inspired by a failed upgrade attempt of a pre-release version.
2023-05-26 20:26:51 +03:00
var DBTypes = [ ] any { wordscore { } } // Stored in DB.
2023-01-30 16:27:06 +03:00
type Filter struct {
Params
2023-12-05 15:35:58 +03:00
log mlog . Log // For logging cid.
2023-01-30 16:27:06 +03:00
closed bool
modified bool // Whether any modifications are pending. Cleared by Save.
hams , spams uint32 // Message count, stored in db under word "-".
cache map [ string ] word // Words read from database or during training.
changed map [ string ] word // Words modified during training.
dbPath , bloomPath string
db * bstore . DB // Always open on a filter.
bloom * Bloom // Only opened when writing.
isNew bool // Set for new filters until their first sync to disk. For faster writing.
}
func ( f * Filter ) ensureBloom ( ) error {
if f . bloom != nil {
return nil
}
var err error
f . bloom , err = openBloom ( f . bloomPath )
return err
}
2023-02-16 11:57:27 +03:00
// CloseDiscard closes the filter, discarding any changes.
func ( f * Filter ) CloseDiscard ( ) error {
if f . closed {
return errClosed
}
err := f . db . Close ( )
* f = Filter { log : f . log , closed : true }
return err
}
2023-01-30 16:27:06 +03:00
// Close first saves the filter if it has modifications, then closes the database
// connection and releases the bloom filter.
func ( f * Filter ) Close ( ) error {
if f . closed {
return errClosed
}
var err error
if f . modified {
err = f . Save ( )
}
if err != nil {
f . db . Close ( )
} else {
err = f . db . Close ( )
}
* f = Filter { log : f . log , closed : true }
return err
}
2023-12-05 15:35:58 +03:00
func OpenFilter ( ctx context . Context , log mlog . Log , params Params , dbPath , bloomPath string , loadBloom bool ) ( * Filter , error ) {
2023-01-30 16:27:06 +03:00
var bloom * Bloom
if loadBloom {
var err error
bloom , err = openBloom ( bloomPath )
if err != nil {
return nil , err
}
} else if fi , err := os . Stat ( bloomPath ) ; err == nil {
if err := BloomValid ( int ( fi . Size ( ) ) , bloomK ) ; err != nil {
return nil , fmt . Errorf ( "bloom: %s" , err )
}
}
2024-05-10 15:44:37 +03:00
db , err := openDB ( ctx , log , dbPath )
2023-01-30 16:27:06 +03:00
if err != nil {
return nil , fmt . Errorf ( "open database: %s" , err )
}
f := & Filter {
Params : params ,
log : log ,
cache : map [ string ] word { } ,
changed : map [ string ] word { } ,
dbPath : dbPath ,
bloomPath : bloomPath ,
db : db ,
bloom : bloom ,
}
2023-05-22 15:40:36 +03:00
err = f . db . Read ( ctx , func ( tx * bstore . Tx ) error {
2023-01-30 16:27:06 +03:00
wc := wordscore { Word : "-" }
err := tx . Get ( & wc )
f . hams = wc . Ham
f . spams = wc . Spam
return err
} )
if err != nil {
2023-02-16 15:22:00 +03:00
cerr := f . Close ( )
log . Check ( cerr , "closing filter after error" )
2023-01-30 16:27:06 +03:00
return nil , fmt . Errorf ( "looking up ham/spam message count: %s" , err )
}
return f , nil
}
// NewFilter creates a new filter with empty bloom filter and database files. The
// filter is marked as new until the first save, will be done automatically if
// TrainDirs is called. If the bloom and/or database files exist, an error is
// returned.
2023-12-05 15:35:58 +03:00
func NewFilter ( ctx context . Context , log mlog . Log , params Params , dbPath , bloomPath string ) ( * Filter , error ) {
2023-01-30 16:27:06 +03:00
var err error
if _ , err := os . Stat ( bloomPath ) ; err == nil {
return nil , fmt . Errorf ( "bloom filter already exists on disk: %s" , bloomPath )
} else if _ , err := os . Stat ( dbPath ) ; err == nil {
return nil , fmt . Errorf ( "database file already exists on disk: %s" , dbPath )
}
bloomSizeBytes := 4 * 1024 * 1024
if err := BloomValid ( bloomSizeBytes , bloomK ) ; err != nil {
return nil , fmt . Errorf ( "bloom: %s" , err )
}
bf , err := os . Create ( bloomPath )
if err != nil {
return nil , fmt . Errorf ( "creating bloom file: %w" , err )
}
if err := bf . Truncate ( 4 * 1024 * 1024 ) ; err != nil {
2023-02-16 15:22:00 +03:00
xerr := bf . Close ( )
log . Check ( xerr , "closing bloom filter file after truncate error" )
xerr = os . Remove ( bloomPath )
log . Check ( xerr , "removing bloom filter file after truncate error" )
2023-01-30 16:27:06 +03:00
return nil , fmt . Errorf ( "making empty bloom filter: %s" , err )
}
2023-02-16 15:22:00 +03:00
err = bf . Close ( )
log . Check ( err , "closing bloomfilter file" )
2023-01-30 16:27:06 +03:00
2023-05-22 15:40:36 +03:00
db , err := newDB ( ctx , log , dbPath )
2023-01-30 16:27:06 +03:00
if err != nil {
2023-02-16 15:22:00 +03:00
xerr := os . Remove ( bloomPath )
log . Check ( xerr , "removing bloom filter file after db init error" )
xerr = os . Remove ( dbPath )
log . Check ( xerr , "removing database file after db init error" )
2023-01-30 16:27:06 +03:00
return nil , fmt . Errorf ( "open database: %s" , err )
}
words := map [ string ] word { } // f.changed is set to new map after training
f := & Filter {
Params : params ,
log : log ,
modified : true , // Ensure ham/spam message count is added for new filter.
cache : words ,
changed : words ,
dbPath : dbPath ,
bloomPath : bloomPath ,
db : db ,
isNew : true ,
}
return f , nil
}
const bloomK = 10
func openBloom ( path string ) ( * Bloom , error ) {
buf , err := os . ReadFile ( path )
if err != nil {
return nil , fmt . Errorf ( "reading bloom file: %w" , err )
}
return NewBloom ( buf , bloomK )
}
2023-12-05 15:35:58 +03:00
func newDB ( ctx context . Context , log mlog . Log , path string ) ( db * bstore . DB , rerr error ) {
2023-01-30 16:27:06 +03:00
// Remove any existing files.
os . Remove ( path )
defer func ( ) {
if rerr != nil {
2023-02-16 15:22:00 +03:00
err := os . Remove ( path )
log . Check ( err , "removing db file after init error" )
2023-01-30 16:27:06 +03:00
}
} ( )
2024-05-10 15:44:37 +03:00
opts := bstore . Options { Timeout : 5 * time . Second , Perm : 0660 , RegisterLogger : log . Logger }
db , err := bstore . Open ( ctx , path , & opts , DBTypes ... )
2023-01-30 16:27:06 +03:00
if err != nil {
return nil , fmt . Errorf ( "open new database: %w" , err )
}
return db , nil
}
2024-05-10 15:44:37 +03:00
func openDB ( ctx context . Context , log mlog . Log , path string ) ( * bstore . DB , error ) {
2023-01-30 16:27:06 +03:00
if _ , err := os . Stat ( path ) ; err != nil {
return nil , fmt . Errorf ( "stat db file: %w" , err )
}
2024-05-10 15:44:37 +03:00
opts := bstore . Options { Timeout : 5 * time . Second , Perm : 0660 , RegisterLogger : log . Logger }
return bstore . Open ( ctx , path , & opts , DBTypes ... )
2023-01-30 16:27:06 +03:00
}
// Save stores modifications, e.g. from training, to the database and bloom
// filter files.
func ( f * Filter ) Save ( ) error {
if f . closed {
return errClosed
}
if ! f . modified {
return nil
}
if f . bloom != nil && f . bloom . Modified ( ) {
if err := f . bloom . Write ( f . bloomPath ) ; err != nil {
return fmt . Errorf ( "writing bloom filter: %w" , err )
}
}
// We need to insert sequentially for reasonable performance.
words := make ( [ ] string , len ( f . changed ) )
i := 0
for w := range f . changed {
words [ i ] = w
i ++
}
sort . Slice ( words , func ( i , j int ) bool {
return words [ i ] < words [ j ]
} )
2023-12-05 15:35:58 +03:00
f . log . Debug ( "inserting words in junkfilter db" , slog . Any ( "words" , len ( f . changed ) ) )
2023-01-30 16:27:06 +03:00
// start := time.Now()
if f . isNew {
if err := f . db . HintAppend ( true , wordscore { } ) ; err != nil {
f . log . Errorx ( "hint appendonly" , err )
} else {
2023-02-16 15:22:00 +03:00
defer func ( ) {
err := f . db . HintAppend ( false , wordscore { } )
f . log . Check ( err , "restoring append hint" )
} ( )
2023-01-30 16:27:06 +03:00
}
}
2023-05-22 15:40:36 +03:00
err := f . db . Write ( context . Background ( ) , func ( tx * bstore . Tx ) error {
2023-01-30 16:27:06 +03:00
update := func ( w string , ham , spam uint32 ) error {
if f . isNew {
return tx . Insert ( & wordscore { w , ham , spam } )
}
wc := wordscore { w , 0 , 0 }
err := tx . Get ( & wc )
if err == bstore . ErrAbsent {
return tx . Insert ( & wordscore { w , ham , spam } )
} else if err != nil {
return err
}
return tx . Update ( & wordscore { w , wc . Ham + ham , wc . Spam + spam } )
}
if err := update ( "-" , f . hams , f . spams ) ; err != nil {
return fmt . Errorf ( "storing total ham/spam message count: %s" , err )
}
for _ , w := range words {
c := f . changed [ w ]
if err := update ( w , c . Ham , c . Spam ) ; err != nil {
return fmt . Errorf ( "updating ham/spam count: %s" , err )
}
}
return nil
} )
if err != nil {
return fmt . Errorf ( "updating database: %w" , err )
}
f . changed = map [ string ] word { }
f . modified = false
f . isNew = false
2023-12-05 15:35:58 +03:00
// f.log.Info("wrote filter to db", slog.Any("duration", time.Since(start)))
2023-01-30 16:27:06 +03:00
return nil
}
2023-05-22 15:40:36 +03:00
func loadWords ( ctx context . Context , db * bstore . DB , l [ ] string , dst map [ string ] word ) error {
2023-01-30 16:27:06 +03:00
sort . Slice ( l , func ( i , j int ) bool {
return l [ i ] < l [ j ]
} )
2023-05-22 15:40:36 +03:00
err := db . Read ( ctx , func ( tx * bstore . Tx ) error {
2023-01-30 16:27:06 +03:00
for _ , w := range l {
wc := wordscore { Word : w }
if err := tx . Get ( & wc ) ; err == nil {
dst [ w ] = word { wc . Ham , wc . Spam }
}
}
return nil
} )
if err != nil {
return fmt . Errorf ( "fetching words: %s" , err )
}
return nil
}
2024-10-04 17:01:30 +03:00
// WordScore is a word with its score as used in classifications, based on
// (historic) training.
type WordScore struct {
Word string
Score float64 // 0 is ham, 1 is spam.
}
2023-01-30 16:27:06 +03:00
// ClassifyWords returns the spam probability for the given words, and number of recognized ham and spam words.
2024-10-04 17:01:30 +03:00
func ( f * Filter ) ClassifyWords ( ctx context . Context , words map [ string ] struct { } ) ( probability float64 , hams , spams [ ] WordScore , rerr error ) {
2023-01-30 16:27:06 +03:00
if f . closed {
2024-10-04 17:01:30 +03:00
return 0 , nil , nil , errClosed
2023-01-30 16:27:06 +03:00
}
var hamHigh float64 = 0
var spamLow float64 = 1
2024-10-04 17:01:30 +03:00
var topHam [ ] WordScore
var topSpam [ ] WordScore
2023-01-30 16:27:06 +03:00
// Find words that should be in the database.
lookupWords := [ ] string { }
expect := map [ string ] struct { } { }
unknowns := map [ string ] struct { } { }
totalUnknown := 0
for w := range words {
if f . bloom != nil && ! f . bloom . Has ( w ) {
totalUnknown ++
if len ( unknowns ) < 50 {
unknowns [ w ] = struct { } { }
}
continue
}
if _ , ok := f . cache [ w ] ; ok {
continue
}
lookupWords = append ( lookupWords , w )
expect [ w ] = struct { } { }
}
if len ( unknowns ) > 0 {
2023-12-05 18:06:50 +03:00
f . log . Debug ( "unknown words in bloom filter, showing max 50" ,
slog . Any ( "words" , unknowns ) ,
slog . Any ( "totalunknown" , totalUnknown ) ,
slog . Any ( "totalwords" , len ( words ) ) )
2023-01-30 16:27:06 +03:00
}
// Fetch words from database.
fetched := map [ string ] word { }
if len ( lookupWords ) > 0 {
2023-05-22 15:40:36 +03:00
if err := loadWords ( ctx , f . db , lookupWords , fetched ) ; err != nil {
2024-10-04 17:01:30 +03:00
return 0 , nil , nil , err
2023-01-30 16:27:06 +03:00
}
for w , c := range fetched {
delete ( expect , w )
f . cache [ w ] = c
}
2023-12-05 18:06:50 +03:00
f . log . Debug ( "unknown words in db" ,
slog . Any ( "words" , expect ) ,
slog . Any ( "totalunknown" , len ( expect ) ) ,
slog . Any ( "totalwords" , len ( words ) ) )
2023-01-30 16:27:06 +03:00
}
for w := range words {
c , ok := f . cache [ w ]
if ! ok {
continue
}
var wS , wH float64
if f . spams > 0 {
wS = float64 ( c . Spam ) / float64 ( f . spams )
}
if f . hams > 0 {
wH = float64 ( c . Ham ) / float64 ( f . hams )
}
r := wS / ( wS + wH )
if r < f . MaxPower {
r = f . MaxPower
} else if r >= 1 - f . MaxPower {
r = 1 - f . MaxPower
}
if c . Ham + c . Spam <= uint32 ( f . RareWords ) {
// Reduce the power of rare words.
r += float64 ( 1 + uint32 ( f . RareWords ) - ( c . Ham + c . Spam ) ) * ( 0.5 - r ) / 10
}
if math . Abs ( 0.5 - r ) < f . IgnoreWords {
continue
}
if r < 0.5 {
if len ( topHam ) >= f . TopWords && r > hamHigh {
continue
}
2024-10-04 17:01:30 +03:00
topHam = append ( topHam , WordScore { w , r } )
2023-01-30 16:27:06 +03:00
if r > hamHigh {
hamHigh = r
}
} else if r > 0.5 {
if len ( topSpam ) >= f . TopWords && r < spamLow {
continue
}
2024-10-04 17:01:30 +03:00
topSpam = append ( topSpam , WordScore { w , r } )
2023-01-30 16:27:06 +03:00
if r < spamLow {
spamLow = r
}
}
}
sort . Slice ( topHam , func ( i , j int ) bool {
a , b := topHam [ i ] , topHam [ j ]
2024-10-04 17:01:30 +03:00
if a . Score == b . Score {
2023-01-30 16:27:06 +03:00
return len ( a . Word ) > len ( b . Word )
}
2024-10-04 17:01:30 +03:00
return a . Score < b . Score
2023-01-30 16:27:06 +03:00
} )
sort . Slice ( topSpam , func ( i , j int ) bool {
a , b := topSpam [ i ] , topSpam [ j ]
2024-10-04 17:01:30 +03:00
if a . Score == b . Score {
2023-01-30 16:27:06 +03:00
return len ( a . Word ) > len ( b . Word )
}
2024-10-04 17:01:30 +03:00
return a . Score > b . Score
2023-01-30 16:27:06 +03:00
} )
2024-10-04 17:01:30 +03:00
nham := f . TopWords
2023-01-30 16:27:06 +03:00
if nham > len ( topHam ) {
nham = len ( topHam )
}
2024-10-04 17:01:30 +03:00
nspam := f . TopWords
2023-01-30 16:27:06 +03:00
if nspam > len ( topSpam ) {
nspam = len ( topSpam )
}
topHam = topHam [ : nham ]
topSpam = topSpam [ : nspam ]
var eta float64
for _ , x := range topHam {
2024-10-04 17:01:30 +03:00
eta += math . Log ( 1 - x . Score ) - math . Log ( x . Score )
2023-01-30 16:27:06 +03:00
}
for _ , x := range topSpam {
2024-10-04 17:01:30 +03:00
eta += math . Log ( 1 - x . Score ) - math . Log ( x . Score )
2023-01-30 16:27:06 +03:00
}
2023-12-05 15:35:58 +03:00
f . log . Debug ( "top words" , slog . Any ( "hams" , topHam ) , slog . Any ( "spams" , topSpam ) )
2023-01-30 16:27:06 +03:00
prob := 1 / ( 1 + math . Pow ( math . E , eta ) )
2024-10-04 17:01:30 +03:00
return prob , topHam , topSpam , nil
2023-01-30 16:27:06 +03:00
}
// ClassifyMessagePath is a convenience wrapper for calling ClassifyMessage on a file.
2024-10-04 17:01:30 +03:00
func ( f * Filter ) ClassifyMessagePath ( ctx context . Context , path string ) ( probability float64 , words map [ string ] struct { } , hams , spams [ ] WordScore , rerr error ) {
2023-01-30 16:27:06 +03:00
if f . closed {
2024-10-04 17:01:30 +03:00
return 0 , nil , nil , nil , errClosed
2023-01-30 16:27:06 +03:00
}
mf , err := os . Open ( path )
if err != nil {
2024-10-04 17:01:30 +03:00
return 0 , nil , nil , nil , err
2023-01-30 16:27:06 +03:00
}
2023-02-16 15:22:00 +03:00
defer func ( ) {
err := mf . Close ( )
f . log . Check ( err , "closing file after classify" )
} ( )
2023-01-30 16:27:06 +03:00
fi , err := mf . Stat ( )
if err != nil {
2024-10-04 17:01:30 +03:00
return 0 , nil , nil , nil , err
2023-01-30 16:27:06 +03:00
}
2023-05-22 15:40:36 +03:00
return f . ClassifyMessageReader ( ctx , mf , fi . Size ( ) )
2023-01-30 16:27:06 +03:00
}
2024-10-04 17:01:30 +03:00
func ( f * Filter ) ClassifyMessageReader ( ctx context . Context , mf io . ReaderAt , size int64 ) ( probability float64 , words map [ string ] struct { } , hams , spams [ ] WordScore , rerr error ) {
2023-12-05 15:35:58 +03:00
m , err := message . EnsurePart ( f . log . Logger , false , mf , size )
2023-01-30 16:27:06 +03:00
if err != nil && errors . Is ( err , message . ErrBadContentType ) {
// Invalid content-type header is a sure sign of spam.
//f.log.Infox("parsing content", err)
2024-10-04 17:01:30 +03:00
return 1 , nil , nil , nil , nil
2023-01-30 16:27:06 +03:00
}
2023-05-22 15:40:36 +03:00
return f . ClassifyMessage ( ctx , m )
2023-01-30 16:27:06 +03:00
}
// ClassifyMessage parses the mail message in r and returns the spam probability
// (between 0 and 1), along with the tokenized words found in the message, and the
2024-10-04 17:01:30 +03:00
// ham and spam words and their scores used.
func ( f * Filter ) ClassifyMessage ( ctx context . Context , m message . Part ) ( probability float64 , words map [ string ] struct { } , hams , spams [ ] WordScore , rerr error ) {
2023-01-30 16:27:06 +03:00
var err error
words , err = f . ParseMessage ( m )
if err != nil {
2024-10-04 17:01:30 +03:00
return 0 , nil , nil , nil , err
2023-01-30 16:27:06 +03:00
}
2024-10-04 17:01:30 +03:00
probability , hams , spams , err = f . ClassifyWords ( ctx , words )
return probability , words , hams , spams , err
2023-01-30 16:27:06 +03:00
}
// Train adds the words of a single message to the filter.
2023-05-22 15:40:36 +03:00
func ( f * Filter ) Train ( ctx context . Context , ham bool , words map [ string ] struct { } ) error {
2023-01-30 16:27:06 +03:00
if err := f . ensureBloom ( ) ; err != nil {
return err
}
var lwords [ ] string
for w := range words {
if ! f . bloom . Has ( w ) {
f . bloom . Add ( w )
continue
}
if _ , ok := f . cache [ w ] ; ! ok {
lwords = append ( lwords , w )
}
}
2023-05-22 15:40:36 +03:00
if err := f . loadCache ( ctx , lwords ) ; err != nil {
2023-01-30 16:27:06 +03:00
return err
}
f . modified = true
if ham {
f . hams ++
} else {
f . spams ++
}
for w := range words {
c := f . cache [ w ]
if ham {
c . Ham ++
} else {
c . Spam ++
}
f . cache [ w ] = c
f . changed [ w ] = c
}
return nil
}
2023-05-22 15:40:36 +03:00
func ( f * Filter ) TrainMessage ( ctx context . Context , r io . ReaderAt , size int64 , ham bool ) error {
2023-12-05 15:35:58 +03:00
p , _ := message . EnsurePart ( f . log . Logger , false , r , size )
2023-01-30 16:27:06 +03:00
words , err := f . ParseMessage ( p )
if err != nil {
return fmt . Errorf ( "parsing mail contents: %v" , err )
}
2023-05-22 15:40:36 +03:00
return f . Train ( ctx , ham , words )
2023-01-30 16:27:06 +03:00
}
2023-05-22 15:40:36 +03:00
func ( f * Filter ) UntrainMessage ( ctx context . Context , r io . ReaderAt , size int64 , ham bool ) error {
2023-12-05 15:35:58 +03:00
p , _ := message . EnsurePart ( f . log . Logger , false , r , size )
2023-01-30 16:27:06 +03:00
words , err := f . ParseMessage ( p )
if err != nil {
return fmt . Errorf ( "parsing mail contents: %v" , err )
}
2023-05-22 15:40:36 +03:00
return f . Untrain ( ctx , ham , words )
2023-01-30 16:27:06 +03:00
}
2023-05-22 15:40:36 +03:00
func ( f * Filter ) loadCache ( ctx context . Context , lwords [ ] string ) error {
2023-01-30 16:27:06 +03:00
if len ( lwords ) == 0 {
return nil
}
2023-05-22 15:40:36 +03:00
return loadWords ( ctx , f . db , lwords , f . cache )
2023-01-30 16:27:06 +03:00
}
// Untrain adjusts the filter to undo a previous training of the words.
2023-05-22 15:40:36 +03:00
func ( f * Filter ) Untrain ( ctx context . Context , ham bool , words map [ string ] struct { } ) error {
2023-01-30 16:27:06 +03:00
if err := f . ensureBloom ( ) ; err != nil {
return err
}
// Lookup any words from the db that aren't in the cache and put them in the cache for modification.
var lwords [ ] string
for w := range words {
if _ , ok := f . cache [ w ] ; ! ok {
lwords = append ( lwords , w )
}
}
2023-05-22 15:40:36 +03:00
if err := f . loadCache ( ctx , lwords ) ; err != nil {
2023-01-30 16:27:06 +03:00
return err
}
// Modify the message count.
f . modified = true
if ham {
f . hams --
} else {
f . spams --
}
// Decrease the word counts.
for w := range words {
c , ok := f . cache [ w ]
if ! ok {
continue
}
if ham {
c . Ham --
} else {
c . Spam --
}
f . cache [ w ] = c
f . changed [ w ] = c
}
return nil
}
// TrainDir parses mail messages from files and trains the filter.
func ( f * Filter ) TrainDir ( dir string , files [ ] string , ham bool ) ( n , malformed uint32 , rerr error ) {
if f . closed {
return 0 , 0 , errClosed
}
if err := f . ensureBloom ( ) ; err != nil {
return 0 , 0 , err
}
for _ , name := range files {
make mox compile on windows, without "mox serve" but with working "mox localserve"
getting mox to compile required changing code in only a few places where
package "syscall" was used: for accessing file access times and for umask
handling. an open problem is how to start a process as an unprivileged user on
windows. that's why "mox serve" isn't implemented yet. and just finding a way
to implement it now may not be good enough in the near future: we may want to
starting using a more complete privilege separation approach, with a process
handling sensitive tasks (handling private keys, authentication), where we may
want to pass file descriptors between processes. how would that work on
windows?
anyway, getting mox to compile for windows doesn't mean it works properly on
windows. the largest issue: mox would normally open a file, rename or remove
it, and finally close it. this happens during message delivery. that doesn't
work on windows, the rename/remove would fail because the file is still open.
so this commit swaps many "remove" and "close" calls. renames are a longer
story: message delivery had two ways to deliver: with "consuming" the
(temporary) message file (which would rename it to its final destination), and
without consuming (by hardlinking the file, falling back to copying). the last
delivery to a recipient of a message (and the only one in the common case of a
single recipient) would consume the message, and the earlier recipients would
not. during delivery, the already open message file was used, to parse the
message. we still want to use that open message file, and the caller now stays
responsible for closing it, but we no longer try to rename (consume) the file.
we always hardlink (or copy) during delivery (this works on windows), and the
caller is responsible for closing and removing (in that order) the original
temporary file. this does cost one syscall more. but it makes the delivery code
(responsibilities) a bit simpler.
there is one more obvious issue: the file system path separator. mox already
used the "filepath" package to join paths in many places, but not everywhere.
and it still used strings with slashes for local file access. with this commit,
the code now uses filepath.FromSlash for path strings with slashes, uses
"filepath" in a few more places where it previously didn't. also switches from
"filepath" to regular "path" package when handling mailbox names in a few
places, because those always use forward slashes, regardless of local file
system conventions. windows can handle forward slashes when opening files, so
test code that passes path strings with forward slashes straight to go stdlib
file i/o functions are left unchanged to reduce code churn. the regular
non-test code, or test code that uses path strings in places other than
standard i/o functions, does have the paths converted for consistent paths
(otherwise we would end up with paths with mixed forward/backward slashes in
log messages).
windows cannot dup a listening socket. for "mox localserve", it isn't
important, and we can work around the issue. the current approach for "mox
serve" (forking a process and passing file descriptors of listening sockets on
"privileged" ports) won't work on windows. perhaps it isn't needed on windows,
and any user can listen on "privileged" ports? that would be welcome.
on windows, os.Open cannot open a directory, so we cannot call Sync on it after
message delivery. a cursory internet search indicates that directories cannot
be synced on windows. the story is probably much more nuanced than that, with
long deep technical details/discussions/disagreement/confusion, like on unix.
for "mox localserve" we can get away with making syncdir a no-op.
2023-10-14 11:54:07 +03:00
p := filepath . Join ( dir , name )
2023-01-30 16:27:06 +03:00
valid , words , err := f . tokenizeMail ( p )
if err != nil {
2023-12-05 15:35:58 +03:00
// f.log.Infox("tokenizing mail", err, slog.Any("path", p))
2023-01-30 16:27:06 +03:00
malformed ++
continue
}
if ! valid {
continue
}
n ++
for w := range words {
if ! f . bloom . Has ( w ) {
f . bloom . Add ( w )
continue
}
c := f . cache [ w ]
f . modified = true
if ham {
c . Ham ++
} else {
c . Spam ++
}
f . cache [ w ] = c
f . changed [ w ] = c
}
}
return
}
// TrainDirs trains and saves a filter with mail messages from different types
// of directories.
func ( f * Filter ) TrainDirs ( hamDir , sentDir , spamDir string , hamFiles , sentFiles , spamFiles [ ] string ) error {
if f . closed {
return errClosed
}
var err error
var start time . Time
var hamMalformed , sentMalformed , spamMalformed uint32
start = time . Now ( )
f . hams , hamMalformed , err = f . TrainDir ( hamDir , hamFiles , true )
if err != nil {
return err
}
tham := time . Since ( start )
var sent uint32
start = time . Now ( )
if sentDir != "" {
sent , sentMalformed , err = f . TrainDir ( sentDir , sentFiles , true )
if err != nil {
return err
}
}
tsent := time . Since ( start )
start = time . Now ( )
f . spams , spamMalformed , err = f . TrainDir ( spamDir , spamFiles , false )
if err != nil {
return err
}
tspam := time . Since ( start )
hams := f . hams
f . hams += sent
if err := f . Save ( ) ; err != nil {
return fmt . Errorf ( "saving filter: %s" , err )
}
dbSize := f . fileSize ( f . dbPath )
bloomSize := f . fileSize ( f . bloomPath )
2023-12-05 15:35:58 +03:00
f . log . Print ( "training done" ,
slog . Any ( "hams" , hams ) ,
slog . Any ( "hamtime" , tham ) ,
slog . Any ( "hammalformed" , hamMalformed ) ,
slog . Any ( "sent" , sent ) ,
slog . Any ( "senttime" , tsent ) ,
slog . Any ( "sentmalformed" , sentMalformed ) ,
slog . Any ( "spams" , f . spams ) ,
slog . Any ( "spamtime" , tspam ) ,
slog . Any ( "spammalformed" , spamMalformed ) ,
slog . Any ( "dbsize" , fmt . Sprintf ( "%.1fmb" , float64 ( dbSize ) / ( 1024 * 1024 ) ) ) ,
slog . Any ( "bloomsize" , fmt . Sprintf ( "%.1fmb" , float64 ( bloomSize ) / ( 1024 * 1024 ) ) ) ,
slog . Any ( "bloom1ratio" , fmt . Sprintf ( "%.4f" , float64 ( f . bloom . Ones ( ) ) / float64 ( len ( f . bloom . Bytes ( ) ) * 8 ) ) ) ,
)
2023-01-30 16:27:06 +03:00
return nil
}
func ( f * Filter ) fileSize ( p string ) int {
fi , err := os . Stat ( p )
if err != nil {
2023-12-05 15:35:58 +03:00
f . log . Infox ( "stat" , err , slog . Any ( "path" , p ) )
2023-01-30 16:27:06 +03:00
return 0
}
return int ( fi . Size ( ) )
}
add a "backup" subcommand to make consistent backups, and a "verifydata" subcommand to verify a backup before restoring, and add tests for future upgrades
the backup command will make consistent snapshots of all the database files. i
had been copying the db files before, and it usually works. but if the file is
modified during the backup, it is inconsistent and is likely to generate errors
when reading (can be at any moment in the future, when reading some db page).
"mox backup" opens the database file and writes out a copy in a transaction.
it also duplicates the message files.
before doing a restore, you could run "mox verifydata" on the to-be-restored
"data" directory. it check the database files, and compares the message files
with the database.
the new "gentestdata" subcommand generates a basic "data" directory, with a
queue and a few accounts. we will use it in the future along with "verifydata"
to test upgrades from old version to the latest version. both when going to the
next version, and when skipping several versions. the script test-upgrades.sh
executes these tests and doesn't do anything at the moment, because no releases
have this subcommand yet.
inspired by a failed upgrade attempt of a pre-release version.
2023-05-26 20:26:51 +03:00
// DB returns the database, for backups.
func ( f * Filter ) DB ( ) * bstore . DB {
return f . db
}