2023-08-21 22:52:35 +03:00
package http
import (
"compress/gzip"
"encoding/base64"
"errors"
"fmt"
"io"
"io/fs"
"net/http"
"os"
"path/filepath"
"sort"
"strconv"
"strings"
"sync"
"time"
"github.com/mjl-/mox/mlog"
)
// todo: consider caching gzipped responses from forward handlers too. we would need to read the responses (handle up to perhaps 2mb), hash the data (blake2b seems fast), check if we have the gzip content for that hash, cache it on second request. keep around entries for non-yet-cached hashes, with some limit and lru eviction policy. we have to recognize some content-types as not applicable and do direct streaming compression, e.g. for text/event-stream. and we need to detect when backend server could be slowly sending out data and abort the caching attempt. downside is always that we need to read the whole response before and hash it before we can send our response. it is best if the backend just responds with gzip itself though. compression needs more cpu than hashing (at least 10x), but it's only worth it with enough hits.
// Cache for gzipped static files.
var staticgzcache gzcache
type gzcache struct {
dir string // Where all files are stored.
// Max total size of combined files in cache. When adding a new entry, the least
// recently used entries are evicted to stay below this size.
maxSize int64
sync . Mutex
// Total on-disk size of compressed data. Not larger than maxSize. We can
// temporarily have more bytes in use because while/after evicting, a writer may
// still have the old removed file open.
size int64
// Indexed by effective path, based on handler.
paths map [ string ] gzfile
// Only with files we completed compressing, kept ordered by atime. We evict from
// oldest. On use, we take entries out and put them at newest.
oldest , newest * pathUse
}
type gzfile struct {
// Whether compressing in progress. If a new request comes in while we are already
// compressing, for simplicity of code we just compress again for that client.
compressing bool
mtime int64 // If mtime changes, we remove entry from cache.
atime int64 // For LRU.
gzsize int64 // Compressed size, used in Content-Length header.
use * pathUse // Only set after compressing finished.
}
type pathUse struct {
prev , next * pathUse // Double-linked list.
path string
}
// Initialize staticgzcache from on-disk directory.
// The path and mtime are in the filename, the atime is in the file itself.
func loadStaticGzipCache ( dir string , maxSize int64 ) {
staticgzcache = gzcache {
dir : dir ,
maxSize : maxSize ,
paths : map [ string ] gzfile { } ,
}
// todo future: should we split cached files in sub directories, so we don't end up with one huge directory?
os . MkdirAll ( dir , 0700 )
entries , err := os . ReadDir ( dir )
if err != nil && ! os . IsNotExist ( err ) {
xlog . Errorx ( "listing static gzip cache files" , err , mlog . Field ( "dir" , dir ) )
}
for _ , e := range entries {
name := e . Name ( )
var err error
if ! strings . HasSuffix ( name , ".gz" ) {
err = errors . New ( "missing .gz suffix" )
}
var path , xpath , mtimestr string
if err == nil {
var ok bool
xpath , mtimestr , ok = strings . Cut ( strings . TrimRight ( name , ".gz" ) , "+" )
if ! ok {
err = fmt . Errorf ( "missing + in filename" )
}
}
if err == nil {
var pathbuf [ ] byte
pathbuf , err = base64 . RawURLEncoding . DecodeString ( xpath )
if err == nil {
path = string ( pathbuf )
}
}
var mtime int64
if err == nil {
mtime , err = strconv . ParseInt ( mtimestr , 16 , 64 )
}
var fi fs . FileInfo
if err == nil {
fi , err = e . Info ( )
}
var atime int64
if err == nil {
make mox compile on windows, without "mox serve" but with working "mox localserve"
getting mox to compile required changing code in only a few places where
package "syscall" was used: for accessing file access times and for umask
handling. an open problem is how to start a process as an unprivileged user on
windows. that's why "mox serve" isn't implemented yet. and just finding a way
to implement it now may not be good enough in the near future: we may want to
starting using a more complete privilege separation approach, with a process
handling sensitive tasks (handling private keys, authentication), where we may
want to pass file descriptors between processes. how would that work on
windows?
anyway, getting mox to compile for windows doesn't mean it works properly on
windows. the largest issue: mox would normally open a file, rename or remove
it, and finally close it. this happens during message delivery. that doesn't
work on windows, the rename/remove would fail because the file is still open.
so this commit swaps many "remove" and "close" calls. renames are a longer
story: message delivery had two ways to deliver: with "consuming" the
(temporary) message file (which would rename it to its final destination), and
without consuming (by hardlinking the file, falling back to copying). the last
delivery to a recipient of a message (and the only one in the common case of a
single recipient) would consume the message, and the earlier recipients would
not. during delivery, the already open message file was used, to parse the
message. we still want to use that open message file, and the caller now stays
responsible for closing it, but we no longer try to rename (consume) the file.
we always hardlink (or copy) during delivery (this works on windows), and the
caller is responsible for closing and removing (in that order) the original
temporary file. this does cost one syscall more. but it makes the delivery code
(responsibilities) a bit simpler.
there is one more obvious issue: the file system path separator. mox already
used the "filepath" package to join paths in many places, but not everywhere.
and it still used strings with slashes for local file access. with this commit,
the code now uses filepath.FromSlash for path strings with slashes, uses
"filepath" in a few more places where it previously didn't. also switches from
"filepath" to regular "path" package when handling mailbox names in a few
places, because those always use forward slashes, regardless of local file
system conventions. windows can handle forward slashes when opening files, so
test code that passes path strings with forward slashes straight to go stdlib
file i/o functions are left unchanged to reduce code churn. the regular
non-test code, or test code that uses path strings in places other than
standard i/o functions, does have the paths converted for consistent paths
(otherwise we would end up with paths with mixed forward/backward slashes in
log messages).
windows cannot dup a listening socket. for "mox localserve", it isn't
important, and we can work around the issue. the current approach for "mox
serve" (forking a process and passing file descriptors of listening sockets on
"privileged" ports) won't work on windows. perhaps it isn't needed on windows,
and any user can listen on "privileged" ports? that would be welcome.
on windows, os.Open cannot open a directory, so we cannot call Sync on it after
message delivery. a cursory internet search indicates that directories cannot
be synced on windows. the story is probably much more nuanced than that, with
long deep technical details/discussions/disagreement/confusion, like on unix.
for "mox localserve" we can get away with making syncdir a no-op.
2023-10-14 11:54:07 +03:00
atime , err = statAtime ( fi . Sys ( ) )
2023-08-21 22:52:35 +03:00
}
if err != nil {
xlog . Infox ( "removing unusable/unrecognized file in static gzip cache dir" , err )
xerr := os . Remove ( filepath . Join ( dir , name ) )
xlog . Check ( xerr , "removing unusable file in static gzip cache dir" , mlog . Field ( "error" , err ) , mlog . Field ( "dir" , dir ) , mlog . Field ( "filename" , name ) )
continue
}
staticgzcache . paths [ path ] = gzfile {
mtime : mtime ,
atime : atime ,
gzsize : fi . Size ( ) ,
use : & pathUse { path : path } ,
}
staticgzcache . size += fi . Size ( )
}
pathatimes := make ( [ ] struct {
path string
atime int64
} , len ( staticgzcache . paths ) )
i := 0
for k , gf := range staticgzcache . paths {
pathatimes [ i ] . path = k
pathatimes [ i ] . atime = gf . atime
i ++
}
sort . Slice ( pathatimes , func ( i , j int ) bool {
return pathatimes [ i ] . atime < pathatimes [ j ] . atime
} )
for _ , pa := range pathatimes {
staticgzcache . push ( staticgzcache . paths [ pa . path ] . use )
}
// Ensure cache size is OK for current config.
staticgzcache . evictFor ( 0 )
}
// Evict entries so size bytes are available.
// Must be called with lock held.
func ( c * gzcache ) evictFor ( size int64 ) {
for c . size + size > c . maxSize && c . oldest != nil {
c . evictPath ( c . oldest . path )
}
}
// remove path from cache.
// Must be called with lock held.
func ( c * gzcache ) evictPath ( path string ) {
gf := c . paths [ path ]
delete ( c . paths , path )
c . unlink ( gf . use )
c . size -= gf . gzsize
err := os . Remove ( staticCachePath ( c . dir , path , gf . mtime ) )
xlog . Check ( err , "removing cached gzipped static file" , mlog . Field ( "path" , path ) )
}
// Open cached file for path, requiring it has mtime. If there is no usable cached
// file, a nil file is returned and the caller should compress and add to the cache
// with startPath and finishPath. No usable cached file means the path isn't in the
// cache, or its mtime is different, or there is an entry but it is new and being
// compressed at the moment. If a usable cached file was found, it is opened and
// returned, along with its compressed/on-disk size.
func ( c * gzcache ) openPath ( path string , mtime int64 ) ( * os . File , int64 ) {
c . Lock ( )
defer c . Unlock ( )
gf , ok := c . paths [ path ]
if ! ok || gf . compressing {
return nil , 0
}
if gf . mtime != mtime {
// File has changed, remove old entry. Caller will add to cache again.
c . evictPath ( path )
return nil , 0
}
p := staticCachePath ( c . dir , path , gf . mtime )
f , err := os . Open ( p )
if err != nil {
xlog . Errorx ( "open static cached gzip file, removing from cache" , err , mlog . Field ( "path" , path ) )
// Perhaps someone removed the file? Remove from cache, it will be recreated.
c . evictPath ( path )
return nil , 0
}
gf . atime = time . Now ( ) . UnixNano ( )
c . unlink ( gf . use )
c . push ( gf . use )
c . paths [ path ] = gf
return f , gf . gzsize
}
// startPath attempts to add an entry to the cache for a new cached compressed
// file. If there is already an entry but it isn't done compressing yet, false is
// returned and the caller can still compress and respond but the entry cannot be
// added to the cache. If the entry is being added, the caller must call finishPath
// or abortPath.
func ( c * gzcache ) startPath ( path string , mtime int64 ) bool {
c . Lock ( )
defer c . Unlock ( )
if _ , ok := c . paths [ path ] ; ok {
return false
}
// note: no "use" yet, we only set that when we finish, so we don't have to clean up on abort.
c . paths [ path ] = gzfile { compressing : true , mtime : mtime }
return true
}
// finishPath completes adding an entry to the cache, marking the entry as
// compressed, accounting for its size, and marking its atime.
func ( c * gzcache ) finishPath ( path string , gzsize int64 ) {
c . Lock ( )
defer c . Unlock ( )
c . evictFor ( gzsize )
gf := c . paths [ path ]
gf . compressing = false
gf . gzsize = gzsize
gf . atime = time . Now ( ) . UnixNano ( )
gf . use = & pathUse { path : path }
c . paths [ path ] = gf
c . size += gzsize
c . push ( gf . use )
}
// abortPath marks an entry as no longer being added to the cache.
func ( c * gzcache ) abortPath ( path string ) {
c . Lock ( )
defer c . Unlock ( )
delete ( c . paths , path )
// note: gzfile.use isn't set yet.
}
// push inserts the "pathUse" to the head of the LRU doubly-linked list, unlinking
// it first if needed.
func ( c * gzcache ) push ( u * pathUse ) {
c . unlink ( u )
u . prev = c . newest
if c . newest != nil {
c . newest . next = u
}
if c . oldest == nil {
c . oldest = u
}
c . newest = u
}
// unlink removes the "pathUse" from the LRU doubly-linked list.
func ( c * gzcache ) unlink ( u * pathUse ) {
if c . oldest == u {
c . oldest = u . next
}
if c . newest == u {
c . newest = u . prev
}
if u . prev != nil {
u . prev . next = u . next
}
if u . next != nil {
u . next . prev = u . prev
}
u . prev = nil
u . next = nil
}
// Return path to the on-disk gzipped cached file.
func staticCachePath ( dir , path string , mtime int64 ) string {
p := base64 . RawURLEncoding . EncodeToString ( [ ] byte ( path ) )
return filepath . Join ( dir , fmt . Sprintf ( "%s+%x.gz" , p , mtime ) )
}
// staticgzcacheReplacer intercepts responses for cacheable static files,
// responding with the cached content if appropriate and failing further writes so
// the regular response writer stops.
type staticgzcacheReplacer struct {
w http . ResponseWriter
r * http . Request // For its context, or logging.
uncomprPath string
uncomprFile * os . File
uncomprMtime time . Time
uncomprSize int64
statusCode int
// Set during WriteHeader to indicate a compressed file has been written, further
// Writes result in an error to stop the writer of the uncompressed content.
handled bool
}
func ( w * staticgzcacheReplacer ) logger ( ) * mlog . Log {
return xlog . WithContext ( w . r . Context ( ) )
}
// Header returns the header of the underlying ResponseWriter.
func ( w * staticgzcacheReplacer ) Header ( ) http . Header {
return w . w . Header ( )
}
2023-08-23 15:59:43 +03:00
// WriteHeader checks whether the response is eligible for compressing. If not,
2023-08-21 22:52:35 +03:00
// WriteHeader on the underlying ResponseWriter is called. If so, headers for gzip
// content are set and the gzip content is written, either from disk or compressed
// and stored in the cache.
func ( w * staticgzcacheReplacer ) WriteHeader ( statusCode int ) {
if w . statusCode != 0 {
return
}
w . statusCode = statusCode
if statusCode != http . StatusOK {
w . w . WriteHeader ( statusCode )
return
}
gzf , gzsize := staticgzcache . openPath ( w . uncomprPath , w . uncomprMtime . UnixNano ( ) )
if gzf == nil {
// Not in cache, or work in progress.
started := staticgzcache . startPath ( w . uncomprPath , w . uncomprMtime . UnixNano ( ) )
if ! started {
// Another request is already compressing and storing this file.
// todo: we should just wait for the other compression to finish, then use its result.
w . w . ( * loggingWriter ) . UncompressedSize = w . uncomprSize
h := w . w . Header ( )
h . Set ( "Content-Encoding" , "gzip" )
h . Del ( "Content-Length" ) // We don't know this, we compress streamingly.
gzw , _ := gzip . NewWriterLevel ( w . w , gzip . BestSpeed )
_ , err := io . Copy ( gzw , w . uncomprFile )
if err == nil {
err = gzw . Close ( )
}
w . handled = true
if err != nil {
w . w . ( * loggingWriter ) . error ( err )
}
return
}
// Compress and write to cache.
p := staticCachePath ( staticgzcache . dir , w . uncomprPath , w . uncomprMtime . UnixNano ( ) )
ngzf , err := os . OpenFile ( p , os . O_CREATE | os . O_EXCL | os . O_RDWR , 0600 )
if err != nil {
w . logger ( ) . Errorx ( "create new static gzip cache file" , err , mlog . Field ( "requestpath" , w . uncomprPath ) , mlog . Field ( "fspath" , p ) )
staticgzcache . abortPath ( w . uncomprPath )
return
}
defer func ( ) {
if ngzf != nil {
staticgzcache . abortPath ( w . uncomprPath )
err := ngzf . Close ( )
w . logger ( ) . Check ( err , "closing failed static gzip cache file" , mlog . Field ( "requestpath" , w . uncomprPath ) , mlog . Field ( "fspath" , p ) )
err = os . Remove ( p )
w . logger ( ) . Check ( err , "removing failed static gzip cache file" , mlog . Field ( "requestpath" , w . uncomprPath ) , mlog . Field ( "fspath" , p ) )
}
} ( )
gzw := gzip . NewWriter ( ngzf )
_ , err = io . Copy ( gzw , w . uncomprFile )
if err == nil {
err = gzw . Close ( )
}
if err == nil {
err = ngzf . Sync ( )
}
if err == nil {
gzsize , err = ngzf . Seek ( 0 , 1 )
}
if err == nil {
_ , err = ngzf . Seek ( 0 , 0 )
}
if err != nil {
w . w . ( * loggingWriter ) . error ( err )
return
}
staticgzcache . finishPath ( w . uncomprPath , gzsize )
gzf = ngzf
ngzf = nil
}
defer func ( ) {
if gzf != nil {
err := gzf . Close ( )
if err != nil {
w . logger ( ) . Errorx ( "closing static gzip cache file" , err )
}
}
} ( )
// Signal to Write that we aleady (attempted to) write the responses.
w . handled = true
w . w . ( * loggingWriter ) . UncompressedSize = w . uncomprSize
h := w . w . Header ( )
h . Set ( "Content-Encoding" , "gzip" )
h . Set ( "Content-Length" , fmt . Sprintf ( "%d" , gzsize ) )
w . w . WriteHeader ( statusCode )
if _ , err := io . Copy ( w . w , gzf ) ; err != nil {
w . w . ( * loggingWriter ) . error ( err )
}
}
var errHandledCompressed = errors . New ( "response written with compression" )
func ( w * staticgzcacheReplacer ) Write ( buf [ ] byte ) ( int , error ) {
if w . statusCode == 0 {
w . WriteHeader ( http . StatusOK )
}
if w . handled {
// For 200 OK, we already wrote the response and just want the caller to stop processing.
return 0 , errHandledCompressed
}
return w . w . Write ( buf )
}