mirror of
https://github.com/octoleo/restic.git
synced 2024-12-12 14:17:57 +00:00
e5d597fd22
The new genericized LRU cache no longer needs to have the IDs separately allocated: name old time/op new time/op delta Add-8 494ns ± 2% 388ns ± 2% -21.46% (p=0.000 n=10+9) name old alloc/op new alloc/op delta Add-8 176B ± 0% 152B ± 0% -13.64% (p=0.000 n=10+10) name old allocs/op new allocs/op delta Add-8 5.00 ± 0% 3.00 ± 0% -40.00% (p=0.000 n=10+10)
92 lines
2.0 KiB
Go
92 lines
2.0 KiB
Go
package bloblru
|
|
|
|
import (
|
|
"sync"
|
|
|
|
"github.com/restic/restic/internal/debug"
|
|
"github.com/restic/restic/internal/restic"
|
|
|
|
"github.com/hashicorp/golang-lru/v2/simplelru"
|
|
)
|
|
|
|
// Crude estimate of the overhead per blob: a SHA-256, a linked list node
|
|
// and some pointers. See comment in Cache.add.
|
|
const overhead = len(restic.ID{}) + 64
|
|
|
|
// A Cache is a fixed-size LRU cache of blob contents.
|
|
// It is safe for concurrent access.
|
|
type Cache struct {
|
|
mu sync.Mutex
|
|
c *simplelru.LRU[restic.ID, []byte]
|
|
|
|
free, size int // Current and max capacity, in bytes.
|
|
}
|
|
|
|
// New constructs a blob cache that stores at most size bytes worth of blobs.
|
|
func New(size int) *Cache {
|
|
c := &Cache{
|
|
free: size,
|
|
size: size,
|
|
}
|
|
|
|
// NewLRU wants us to specify some max. number of entries, else it errors.
|
|
// The actual maximum will be smaller than size/overhead, because we
|
|
// evict entries (RemoveOldest in add) to maintain our size bound.
|
|
maxEntries := size / overhead
|
|
lru, err := simplelru.NewLRU[restic.ID, []byte](maxEntries, c.evict)
|
|
if err != nil {
|
|
panic(err) // Can only be maxEntries <= 0.
|
|
}
|
|
c.c = lru
|
|
|
|
return c
|
|
}
|
|
|
|
// Add adds key id with value blob to c.
|
|
// It may return an evicted buffer for reuse.
|
|
func (c *Cache) Add(id restic.ID, blob []byte) (old []byte) {
|
|
debug.Log("bloblru.Cache: add %v", id)
|
|
|
|
size := cap(blob) + overhead
|
|
if size > c.size {
|
|
return
|
|
}
|
|
|
|
c.mu.Lock()
|
|
defer c.mu.Unlock()
|
|
|
|
if c.c.Contains(id) { // Doesn't update the recency list.
|
|
return
|
|
}
|
|
|
|
// This loop takes at most min(maxEntries, maxchunksize/overhead)
|
|
// iterations.
|
|
for size > c.free {
|
|
_, b, _ := c.c.RemoveOldest()
|
|
if cap(b) > cap(old) {
|
|
// We can only return one buffer, so pick the largest.
|
|
old = b
|
|
}
|
|
}
|
|
|
|
c.c.Add(id, blob)
|
|
c.free -= size
|
|
|
|
return old
|
|
}
|
|
|
|
func (c *Cache) Get(id restic.ID) ([]byte, bool) {
|
|
c.mu.Lock()
|
|
blob, ok := c.c.Get(id)
|
|
c.mu.Unlock()
|
|
|
|
debug.Log("bloblru.Cache: get %v, hit %v", id, ok)
|
|
|
|
return blob, ok
|
|
}
|
|
|
|
func (c *Cache) evict(key restic.ID, blob []byte) {
|
|
debug.Log("bloblru.Cache: evict %v, %d bytes", key, cap(blob))
|
|
c.free += cap(blob) + overhead
|
|
}
|