Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
112 changes: 112 additions & 0 deletions pkg/odp/lru_cache.go
Original file line number Diff line number Diff line change
@@ -0,0 +1,112 @@
/****************************************************************************
* Copyright 2022, Optimizely, Inc. and contributors *
* *
* Licensed under the Apache License, Version 2.0 (the "License"); *
* you may not use this file except in compliance with the License. *
* You may obtain a copy of the License at *
* *
* http://www.apache.org/licenses/LICENSE-2.0 *
* *
* Unless required by applicable law or agreed to in writing, software *
* distributed under the License is distributed on an "AS IS" BASIS, *
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. *
* See the License for the specific language governing permissions and *
* limitations under the License. *
***************************************************************************/

// Package odp //
package odp

import (
"container/list"
"sync"
"time"
)

// Cache is used for caching ODP segments
type Cache interface {
Save(key string, value interface{})
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

can we add specific type of interface instead of interface{}

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

interface has no further types in go.

Lookup(key string) interface{}
Reset()
}

type cacheElement struct {
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Cache ?

data interface{}
time int64
keyPtr *list.Element
}

// LRUCache a Least Recently Used in-memory cache
type LRUCache struct {
queue *list.List
items map[string]*cacheElement
maxSize int
timeoutInSecs int64
lock *sync.RWMutex
}

// NewLRUCache returns a new instance of Least Recently Used in-memory cache
func NewLRUCache(size int, timeoutInSecs int64) LRUCache {
Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

int64 to int? Range will be less than ~10000

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

To avoid converting timeoutInSecs to in64 every time isValid is called since time.Now().Unix() returns a int64 value and calculations are only allowed between values of similar type's.

return LRUCache{queue: list.New(), items: make(map[string]*cacheElement), maxSize: size, timeoutInSecs: timeoutInSecs, lock: new(sync.RWMutex)}
}

// Save stores a new element into the cache
func (l *LRUCache) Save(key string, value interface{}) {
if l.maxSize <= 0 {
return
}
l.lock.Lock()
defer l.lock.Unlock()
if item, ok := l.items[key]; !ok {
// remove the last object if queue is full
if l.maxSize == len(l.items) {
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

add comment. what this logic does.

back := l.queue.Back()
l.queue.Remove(back)
delete(l.items, back.Value.(string))
}
// push the new object to the front of the queue
l.items[key] = &cacheElement{data: value, keyPtr: l.queue.PushFront(key), time: time.Now().Unix()}
} else {
item.data = value
l.items[key] = item
l.queue.MoveToFront(item.keyPtr)
}
}

// Lookup retrieves an element from the cache, reordering the elements
func (l *LRUCache) Lookup(key string) interface{} {
if l.maxSize <= 0 {
return nil
}
l.lock.Lock()
defer l.lock.Unlock()
if item, ok := l.items[key]; ok {
if l.isValid(item) {
l.queue.MoveToFront(item.keyPtr)
return item.data
}
l.queue.Remove(item.keyPtr)
delete(l.items, item.keyPtr.Value.(string))
}
return nil
}

// Reset clears all the elements from the cache
func (l *LRUCache) Reset() {
if l.maxSize <= 0 {
return
}
l.lock.Lock()
defer l.lock.Unlock()
l.queue = list.New()
l.items = make(map[string]*cacheElement)
}

func (l *LRUCache) isValid(e *cacheElement) bool {
if l.timeoutInSecs <= 0 {
return true
}
currenttime := time.Now().Unix()
elapsedtime := currenttime - e.time
return l.timeoutInSecs > elapsedtime
}
189 changes: 189 additions & 0 deletions pkg/odp/lru_cache_test.go
Original file line number Diff line number Diff line change
@@ -0,0 +1,189 @@
package odp

import (
"fmt"
"sync"
"testing"
"time"

"github.com/stretchr/testify/assert"
)

func TestMinConfig(t *testing.T) {
cache := NewLRUCache(1000, 2000)
assert.Equal(t, 1000, cache.maxSize)
assert.Equal(t, int64(2000), cache.timeoutInSecs)

cache = NewLRUCache(0, 0)
assert.Equal(t, 0, cache.maxSize)
assert.Equal(t, int64(0), cache.timeoutInSecs)
}

func TestSaveAndLookupConfig(t *testing.T) {
maxSize := 2
cache := NewLRUCache(maxSize, 1000)

cache.Save("1", 100) // [1]
cache.Save("2", 200) // [1, 2]
cache.Save("3", 300) // [2, 3]
assert.Nil(t, cache.Lookup("1"))
assert.Equal(t, 200, cache.Lookup("2")) // [3, 2]
assert.Equal(t, 300, cache.Lookup("3")) // [2, 3]

cache.Save("2", 201) // [3, 2]
cache.Save("1", 101) // [2, 1]
assert.Equal(t, 101, cache.Lookup("1"))
assert.Equal(t, 201, cache.Lookup("2")) // [1, 2]
assert.Nil(t, cache.Lookup("3"))

cache.Save("3", 302) // [2, 3]
assert.Nil(t, cache.Lookup("1"))
assert.Equal(t, 201, cache.Lookup("2")) // [3, 2]
assert.Equal(t, 302, cache.Lookup("3")) // [2, 3]

cache.Save("1", 103) // [3, 1]
assert.Equal(t, 103, cache.Lookup("1"))
assert.Nil(t, cache.Lookup("2"))
assert.Equal(t, 302, cache.Lookup("3")) // [1, 3]

// Check if old items were deleted
assert.Equal(t, maxSize, cache.queue.Len())
assert.Equal(t, maxSize, len(cache.items))
}

func TestReset(t *testing.T) {
maxSize := 2
cache := NewLRUCache(maxSize, 1000)

cache.Save("1", 100) // [1]
cache.Save("2", 200) // [1, 2]

assert.Equal(t, maxSize, cache.queue.Len())

// cache reset
cache.Reset()
assert.Equal(t, 0, cache.queue.Len())
assert.Equal(t, 0, len(cache.items))

// validate cache fully functional after reset
cache.Save("1", 100) // [1]
cache.Save("2", 200) // [1, 2]
cache.Save("3", 300) // [2, 3]
assert.Nil(t, cache.Lookup("1"))
assert.Equal(t, 200, cache.Lookup("2")) // [3, 2]
assert.Equal(t, 300, cache.Lookup("3")) // [2, 3]

cache.Save("2", 201) // [3, 2]
cache.Save("1", 101) // [2, 1]
assert.Equal(t, 101, cache.Lookup("1"))
assert.Equal(t, 201, cache.Lookup("2")) // [1, 2]
assert.Nil(t, cache.Lookup("3"))
}

func TestSizeZero(t *testing.T) {
cache := NewLRUCache(0, 1000)
cache.Save("1", 100)
assert.Nil(t, cache.Lookup("1"))
cache.Save("2", 200)
assert.Nil(t, cache.Lookup("2"))
cache.Reset()
assert.Nil(t, cache.Lookup("1"))
assert.Nil(t, cache.Lookup("2"))
}

func TestThreadSafe(t *testing.T) {
maxSize := 1000
cache := NewLRUCache(maxSize, 1000)
wg := sync.WaitGroup{}

save := func(k int, v interface{}, wg *sync.WaitGroup) {
defer wg.Done()
strKey := fmt.Sprintf("%d", k)
cache.Save(strKey, v)
}
lookup := func(k int, wg *sync.WaitGroup, checkValue bool) {
defer wg.Done()
strKey := fmt.Sprintf("%d", k)
v := cache.Lookup(strKey)
if checkValue {
assert.Equal(t, k*100, v)
}
}
reset := func(wg *sync.WaitGroup) {
defer wg.Done()
cache.Reset()
}

// Add entries
wg.Add(maxSize)
for i := 1; i <= maxSize; i++ {
go save(i, i*100, &wg)
}
wg.Wait()

// Lookup previous entries
wg.Add(maxSize)
for i := 1; i <= maxSize; i++ {
go lookup(i, &wg, true)
}
wg.Wait()

// Add more entries then the max size
wg.Add(maxSize)
for i := maxSize + 1; i <= maxSize*2; i++ {
go save(i, i*100, &wg)
}
wg.Wait()

// Check if new entries replaced the old ones
wg.Add(maxSize)
for i := maxSize + 1; i <= maxSize*2; i++ {
go lookup(i, &wg, true)
}
wg.Wait()

// Check if old items were deleted
assert.Equal(t, maxSize, cache.queue.Len())
assert.Equal(t, maxSize, len(cache.items))

wg.Add(maxSize * 3)
// Check all api's simultaneously for race conditions
for i := 1; i <= maxSize; i++ {
go save(i, i*100, &wg)
go lookup(i, &wg, false)
go reset(&wg)
}
wg.Wait()
}

func TestTimeout(t *testing.T) {
var maxTimeout int64 = 1
// cache with timeout
cache1 := NewLRUCache(1000, maxTimeout)
// Zero timeout cache
cache2 := NewLRUCache(1000, 0)

cache1.Save("1", 100) // [1]
cache1.Save("2", 200) // [1, 2]
cache1.Save("3", 300) // [1,2,3]
cache2.Save("1", 100) // [1]
cache2.Save("2", 200) // [1, 2]
cache2.Save("3", 300) // [1,2,3]

// cache1 should expire while cache2 should not
time.Sleep(1 * time.Second)

// cache1 should expire
assert.Nil(t, cache1.Lookup("1"))
assert.Nil(t, cache1.Lookup("2"))
assert.Nil(t, cache1.Lookup("3"))
cache1.Save("1", 100) // [1]
cache1.Save("4", 400) // [1,4]
assert.Equal(t, 100, cache1.Lookup("1")) // [4,1]
assert.Equal(t, 400, cache1.Lookup("4")) // [1,4]

// cache2 should not expire
assert.Equal(t, 100, cache2.Lookup("1"))
assert.Equal(t, 200, cache2.Lookup("2"))
assert.Equal(t, 300, cache2.Lookup("3"))
}