mirror of
https://github.com/golang/go.git
synced 2025-12-08 06:10:04 +00:00
In the original BoringCrypto port, ecdsa and rsa's public and private keys added a 'boring unsafe.Pointer' field to cache the BoringCrypto form of the key. This led to problems with code that “knew” the layout of those structs and in particular that they had no unexported fields. In response, as an awful kludge, I changed the compiler to pretend that field did not exist when laying out reflect data. Because we want to merge BoringCrypto in the main tree, we need a different solution. Using boring.Cache is that solution. For #51940. Change-Id: Ideb2b40b599a1dc223082eda35a5ea9abcc01e30 Reviewed-on: https://go-review.googlesource.com/c/go/+/395883 Run-TryBot: Russ Cox <rsc@golang.org> TryBot-Result: Gopher Robot <gobot@golang.org> Reviewed-by: Roland Shoemaker <roland@golang.org>
120 lines
2.3 KiB
Go
120 lines
2.3 KiB
Go
// Copyright 2022 The Go Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style
|
|
// license that can be found in the LICENSE file.
|
|
|
|
package boring
|
|
|
|
import (
|
|
"fmt"
|
|
"runtime"
|
|
"sync"
|
|
"sync/atomic"
|
|
"testing"
|
|
"unsafe"
|
|
)
|
|
|
|
var registeredCache Cache
|
|
|
|
func init() {
|
|
registeredCache.Register()
|
|
}
|
|
|
|
func TestCache(t *testing.T) {
|
|
// Use unregistered cache for functionality tests,
|
|
// to keep the runtime from clearing behind our backs.
|
|
c := new(Cache)
|
|
|
|
// Create many entries.
|
|
seq := uint32(0)
|
|
next := func() unsafe.Pointer {
|
|
x := new(int)
|
|
*x = int(atomic.AddUint32(&seq, 1))
|
|
return unsafe.Pointer(x)
|
|
}
|
|
m := make(map[unsafe.Pointer]unsafe.Pointer)
|
|
for i := 0; i < 10000; i++ {
|
|
k := next()
|
|
v := next()
|
|
m[k] = v
|
|
c.Put(k, v)
|
|
}
|
|
|
|
// Overwrite a random 20% of those.
|
|
n := 0
|
|
for k := range m {
|
|
v := next()
|
|
m[k] = v
|
|
c.Put(k, v)
|
|
if n++; n >= 2000 {
|
|
break
|
|
}
|
|
}
|
|
|
|
// Check results.
|
|
str := func(p unsafe.Pointer) string {
|
|
if p == nil {
|
|
return "nil"
|
|
}
|
|
return fmt.Sprint(*(*int)(p))
|
|
}
|
|
for k, v := range m {
|
|
if cv := c.Get(k); cv != v {
|
|
t.Fatalf("c.Get(%v) = %v, want %v", str(k), str(cv), str(v))
|
|
}
|
|
}
|
|
|
|
c.Clear()
|
|
for k := range m {
|
|
if cv := c.Get(k); cv != nil {
|
|
t.Fatalf("after GC, c.Get(%v) = %v, want nil", str(k), str(cv))
|
|
}
|
|
}
|
|
|
|
// Check that registered cache is cleared at GC.
|
|
c = ®isteredCache
|
|
for k, v := range m {
|
|
c.Put(k, v)
|
|
}
|
|
runtime.GC()
|
|
for k := range m {
|
|
if cv := c.Get(k); cv != nil {
|
|
t.Fatalf("after Clear, c.Get(%v) = %v, want nil", str(k), str(cv))
|
|
}
|
|
}
|
|
|
|
// Check that cache works for concurrent access.
|
|
// Lists are discarded if they reach 1000 entries,
|
|
// and there are cacheSize list heads, so we should be
|
|
// able to do 100 * cacheSize entries with no problem at all.
|
|
c = new(Cache)
|
|
var barrier, wg sync.WaitGroup
|
|
const N = 100
|
|
barrier.Add(N)
|
|
wg.Add(N)
|
|
var lost int32
|
|
for i := 0; i < N; i++ {
|
|
go func() {
|
|
defer wg.Done()
|
|
|
|
m := make(map[unsafe.Pointer]unsafe.Pointer)
|
|
for j := 0; j < cacheSize; j++ {
|
|
k, v := next(), next()
|
|
m[k] = v
|
|
c.Put(k, v)
|
|
}
|
|
barrier.Done()
|
|
barrier.Wait()
|
|
|
|
for k, v := range m {
|
|
if cv := c.Get(k); cv != v {
|
|
t.Errorf("c.Get(%v) = %v, want %v", str(k), str(cv), str(v))
|
|
atomic.AddInt32(&lost, +1)
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
wg.Wait()
|
|
if lost != 0 {
|
|
t.Errorf("lost %d entries", lost)
|
|
}
|
|
}
|