Comments (11)
@ineiti well yes, but everything depends on a context.
AppEngine do not allow you to use anything from unsafe
package, but when all your product is build on such platform, you do not have many options. And obvious one is to remove any unsafe
things.
IMO: there is no sense to add // build !appengine
and remove unsafe
until you need it heavily. For BigCache it make sense due to wide usage and different customer's needs.
from bigcache.
I can take a stab at this. Just wondering, how critical is it to avoid the string
conversion here? Is bytesToString
being called frequently enough to warrant the optimization?
from bigcache.
@lukechampine - currently looking at all dependencies with unsafe
from c4dt/byzcoin, and I'm also wondering if it's really worth it to use unsafe
here?
from bigcache.
I did some benchmarks
name old time/op new time/op delta
WriteToCacheWith1Shard-4 537ns ± 3% 523ns ± 4% ~ (p=0.400 n=3+3)
WriteToLimitedCacheWithSmallInitSizeAnd1Shard-4 400ns ± 1% 401ns ± 0% ~ (p=0.600 n=3+3)
WriteToUnlimitedCacheWithSmallInitSizeAnd1Shard-4 836ns ±11% 814ns ±12% ~ (p=0.700 n=3+3)
WriteToCache/1-shards-4 519ns ± 1% 503ns ± 1% ~ (p=0.100 n=3+3)
WriteToCache/512-shards-4 215ns ± 0% 214ns ± 2% ~ (p=0.600 n=3+3)
WriteToCache/1024-shards-4 217ns ± 2% 218ns ± 1% ~ (p=0.700 n=3+3)
WriteToCache/8192-shards-4 280ns ± 0% 277ns ± 1% ~ (p=0.200 n=3+3)
ReadFromCache/1-shards-4 241ns ± 2% 237ns ± 1% ~ (p=0.500 n=3+3)
ReadFromCache/512-shards-4 250ns ± 0% 241ns ± 2% ~ (p=0.100 n=3+3)
ReadFromCache/1024-shards-4 250ns ± 2% 240ns ± 1% ~ (p=0.100 n=3+3)
ReadFromCache/8192-shards-4 276ns ± 1% 264ns ± 1% ~ (p=0.100 n=3+3)
ReadFromCacheWithInfo/1-shards-4 250ns ± 1% 242ns ± 1% ~ (p=0.100 n=3+3)
ReadFromCacheWithInfo/512-shards-4 254ns ± 2% 254ns ± 1% ~ (p=1.000 n=3+3)
ReadFromCacheWithInfo/1024-shards-4 253ns ± 2% 251ns ± 1% ~ (p=1.000 n=3+3)
ReadFromCacheWithInfo/8192-shards-4 279ns ± 1% 275ns ± 1% ~ (p=0.200 n=3+3)
IterateOverCache/512-shards-4 159ns ± 1% 180ns ± 0% ~ (p=0.100 n=3+3)
IterateOverCache/1024-shards-4 159ns ± 0% 179ns ± 1% ~ (p=0.100 n=3+3)
IterateOverCache/8192-shards-4 155ns ± 0% 179ns ± 0% ~ (p=0.100 n=3+3)
WriteToCacheWith1024ShardsAndSmallShardInitSize-4 320ns ± 8% 297ns ± 4% ~ (p=0.200 n=3+3)
ReadFromCacheNonExistentKeys/1-shards-4 114ns ± 1% 111ns ± 1% ~ (p=0.100 n=3+3)
ReadFromCacheNonExistentKeys/512-shards-4 112ns ± 0% 110ns ± 1% ~ (p=0.100 n=3+3)
ReadFromCacheNonExistentKeys/1024-shards-4 113ns ± 3% 112ns ± 1% ~ (p=1.000 n=3+3)
ReadFromCacheNonExistentKeys/8192-shards-4 113ns ± 2% 111ns ± 1% ~ (p=0.300 n=3+3)
FnvHashSum64-4 6.51ns ± 0% 6.70ns ± 0% ~ (p=0.100 n=3+3)
FnvHashStdLibSum64-4 40.0ns ± 0% 40.0ns ± 0% ~ (p=0.600 n=3+3)
name old alloc/op new alloc/op delta
WriteToCacheWith1Shard-4 602B ± 0% 602B ± 0% ~ (p=1.000 n=3+3)
WriteToLimitedCacheWithSmallInitSizeAnd1Shard-4 41.0B ± 0% 41.0B ± 0% ~ (all equal)
WriteToUnlimitedCacheWithSmallInitSizeAnd1Shard-4 3.94kB ± 7% 3.90kB ± 7% ~ (p=0.400 n=3+3)
WriteToCache/1-shards-4 602B ± 0% 600B ± 0% ~ (p=0.100 n=3+3)
WriteToCache/512-shards-4 597B ± 0% 596B ± 0% ~ (p=0.700 n=3+3)
WriteToCache/1024-shards-4 596B ± 0% 597B ± 0% ~ (p=1.000 n=3+3)
WriteToCache/8192-shards-4 628B ± 2% 619B ± 1% ~ (p=0.300 n=3+3)
ReadFromCache/1-shards-4 271B ± 0% 279B ± 0% ~ (p=0.100 n=3+3)
ReadFromCache/512-shards-4 271B ± 0% 279B ± 0% ~ (p=0.100 n=3+3)
ReadFromCache/1024-shards-4 271B ± 0% 279B ± 0% ~ (p=0.100 n=3+3)
ReadFromCache/8192-shards-4 271B ± 0% 279B ± 0% ~ (p=0.100 n=3+3)
ReadFromCacheWithInfo/1-shards-4 271B ± 0% 279B ± 0% ~ (p=0.100 n=3+3)
ReadFromCacheWithInfo/512-shards-4 271B ± 0% 279B ± 0% ~ (p=0.100 n=3+3)
ReadFromCacheWithInfo/1024-shards-4 271B ± 0% 279B ± 0% ~ (p=0.100 n=3+3)
ReadFromCacheWithInfo/8192-shards-4 271B ± 0% 279B ± 0% ~ (p=0.100 n=3+3)
IterateOverCache/512-shards-4 20.0B ± 0% 36.0B ± 0% ~ (p=0.100 n=3+3)
IterateOverCache/1024-shards-4 20.0B ± 0% 36.0B ± 0% ~ (p=0.100 n=3+3)
IterateOverCache/8192-shards-4 20.0B ± 0% 36.0B ± 0% ~ (p=0.100 n=3+3)
WriteToCacheWith1024ShardsAndSmallShardInitSize-4 1.07kB ±24% 0.89kB ±14% ~ (p=0.400 n=3+3)
ReadFromCacheNonExistentKeys/1-shards-4 7.00B ± 0% 7.00B ± 0% ~ (all equal)
ReadFromCacheNonExistentKeys/512-shards-4 7.00B ± 0% 7.00B ± 0% ~ (all equal)
ReadFromCacheNonExistentKeys/1024-shards-4 7.00B ± 0% 7.00B ± 0% ~ (all equal)
ReadFromCacheNonExistentKeys/8192-shards-4 7.00B ± 0% 7.00B ± 0% ~ (all equal)
FnvHashSum64-4 0.00B 0.00B ~ (all equal)
FnvHashStdLibSum64-4 16.0B ± 0% 16.0B ± 0% ~ (all equal)
name old allocs/op new allocs/op delta
WriteToCacheWith1Shard-4 3.00 ± 0% 3.00 ± 0% ~ (all equal)
WriteToLimitedCacheWithSmallInitSizeAnd1Shard-4 3.00 ± 0% 3.00 ± 0% ~ (all equal)
WriteToUnlimitedCacheWithSmallInitSizeAnd1Shard-4 2.00 ± 0% 2.00 ± 0% ~ (all equal)
WriteToCache/1-shards-4 3.00 ± 0% 3.00 ± 0% ~ (all equal)
WriteToCache/512-shards-4 3.00 ± 0% 3.00 ± 0% ~ (all equal)
WriteToCache/1024-shards-4 3.00 ± 0% 3.00 ± 0% ~ (all equal)
WriteToCache/8192-shards-4 3.00 ± 0% 3.00 ± 0% ~ (all equal)
ReadFromCache/1-shards-4 2.00 ± 0% 3.00 ± 0% ~ (p=0.100 n=3+3)
ReadFromCache/512-shards-4 2.00 ± 0% 3.00 ± 0% ~ (p=0.100 n=3+3)
ReadFromCache/1024-shards-4 2.00 ± 0% 3.00 ± 0% ~ (p=0.100 n=3+3)
ReadFromCache/8192-shards-4 2.00 ± 0% 3.00 ± 0% ~ (p=0.100 n=3+3)
ReadFromCacheWithInfo/1-shards-4 2.00 ± 0% 3.00 ± 0% ~ (p=0.100 n=3+3)
ReadFromCacheWithInfo/512-shards-4 2.00 ± 0% 3.00 ± 0% ~ (p=0.100 n=3+3)
ReadFromCacheWithInfo/1024-shards-4 2.00 ± 0% 3.00 ± 0% ~ (p=0.100 n=3+3)
ReadFromCacheWithInfo/8192-shards-4 2.00 ± 0% 3.00 ± 0% ~ (p=0.100 n=3+3)
IterateOverCache/512-shards-4 2.00 ± 0% 3.00 ± 0% ~ (p=0.100 n=3+3)
IterateOverCache/1024-shards-4 2.00 ± 0% 3.00 ± 0% ~ (p=0.100 n=3+3)
IterateOverCache/8192-shards-4 2.00 ± 0% 3.00 ± 0% ~ (p=0.100 n=3+3)
WriteToCacheWith1024ShardsAndSmallShardInitSize-4 3.00 ± 0% 3.00 ± 0% ~ (all equal)
ReadFromCacheNonExistentKeys/1-shards-4 0.00 0.00 ~ (all equal)
ReadFromCacheNonExistentKeys/512-shards-4 0.00 0.00 ~ (all equal)
ReadFromCacheNonExistentKeys/1024-shards-4 0.00 0.00 ~ (all equal)
ReadFromCacheNonExistentKeys/8192-shards-4 0.00 0.00 ~ (all equal)
FnvHashSum64-4 0.00 0.00 ~ (all equal)
FnvHashStdLibSum64-4 2.00 ± 0% 2.00 ± 0% ~ (all equal)
from bigcache.
@janisz what exactly do you compare? :)
from bigcache.
go test -bench=. -benchmem ./... -count 3 -cpu 4 > new.txt && \
git rh && \
go test -bench=. -benchmem ./... -count 3 -cpu 4 > old.txt && \
benchstat old.txt new.txt
diff --git a/bytes.go b/bytes.go
deleted file mode 100644
index 3944bfe..0000000
--- a/bytes.go
+++ /dev/null
@@ -1,14 +0,0 @@
-// +build !appengine
-
-package bigcache
-
-import (
- "reflect"
- "unsafe"
-)
-
-func bytesToString(b []byte) string {
- bytesHeader := (*reflect.SliceHeader)(unsafe.Pointer(&b))
- strHeader := reflect.StringHeader{Data: bytesHeader.Data, Len: bytesHeader.Len}
- return *(*string)(unsafe.Pointer(&strHeader))
-}
diff --git a/bytes_appengine.go b/bytes_appengine.go
index 3892f3b..6718f6b 100644
--- a/bytes_appengine.go
+++ b/bytes_appengine.go
@@ -1,5 +1,3 @@
-// +build appengine
-
package bigcache
func bytesToString(b []byte) string {
from bigcache.
So I guess your diff misses the return string(b)
...
Anyway, looking at the times and memory allocated, it seems that the unsafe bytesToString
is not really worth it.
Funny thing is that https://github.com/prataprc/goparsec/blob/master/scanner.go#L228 is doing nearly the same. Was there an old version of go that misbehaved when you did a string(b)
?
from bigcache.
package bigcache_test
import (
"bytes"
"reflect"
"testing"
"unsafe"
)
func bytesToString(b []byte) string {
return string(b)
}
func bytesToStringUnsafe(b []byte) string {
bytesHeader := (*reflect.SliceHeader)(unsafe.Pointer(&b))
strHeader := reflect.StringHeader{Data: bytesHeader.Data, Len: bytesHeader.Len}
return *(*string)(unsafe.Pointer(&strHeader))
}
var a = bytes.Repeat([]byte("abcd"), 1024)
var sink = ""
func BenchmarkBytesToString(b *testing.B) {
for i := 0; i < b.N; i++ {
sink = bytesToString(a)
}
}
func BenchmarkBytesToStringUnsafe(b *testing.B) {
for i := 0; i < b.N; i++ {
sink = bytesToStringUnsafe(a)
}
}
# go test -bench="BenchmarkBytesToString" -benchmem x_test.go
goos: linux
goarch: amd64
BenchmarkBytesToString-8 1074472 977 ns/op 4096 B/op 1 allocs/op
BenchmarkBytesToStringUnsafe-8 546898618 2.06 ns/op 0 B/op 0 allocs/op
PASS
ok command-line-arguments 3.296s
from bigcache.
^^^ it does make sense, 'cause converting []byte
to string
must create an immutable object (a string). And with a current Go memory model there is no other way to do this without an allocation.
Converting []byte
to string
via unsafe, via direct memory addressing solve this. It's possible that we do not have such place, where we convert bytes to string and that's why we do not see any changes in benchmarks.
Probably we can get rid of unsafe at all, wdyt @janisz ?
from bigcache.
We need more benchmarks, The only usage of bytesToString
is in readKeyFromEntry
. We need to investigate the impact. In current benchmarks we can observe additional allocation when not using unsafe
. I did some benchmarks for older versions as well
name \ time/op go1.10.8 go1.11.13 go1.12.16 go1.13.7 go1.14rc go1.5.4 go1.6.4 go1.7.6 go1.8.7 go1.9.7
goos:linux goarch:amd64
BytesToString 757ns ± 2% 750ns ± 1% 855ns ± 2% 540ns ± 2% 510ns ± 1% 641ns ± 2%
BytesToStringUnsafe 2.05ns ± 0% 2.05ns ± 0% 2.05ns ± 0% 2.09ns ± 2% 2.05ns ± 0% 2.30ns ± 0%
BytesToString 363ns ± 1% 638ns ±12% 352ns ± 7% 665ns ± 6%
BytesToStringUnsafe 3.07ns ± 0% 3.08ns ± 0% 2.57ns ± 1% 2.57ns ± 1%
name \ alloc/op go1.10.8 go1.11.13 go1.12.16 go1.13.7 go1.14rc go1.5.4 go1.6.4 go1.7.6 go1.8.7 go1.9.7
goos:linux goarch:amd64
BytesToString 4.10kB ± 0% 4.10kB ± 0% 4.10kB ± 0% 4.10kB ± 0% 4.10kB ± 0% 4.10kB ± 0%
BytesToStringUnsafe 0.00B 0.00B 0.00B 0.00B 0.00B 0.00B
BytesToString 4.10kB ± 0% 4.10kB ± 0% 4.10kB ± 0% 4.10kB ± 0%
BytesToStringUnsafe 0.00B 0.00B 0.00B 0.00B
name \ allocs/op go1.10.8 go1.11.13 go1.12.16 go1.13.7 go1.14rc go1.5.4 go1.6.4 go1.7.6 go1.8.7 go1.9.7
goos:linux goarch:amd64
BytesToString 1.00 ± 0% 1.00 ± 0% 1.00 ± 0% 1.00 ± 0% 1.00 ± 0% 1.00 ± 0%
BytesToStringUnsafe 0.00 0.00 0.00 0.00 0.00 0.00
BytesToString 1.00 ± 0% 1.00 ± 0% 1.00 ± 0% 1.00 ± 0%
BytesToStringUnsafe 0.00 0.00 0.00 0.00
And unsafe time is constant across the versions while safe differs from 352ns ± 7% (1.9.7) to 855ns ± 2% (1.12.16) so I prefer to stay with unsafe to reduce the risk of this having impact on our codebase under different go versions.
I still do not understand why this have so minimal impact on our benchmarks
from bigcache.
I prefer to stay with unsafe to reduce the risk of this having impact on our codebase under different go versions.
I think this is the right methodology, constant performance is the right approach.
I still do not understand why this have so minimal impact on our benchmarks
Do share your results here when you find it, I am curious.
from bigcache.
Related Issues (20)
- LifeWindow Not valid HOT 1
- Error logs about bytes allocation? HOT 1
- why did my key expired HOT 7
- resetKeyFromEntry HOT 1
- Excessive memory allocation HOT 1
- The index(int) of entries(queue.BytesQueue) overflows the hashmap's uint32 value HOT 4
- Memory so high, and when clean not reduce size HOT 5
- Benchmark is not fair, So the speed can't be trusted HOT 4
- Whether it is a bug ? HOT 1
- Why doesn't bigcache consider designing expiration time for each cache key? HOT 4
- How does bigcache handle hash collisions HOT 1
- BigCache is logging allocations it should not HOT 4
- Support for the map? HOT 1
- compile a 32-bit LSB executable on arm and Execution error-> panic: unaligned 64-bit atomic operation HOT 4
- Is there a memory leak in bigcache? HOT 4
- Add multi-level cache design HOT 2
- Why do we need to configure both lifeWindow and cleanWindow? HOT 2
- GRPC server support HOT 4
- feature: Support uint64 keys without converting to string then back again HOT 6
- Proposal to Integrate SIEVE Eviction Algorithm
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from bigcache.