crypto, pow, vendor: hash optimizations, mmap ethash
This commit is contained in:
parent
b7d93500f1
commit
5c8fa6ae1a
@ -37,10 +37,6 @@ var (
|
|||||||
secp256k1_halfN = new(big.Int).Div(secp256k1_N, big.NewInt(2))
|
secp256k1_halfN = new(big.Int).Div(secp256k1_N, big.NewInt(2))
|
||||||
)
|
)
|
||||||
|
|
||||||
// Hasher is a repetitive hasher allowing the same hash data structures to be
|
|
||||||
// reused between hash runs instead of requiring new ones to be created.
|
|
||||||
type Hasher func(data []byte) []byte
|
|
||||||
|
|
||||||
// Keccak256 calculates and returns the Keccak256 hash of the input data.
|
// Keccak256 calculates and returns the Keccak256 hash of the input data.
|
||||||
func Keccak256(data ...[]byte) []byte {
|
func Keccak256(data ...[]byte) []byte {
|
||||||
d := sha3.NewKeccak256()
|
d := sha3.NewKeccak256()
|
||||||
@ -61,22 +57,6 @@ func Keccak256Hash(data ...[]byte) (h common.Hash) {
|
|||||||
return h
|
return h
|
||||||
}
|
}
|
||||||
|
|
||||||
// Keccak256Hasher creates a repetitive Keccak256 hasher, allowing the same hash
|
|
||||||
// data structures to be reused between hash runs instead of requiring new ones
|
|
||||||
// to be created.
|
|
||||||
//
|
|
||||||
// The returned function is not thread safe!
|
|
||||||
func Keccak256Hasher() Hasher {
|
|
||||||
hasher := sha3.NewKeccak256()
|
|
||||||
|
|
||||||
return func(data []byte) []byte {
|
|
||||||
hasher.Write(data)
|
|
||||||
result := hasher.Sum(nil)
|
|
||||||
hasher.Reset()
|
|
||||||
return result
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Keccak512 calculates and returns the Keccak512 hash of the input data.
|
// Keccak512 calculates and returns the Keccak512 hash of the input data.
|
||||||
func Keccak512(data ...[]byte) []byte {
|
func Keccak512(data ...[]byte) []byte {
|
||||||
d := sha3.NewKeccak512()
|
d := sha3.NewKeccak512()
|
||||||
@ -86,22 +66,6 @@ func Keccak512(data ...[]byte) []byte {
|
|||||||
return d.Sum(nil)
|
return d.Sum(nil)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Keccak512Hasher creates a repetitive Keccak512 hasher, allowing the same hash
|
|
||||||
// data structures to be reused between hash runs instead of requiring new ones
|
|
||||||
// to be created.
|
|
||||||
//
|
|
||||||
// The returned function is not thread safe!
|
|
||||||
func Keccak512Hasher() Hasher {
|
|
||||||
hasher := sha3.NewKeccak512()
|
|
||||||
|
|
||||||
return func(data []byte) []byte {
|
|
||||||
hasher.Write(data)
|
|
||||||
result := hasher.Sum(nil)
|
|
||||||
hasher.Reset()
|
|
||||||
return result
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Deprecated: For backward compatibility as other packages depend on these
|
// Deprecated: For backward compatibility as other packages depend on these
|
||||||
func Sha3Hash(data ...[]byte) common.Hash { return Keccak256Hash(data...) }
|
func Sha3Hash(data ...[]byte) common.Hash { return Keccak256Hash(data...) }
|
||||||
|
|
||||||
|
310
pow/ethash.go
310
pow/ethash.go
@ -17,20 +17,21 @@
|
|||||||
package pow
|
package pow
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bufio"
|
|
||||||
"bytes"
|
"bytes"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io/ioutil"
|
|
||||||
"math"
|
"math"
|
||||||
"math/big"
|
"math/big"
|
||||||
"math/rand"
|
"math/rand"
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
|
"reflect"
|
||||||
|
"strconv"
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
"unsafe"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
mmap "github.com/edsrzf/mmap-go"
|
||||||
"github.com/ethereum/go-ethereum/common/hexutil"
|
"github.com/ethereum/go-ethereum/common/hexutil"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
metrics "github.com/rcrowley/go-metrics"
|
metrics "github.com/rcrowley/go-metrics"
|
||||||
@ -57,10 +58,89 @@ var (
|
|||||||
dumpMagic = hexutil.MustDecode("0xfee1deadbaddcafe")
|
dumpMagic = hexutil.MustDecode("0xfee1deadbaddcafe")
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// isLittleEndian returns whether the local system is running in little or big
|
||||||
|
// endian byte order.
|
||||||
|
func isLittleEndian() bool {
|
||||||
|
n := uint32(0x01020304)
|
||||||
|
return *(*byte)(unsafe.Pointer(&n)) == 0x04
|
||||||
|
}
|
||||||
|
|
||||||
|
// memoryMap tries to memory map a file of uint32s for read only access.
|
||||||
|
func memoryMap(path string) (*os.File, mmap.MMap, []uint32, error) {
|
||||||
|
file, err := os.OpenFile(path, os.O_RDONLY, 0644)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, nil, err
|
||||||
|
}
|
||||||
|
mem, buffer, err := memoryMapFile(file, false)
|
||||||
|
if err != nil {
|
||||||
|
file.Close()
|
||||||
|
return nil, nil, nil, err
|
||||||
|
}
|
||||||
|
return file, mem, buffer, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// memoryMapFile tries to memory map an already opened file descriptor.
|
||||||
|
func memoryMapFile(file *os.File, write bool) (mmap.MMap, []uint32, error) {
|
||||||
|
// Try to memory map the file
|
||||||
|
flag := mmap.RDONLY
|
||||||
|
if write {
|
||||||
|
flag = mmap.RDWR
|
||||||
|
}
|
||||||
|
mem, err := mmap.Map(file, flag, 0)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, err
|
||||||
|
}
|
||||||
|
// Yay, we managed to memory map the file, here be dragons
|
||||||
|
header := *(*reflect.SliceHeader)(unsafe.Pointer(&mem))
|
||||||
|
header.Len /= 4
|
||||||
|
header.Cap /= 4
|
||||||
|
|
||||||
|
return mem, *(*[]uint32)(unsafe.Pointer(&header)), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// memoryMapAndGenerate tries to memory map a temporary file of uint32s for write
|
||||||
|
// access, fill it with the data from a generator and then move it into the final
|
||||||
|
// path requested.
|
||||||
|
func memoryMapAndGenerate(path string, size uint64, generator func(buffer []uint32)) (*os.File, mmap.MMap, []uint32, error) {
|
||||||
|
// Ensure the data folder exists
|
||||||
|
if err := os.MkdirAll(filepath.Dir(path), 0755); err != nil {
|
||||||
|
return nil, nil, nil, err
|
||||||
|
}
|
||||||
|
// Create a huge temporary empty file to fill with data
|
||||||
|
temp := path + "." + strconv.Itoa(rand.Int())
|
||||||
|
|
||||||
|
dump, err := os.Create(temp)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, nil, err
|
||||||
|
}
|
||||||
|
if err = dump.Truncate(int64(size)); err != nil {
|
||||||
|
return nil, nil, nil, err
|
||||||
|
}
|
||||||
|
// Memory map the file for writing and fill it with the generator
|
||||||
|
mem, buffer, err := memoryMapFile(dump, true)
|
||||||
|
if err != nil {
|
||||||
|
dump.Close()
|
||||||
|
return nil, nil, nil, err
|
||||||
|
}
|
||||||
|
generator(buffer)
|
||||||
|
|
||||||
|
if err := mem.Flush(); err != nil {
|
||||||
|
mem.Unmap()
|
||||||
|
dump.Close()
|
||||||
|
return nil, nil, nil, err
|
||||||
|
}
|
||||||
|
os.Rename(temp, path)
|
||||||
|
return dump, mem, buffer, nil
|
||||||
|
}
|
||||||
|
|
||||||
// cache wraps an ethash cache with some metadata to allow easier concurrent use.
|
// cache wraps an ethash cache with some metadata to allow easier concurrent use.
|
||||||
type cache struct {
|
type cache struct {
|
||||||
epoch uint64 // Epoch for which this cache is relevant
|
epoch uint64 // Epoch for which this cache is relevant
|
||||||
cache []uint32 // The actual cache data content
|
|
||||||
|
dump *os.File // File descriptor of the memory mapped cache
|
||||||
|
mmap mmap.MMap // Memory map itself to unmap before releasing
|
||||||
|
|
||||||
|
cache []uint32 // The actual cache data content (may be memory mapped)
|
||||||
used time.Time // Timestamp of the last use for smarter eviction
|
used time.Time // Timestamp of the last use for smarter eviction
|
||||||
once sync.Once // Ensures the cache is generated only once
|
once sync.Once // Ensures the cache is generated only once
|
||||||
lock sync.Mutex // Ensures thread safety for updating the usage time
|
lock sync.Mutex // Ensures thread safety for updating the usage time
|
||||||
@ -71,57 +151,72 @@ func (c *cache) generate(dir string, limit int, test bool) {
|
|||||||
c.once.Do(func() {
|
c.once.Do(func() {
|
||||||
// If we have a testing cache, generate and return
|
// If we have a testing cache, generate and return
|
||||||
if test {
|
if test {
|
||||||
rawCache := generateCache(1024, seedHash(c.epoch*epochLength+1))
|
c.cache = make([]uint32, 1024/4)
|
||||||
c.cache = prepare(1024, bytes.NewReader(rawCache))
|
generateCache(c.cache, c.epoch, seedHash(c.epoch*epochLength+1))
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
// Full cache generation is needed, check cache dir for existing data
|
// If we don't store anything on disk, generate and return
|
||||||
size := cacheSize(c.epoch*epochLength + 1)
|
size := cacheSize(c.epoch*epochLength + 1)
|
||||||
seed := seedHash(c.epoch*epochLength + 1)
|
seed := seedHash(c.epoch*epochLength + 1)
|
||||||
|
|
||||||
path := filepath.Join(dir, fmt.Sprintf("cache-R%d-%x", algorithmRevision, seed))
|
if dir == "" {
|
||||||
logger := log.New("seed", hexutil.Bytes(seed))
|
c.cache = make([]uint32, size/4)
|
||||||
|
generateCache(c.cache, c.epoch, seed)
|
||||||
if dir != "" {
|
return
|
||||||
dump, err := os.Open(path)
|
|
||||||
if err == nil {
|
|
||||||
logger.Info("Loading ethash cache from disk")
|
|
||||||
start := time.Now()
|
|
||||||
c.cache = prepare(size, bufio.NewReader(dump))
|
|
||||||
logger.Info("Loaded ethash cache from disk", "elapsed", common.PrettyDuration(time.Since(start)))
|
|
||||||
|
|
||||||
dump.Close()
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
// No previous disk cache was available, generate on the fly
|
// Disk storage is needed, this will get fancy
|
||||||
rawCache := generateCache(size, seed)
|
endian := "le"
|
||||||
c.cache = prepare(size, bytes.NewReader(rawCache))
|
if !isLittleEndian() {
|
||||||
|
endian = "be"
|
||||||
|
}
|
||||||
|
path := filepath.Join(dir, fmt.Sprintf("cache-R%d-%x.%s", algorithmRevision, seed, endian))
|
||||||
|
logger := log.New("epoch", c.epoch)
|
||||||
|
|
||||||
// If a cache directory is given, attempt to serialize for next time
|
// Try to load the file from disk and memory map it
|
||||||
if dir != "" {
|
var err error
|
||||||
// Store the ethash cache to disk
|
c.dump, c.mmap, c.cache, err = memoryMap(path)
|
||||||
start := time.Now()
|
if err == nil {
|
||||||
if err := os.MkdirAll(filepath.Dir(path), os.ModePerm); err != nil {
|
logger.Debug("Loaded old ethash cache from disk")
|
||||||
logger.Error("Failed to create ethash cache dir", "err", err)
|
return
|
||||||
} else if err := ioutil.WriteFile(path, rawCache, os.ModePerm); err != nil {
|
}
|
||||||
logger.Error("Failed to write ethash cache to disk", "err", err)
|
logger.Debug("Failed to load old ethash cache", "err", err)
|
||||||
} else {
|
|
||||||
logger.Info("Stored ethash cache to disk", "elapsed", common.PrettyDuration(time.Since(start)))
|
// No previous cache available, create a new cache file to fill
|
||||||
}
|
c.dump, c.mmap, c.cache, err = memoryMapAndGenerate(path, size, func(buffer []uint32) { generateCache(buffer, c.epoch, seed) })
|
||||||
// Iterate over all previous instances and delete old ones
|
if err != nil {
|
||||||
for ep := int(c.epoch) - limit; ep >= 0; ep-- {
|
logger.Error("Failed to generate mapped ethash cache", "err", err)
|
||||||
seed := seedHash(uint64(ep)*epochLength + 1)
|
|
||||||
path := filepath.Join(dir, fmt.Sprintf("cache-R%d-%x", algorithmRevision, seed))
|
c.cache = make([]uint32, size/4)
|
||||||
os.Remove(path)
|
generateCache(c.cache, c.epoch, seed)
|
||||||
}
|
}
|
||||||
|
// Iterate over all previous instances and delete old ones
|
||||||
|
for ep := int(c.epoch) - limit; ep >= 0; ep-- {
|
||||||
|
seed := seedHash(uint64(ep)*epochLength + 1)
|
||||||
|
path := filepath.Join(dir, fmt.Sprintf("cache-R%d-%x.%s", algorithmRevision, seed, endian))
|
||||||
|
os.Remove(path)
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// release closes any file handlers and memory maps open.
|
||||||
|
func (c *cache) release() {
|
||||||
|
if c.mmap != nil {
|
||||||
|
c.mmap.Unmap()
|
||||||
|
c.mmap = nil
|
||||||
|
}
|
||||||
|
if c.dump != nil {
|
||||||
|
c.dump.Close()
|
||||||
|
c.dump = nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// dataset wraps an ethash dataset with some metadata to allow easier concurrent use.
|
// dataset wraps an ethash dataset with some metadata to allow easier concurrent use.
|
||||||
type dataset struct {
|
type dataset struct {
|
||||||
epoch uint64 // Epoch for which this cache is relevant
|
epoch uint64 // Epoch for which this cache is relevant
|
||||||
|
|
||||||
|
dump *os.File // File descriptor of the memory mapped cache
|
||||||
|
mmap mmap.MMap // Memory map itself to unmap before releasing
|
||||||
|
|
||||||
dataset []uint32 // The actual cache data content
|
dataset []uint32 // The actual cache data content
|
||||||
used time.Time // Timestamp of the last use for smarter eviction
|
used time.Time // Timestamp of the last use for smarter eviction
|
||||||
once sync.Once // Ensures the cache is generated only once
|
once sync.Once // Ensures the cache is generated only once
|
||||||
@ -129,78 +224,91 @@ type dataset struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// generate ensures that the dataset content is generated before use.
|
// generate ensures that the dataset content is generated before use.
|
||||||
func (d *dataset) generate(dir string, limit int, test bool, discard bool) {
|
func (d *dataset) generate(dir string, limit int, test bool) {
|
||||||
d.once.Do(func() {
|
d.once.Do(func() {
|
||||||
// If we have a testing dataset, generate and return
|
// If we have a testing dataset, generate and return
|
||||||
if test {
|
if test {
|
||||||
rawCache := generateCache(1024, seedHash(d.epoch*epochLength+1))
|
cache := make([]uint32, 1024/4)
|
||||||
intCache := prepare(1024, bytes.NewReader(rawCache))
|
generateCache(cache, d.epoch, seedHash(d.epoch*epochLength+1))
|
||||||
|
|
||||||
rawDataset := generateDataset(32*1024, intCache)
|
d.dataset = make([]uint32, 32*1024/4)
|
||||||
d.dataset = prepare(32*1024, bytes.NewReader(rawDataset))
|
generateDataset(d.dataset, d.epoch, cache)
|
||||||
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
// Full dataset generation is needed, check dataset dir for existing data
|
// If we don't store anything on disk, generate and return
|
||||||
csize := cacheSize(d.epoch*epochLength + 1)
|
csize := cacheSize(d.epoch*epochLength + 1)
|
||||||
dsize := datasetSize(d.epoch*epochLength + 1)
|
dsize := datasetSize(d.epoch*epochLength + 1)
|
||||||
seed := seedHash(d.epoch*epochLength + 1)
|
seed := seedHash(d.epoch*epochLength + 1)
|
||||||
|
|
||||||
path := filepath.Join(dir, fmt.Sprintf("full-R%d-%x", algorithmRevision, seed))
|
if dir == "" {
|
||||||
logger := log.New("seed", hexutil.Bytes(seed))
|
cache := make([]uint32, csize/4)
|
||||||
|
generateCache(cache, d.epoch, seed)
|
||||||
|
|
||||||
if dir != "" {
|
d.dataset = make([]uint32, dsize/4)
|
||||||
dump, err := os.Open(path)
|
generateDataset(d.dataset, d.epoch, cache)
|
||||||
if err == nil {
|
|
||||||
if !discard {
|
|
||||||
logger.Info("Loading ethash DAG from disk")
|
|
||||||
start := time.Now()
|
|
||||||
d.dataset = prepare(dsize, bufio.NewReader(dump))
|
|
||||||
logger.Info("Loaded ethash DAG from disk", "elapsed", common.PrettyDuration(time.Since(start)))
|
|
||||||
}
|
|
||||||
dump.Close()
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
// No previous disk dataset was available, generate on the fly
|
// Disk storage is needed, this will get fancy
|
||||||
rawCache := generateCache(csize, seed)
|
endian := "le"
|
||||||
intCache := prepare(csize, bytes.NewReader(rawCache))
|
if !isLittleEndian() {
|
||||||
|
endian = "be"
|
||||||
|
}
|
||||||
|
path := filepath.Join(dir, fmt.Sprintf("full-R%d-%x.%s", algorithmRevision, seed, endian))
|
||||||
|
logger := log.New("epoch", d.epoch)
|
||||||
|
|
||||||
rawDataset := generateDataset(dsize, intCache)
|
// Try to load the file from disk and memory map it
|
||||||
if !discard {
|
var err error
|
||||||
d.dataset = prepare(dsize, bytes.NewReader(rawDataset))
|
d.dump, d.mmap, d.dataset, err = memoryMap(path)
|
||||||
|
if err == nil {
|
||||||
|
logger.Debug("Loaded old ethash dataset from disk")
|
||||||
|
return
|
||||||
}
|
}
|
||||||
// If a dataset directory is given, attempt to serialize for next time
|
logger.Debug("Failed to load old ethash dataset", "err", err)
|
||||||
if dir != "" {
|
|
||||||
// Store the ethash dataset to disk
|
// No previous dataset available, create a new dataset file to fill
|
||||||
start := time.Now()
|
cache := make([]uint32, csize/4)
|
||||||
if err := os.MkdirAll(filepath.Dir(path), os.ModePerm); err != nil {
|
generateCache(cache, d.epoch, seed)
|
||||||
logger.Error("Failed to create ethash DAG dir", "err", err)
|
|
||||||
} else if err := ioutil.WriteFile(path, rawDataset, os.ModePerm); err != nil {
|
d.dump, d.mmap, d.dataset, err = memoryMapAndGenerate(path, dsize, func(buffer []uint32) { generateDataset(buffer, d.epoch, cache) })
|
||||||
logger.Error("Failed to write ethash DAG to disk", "err", err)
|
if err != nil {
|
||||||
} else {
|
logger.Error("Failed to generate mapped ethash dataset", "err", err)
|
||||||
logger.Info("Stored ethash DAG to disk", "elapsed", common.PrettyDuration(time.Since(start)))
|
|
||||||
}
|
d.dataset = make([]uint32, dsize/2)
|
||||||
// Iterate over all previous instances and delete old ones
|
generateDataset(d.dataset, d.epoch, cache)
|
||||||
for ep := int(d.epoch) - limit; ep >= 0; ep-- {
|
}
|
||||||
seed := seedHash(uint64(ep)*epochLength + 1)
|
// Iterate over all previous instances and delete old ones
|
||||||
path := filepath.Join(dir, fmt.Sprintf("full-R%d-%x", algorithmRevision, seed))
|
for ep := int(d.epoch) - limit; ep >= 0; ep-- {
|
||||||
os.Remove(path)
|
seed := seedHash(uint64(ep)*epochLength + 1)
|
||||||
}
|
path := filepath.Join(dir, fmt.Sprintf("full-R%d-%x.%s", algorithmRevision, seed, endian))
|
||||||
|
os.Remove(path)
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// release closes any file handlers and memory maps open.
|
||||||
|
func (d *dataset) release() {
|
||||||
|
if d.mmap != nil {
|
||||||
|
d.mmap.Unmap()
|
||||||
|
d.mmap = nil
|
||||||
|
}
|
||||||
|
if d.dump != nil {
|
||||||
|
d.dump.Close()
|
||||||
|
d.dump = nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// MakeCache generates a new ethash cache and optionally stores it to disk.
|
// MakeCache generates a new ethash cache and optionally stores it to disk.
|
||||||
func MakeCache(block uint64, dir string) {
|
func MakeCache(block uint64, dir string) {
|
||||||
c := cache{epoch: block/epochLength + 1}
|
c := cache{epoch: block/epochLength + 1}
|
||||||
c.generate(dir, math.MaxInt32, false)
|
c.generate(dir, math.MaxInt32, false)
|
||||||
|
c.release()
|
||||||
}
|
}
|
||||||
|
|
||||||
// MakeDataset generates a new ethash dataset and optionally stores it to disk.
|
// MakeDataset generates a new ethash dataset and optionally stores it to disk.
|
||||||
func MakeDataset(block uint64, dir string) {
|
func MakeDataset(block uint64, dir string) {
|
||||||
d := dataset{epoch: block/epochLength + 1}
|
d := dataset{epoch: block/epochLength + 1}
|
||||||
d.generate(dir, math.MaxInt32, false, true)
|
d.generate(dir, math.MaxInt32, false)
|
||||||
|
d.release()
|
||||||
}
|
}
|
||||||
|
|
||||||
// Ethash is a PoW data struture implementing the ethash algorithm.
|
// Ethash is a PoW data struture implementing the ethash algorithm.
|
||||||
@ -318,22 +426,26 @@ func (ethash *Ethash) cache(block uint64) []uint32 {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
delete(ethash.caches, evict.epoch)
|
delete(ethash.caches, evict.epoch)
|
||||||
|
evict.release()
|
||||||
|
|
||||||
log.Debug("Evicted ethash cache", "epoch", evict.epoch, "used", evict.used)
|
log.Trace("Evicted ethash cache", "epoch", evict.epoch, "used", evict.used)
|
||||||
}
|
}
|
||||||
// If we have the new cache pre-generated, use that, otherwise create a new one
|
// If we have the new cache pre-generated, use that, otherwise create a new one
|
||||||
if ethash.fcache != nil && ethash.fcache.epoch == epoch {
|
if ethash.fcache != nil && ethash.fcache.epoch == epoch {
|
||||||
log.Debug("Using pre-generated cache", "epoch", epoch)
|
log.Trace("Using pre-generated cache", "epoch", epoch)
|
||||||
current, ethash.fcache = ethash.fcache, nil
|
current, ethash.fcache = ethash.fcache, nil
|
||||||
} else {
|
} else {
|
||||||
log.Debug("Requiring new ethash cache", "epoch", epoch)
|
log.Trace("Requiring new ethash cache", "epoch", epoch)
|
||||||
current = &cache{epoch: epoch}
|
current = &cache{epoch: epoch}
|
||||||
}
|
}
|
||||||
ethash.caches[epoch] = current
|
ethash.caches[epoch] = current
|
||||||
|
|
||||||
// If we just used up the future cache, or need a refresh, regenerate
|
// If we just used up the future cache, or need a refresh, regenerate
|
||||||
if ethash.fcache == nil || ethash.fcache.epoch <= epoch {
|
if ethash.fcache == nil || ethash.fcache.epoch <= epoch {
|
||||||
log.Debug("Requiring new future ethash cache", "epoch", epoch+1)
|
if ethash.fcache != nil {
|
||||||
|
ethash.fcache.release()
|
||||||
|
}
|
||||||
|
log.Trace("Requiring new future ethash cache", "epoch", epoch+1)
|
||||||
future = &cache{epoch: epoch + 1}
|
future = &cache{epoch: epoch + 1}
|
||||||
ethash.fcache = future
|
ethash.fcache = future
|
||||||
}
|
}
|
||||||
@ -418,23 +530,27 @@ func (ethash *Ethash) dataset(block uint64) []uint32 {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
delete(ethash.datasets, evict.epoch)
|
delete(ethash.datasets, evict.epoch)
|
||||||
|
evict.release()
|
||||||
|
|
||||||
log.Debug("Evicted ethash dataset", "epoch", evict.epoch, "used", evict.used)
|
log.Trace("Evicted ethash dataset", "epoch", evict.epoch, "used", evict.used)
|
||||||
}
|
}
|
||||||
// If we have the new cache pre-generated, use that, otherwise create a new one
|
// If we have the new cache pre-generated, use that, otherwise create a new one
|
||||||
if ethash.fdataset != nil && ethash.fdataset.epoch == epoch {
|
if ethash.fdataset != nil && ethash.fdataset.epoch == epoch {
|
||||||
log.Debug("Using pre-generated dataset", "epoch", epoch)
|
log.Trace("Using pre-generated dataset", "epoch", epoch)
|
||||||
current = &dataset{epoch: ethash.fdataset.epoch} // Reload from disk
|
current = &dataset{epoch: ethash.fdataset.epoch} // Reload from disk
|
||||||
ethash.fdataset = nil
|
ethash.fdataset = nil
|
||||||
} else {
|
} else {
|
||||||
log.Debug("Requiring new ethash dataset", "epoch", epoch)
|
log.Trace("Requiring new ethash dataset", "epoch", epoch)
|
||||||
current = &dataset{epoch: epoch}
|
current = &dataset{epoch: epoch}
|
||||||
}
|
}
|
||||||
ethash.datasets[epoch] = current
|
ethash.datasets[epoch] = current
|
||||||
|
|
||||||
// If we just used up the future dataset, or need a refresh, regenerate
|
// If we just used up the future dataset, or need a refresh, regenerate
|
||||||
if ethash.fdataset == nil || ethash.fdataset.epoch <= epoch {
|
if ethash.fdataset == nil || ethash.fdataset.epoch <= epoch {
|
||||||
log.Debug("Requiring new future ethash dataset", "epoch", epoch+1)
|
if ethash.fdataset != nil {
|
||||||
|
ethash.fdataset.release()
|
||||||
|
}
|
||||||
|
log.Trace("Requiring new future ethash dataset", "epoch", epoch+1)
|
||||||
future = &dataset{epoch: epoch + 1}
|
future = &dataset{epoch: epoch + 1}
|
||||||
ethash.fdataset = future
|
ethash.fdataset = future
|
||||||
}
|
}
|
||||||
@ -443,7 +559,7 @@ func (ethash *Ethash) dataset(block uint64) []uint32 {
|
|||||||
ethash.lock.Unlock()
|
ethash.lock.Unlock()
|
||||||
|
|
||||||
// Wait for generation finish, bump the timestamp and finalize the cache
|
// Wait for generation finish, bump the timestamp and finalize the cache
|
||||||
current.generate(ethash.dagdir, ethash.dagsondisk, ethash.tester, false)
|
current.generate(ethash.dagdir, ethash.dagsondisk, ethash.tester)
|
||||||
|
|
||||||
current.lock.Lock()
|
current.lock.Lock()
|
||||||
current.used = time.Now()
|
current.used = time.Now()
|
||||||
@ -451,7 +567,7 @@ func (ethash *Ethash) dataset(block uint64) []uint32 {
|
|||||||
|
|
||||||
// If we exhausted the future dataset, now's a good time to regenerate it
|
// If we exhausted the future dataset, now's a good time to regenerate it
|
||||||
if future != nil {
|
if future != nil {
|
||||||
go future.generate(ethash.dagdir, ethash.dagsondisk, ethash.tester, true) // Discard results from memorys
|
go future.generate(ethash.dagdir, ethash.dagsondisk, ethash.tester)
|
||||||
}
|
}
|
||||||
return current.dataset
|
return current.dataset
|
||||||
}
|
}
|
||||||
|
@ -18,15 +18,17 @@ package pow
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/binary"
|
"encoding/binary"
|
||||||
"io"
|
"hash"
|
||||||
|
"reflect"
|
||||||
"runtime"
|
"runtime"
|
||||||
"sync"
|
"sync"
|
||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
"unsafe"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/common/hexutil"
|
|
||||||
"github.com/ethereum/go-ethereum/crypto"
|
"github.com/ethereum/go-ethereum/crypto"
|
||||||
|
"github.com/ethereum/go-ethereum/crypto/sha3"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -44,6 +46,22 @@ const (
|
|||||||
loopAccesses = 64 // Number of accesses in hashimoto loop
|
loopAccesses = 64 // Number of accesses in hashimoto loop
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// hasher is a repetitive hasher allowing the same hash data structures to be
|
||||||
|
// reused between hash runs instead of requiring new ones to be created.
|
||||||
|
type hasher func(dest []byte, data []byte)
|
||||||
|
|
||||||
|
// makeHasher creates a repetitive hasher, allowing the same hash data structures
|
||||||
|
// to be reused between hash runs instead of requiring new ones to be created.
|
||||||
|
//
|
||||||
|
// The returned function is not thread safe!
|
||||||
|
func makeHasher(h hash.Hash) hasher {
|
||||||
|
return func(dest []byte, data []byte) {
|
||||||
|
h.Write(data)
|
||||||
|
h.Sum(dest[:0])
|
||||||
|
h.Reset()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// seedHash is the seed to use for generating a verification cache and the mining
|
// seedHash is the seed to use for generating a verification cache and the mining
|
||||||
// dataset.
|
// dataset.
|
||||||
func seedHash(block uint64) []byte {
|
func seedHash(block uint64) []byte {
|
||||||
@ -51,9 +69,9 @@ func seedHash(block uint64) []byte {
|
|||||||
if block < epochLength {
|
if block < epochLength {
|
||||||
return seed
|
return seed
|
||||||
}
|
}
|
||||||
keccak256 := crypto.Keccak256Hasher()
|
keccak256 := makeHasher(sha3.NewKeccak256())
|
||||||
for i := 0; i < int(block/epochLength); i++ {
|
for i := 0; i < int(block/epochLength); i++ {
|
||||||
seed = keccak256(seed)
|
keccak256(seed, seed)
|
||||||
}
|
}
|
||||||
return seed
|
return seed
|
||||||
}
|
}
|
||||||
@ -63,17 +81,30 @@ func seedHash(block uint64) []byte {
|
|||||||
// memory, then performing two passes of Sergio Demian Lerner's RandMemoHash
|
// memory, then performing two passes of Sergio Demian Lerner's RandMemoHash
|
||||||
// algorithm from Strict Memory Hard Hashing Functions (2014). The output is a
|
// algorithm from Strict Memory Hard Hashing Functions (2014). The output is a
|
||||||
// set of 524288 64-byte values.
|
// set of 524288 64-byte values.
|
||||||
func generateCache(size uint64, seed []byte) []byte {
|
//
|
||||||
|
// This method places the result into dest in machine byte order.
|
||||||
|
func generateCache(dest []uint32, epoch uint64, seed []byte) {
|
||||||
// Print some debug logs to allow analysis on low end devices
|
// Print some debug logs to allow analysis on low end devices
|
||||||
logger := log.New("seed", hexutil.Bytes(seed))
|
logger := log.New("epoch", epoch)
|
||||||
logger.Debug("Generating ethash verification cache")
|
|
||||||
|
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
defer func() {
|
defer func() {
|
||||||
logger.Info("Generated ethash verification cache", "elapsed", common.PrettyDuration(time.Since(start)))
|
elapsed := time.Since(start)
|
||||||
|
|
||||||
|
logFn := logger.Debug
|
||||||
|
if elapsed > 3*time.Second {
|
||||||
|
logFn = logger.Info
|
||||||
|
}
|
||||||
|
logFn("Generated ethash verification cache", "elapsed", common.PrettyDuration(elapsed))
|
||||||
}()
|
}()
|
||||||
|
// Convert our destination slice to a byte buffer
|
||||||
|
header := *(*reflect.SliceHeader)(unsafe.Pointer(&dest))
|
||||||
|
header.Len *= 4
|
||||||
|
header.Cap *= 4
|
||||||
|
cache := *(*[]byte)(unsafe.Pointer(&header))
|
||||||
|
|
||||||
// Calculate the number of thoretical rows (we'll store in one buffer nonetheless)
|
// Calculate the number of thoretical rows (we'll store in one buffer nonetheless)
|
||||||
|
size := uint64(len(cache))
|
||||||
rows := int(size) / hashBytes
|
rows := int(size) / hashBytes
|
||||||
|
|
||||||
// Start a monitoring goroutine to report progress on low end devices
|
// Start a monitoring goroutine to report progress on low end devices
|
||||||
@ -93,13 +124,12 @@ func generateCache(size uint64, seed []byte) []byte {
|
|||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
// Create a hasher to reuse between invocations
|
// Create a hasher to reuse between invocations
|
||||||
keccak512 := crypto.Keccak512Hasher()
|
keccak512 := makeHasher(sha3.NewKeccak512())
|
||||||
|
|
||||||
// Sequentially produce the initial dataset
|
// Sequentially produce the initial dataset
|
||||||
cache := make([]byte, size)
|
keccak512(cache, seed)
|
||||||
copy(cache, keccak512(seed))
|
|
||||||
for offset := uint64(hashBytes); offset < size; offset += hashBytes {
|
for offset := uint64(hashBytes); offset < size; offset += hashBytes {
|
||||||
copy(cache[offset:], keccak512(cache[offset-hashBytes:offset]))
|
keccak512(cache[offset:], cache[offset-hashBytes:offset])
|
||||||
atomic.AddUint32(&progress, 1)
|
atomic.AddUint32(&progress, 1)
|
||||||
}
|
}
|
||||||
// Use a low-round version of randmemohash
|
// Use a low-round version of randmemohash
|
||||||
@ -113,26 +143,31 @@ func generateCache(size uint64, seed []byte) []byte {
|
|||||||
xorOff = (binary.LittleEndian.Uint32(cache[dstOff:]) % uint32(rows)) * hashBytes
|
xorOff = (binary.LittleEndian.Uint32(cache[dstOff:]) % uint32(rows)) * hashBytes
|
||||||
)
|
)
|
||||||
xorBytes(temp, cache[srcOff:srcOff+hashBytes], cache[xorOff:xorOff+hashBytes])
|
xorBytes(temp, cache[srcOff:srcOff+hashBytes], cache[xorOff:xorOff+hashBytes])
|
||||||
copy(cache[dstOff:], keccak512(temp))
|
keccak512(cache[dstOff:], temp)
|
||||||
|
|
||||||
atomic.AddUint32(&progress, 1)
|
atomic.AddUint32(&progress, 1)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return cache
|
// Swap the byte order on big endian systems and return
|
||||||
|
if !isLittleEndian() {
|
||||||
|
swap(cache)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// swap changes the byte order of the buffer assuming a uint32 representation.
|
||||||
|
func swap(buffer []byte) {
|
||||||
|
for i := 0; i < len(buffer); i += 4 {
|
||||||
|
binary.BigEndian.PutUint32(buffer[i:], binary.LittleEndian.Uint32(buffer[i:]))
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// prepare converts an ethash cache or dataset from a byte stream into the internal
|
// prepare converts an ethash cache or dataset from a byte stream into the internal
|
||||||
// int representation. All ethash methods work with ints to avoid constant byte to
|
// int representation. All ethash methods work with ints to avoid constant byte to
|
||||||
// int conversions as well as to handle both little and big endian systems.
|
// int conversions as well as to handle both little and big endian systems.
|
||||||
func prepare(size uint64, r io.Reader) []uint32 {
|
func prepare(dest []uint32, src []byte) {
|
||||||
ints := make([]uint32, size/4)
|
for i := 0; i < len(dest); i++ {
|
||||||
|
dest[i] = binary.LittleEndian.Uint32(src[i*4:])
|
||||||
buffer := make([]byte, 4)
|
|
||||||
for i := 0; i < len(ints); i++ {
|
|
||||||
io.ReadFull(r, buffer)
|
|
||||||
ints[i] = binary.LittleEndian.Uint32(buffer)
|
|
||||||
}
|
}
|
||||||
return ints
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// fnv is an algorithm inspired by the FNV hash, which in some cases is used as
|
// fnv is an algorithm inspired by the FNV hash, which in some cases is used as
|
||||||
@ -152,7 +187,7 @@ func fnvHash(mix []uint32, data []uint32) {
|
|||||||
|
|
||||||
// generateDatasetItem combines data from 256 pseudorandomly selected cache nodes,
|
// generateDatasetItem combines data from 256 pseudorandomly selected cache nodes,
|
||||||
// and hashes that to compute a single dataset node.
|
// and hashes that to compute a single dataset node.
|
||||||
func generateDatasetItem(cache []uint32, index uint32, keccak512 crypto.Hasher) []byte {
|
func generateDatasetItem(cache []uint32, index uint32, keccak512 hasher) []byte {
|
||||||
// Calculate the number of thoretical rows (we use one buffer nonetheless)
|
// Calculate the number of thoretical rows (we use one buffer nonetheless)
|
||||||
rows := uint32(len(cache) / hashWords)
|
rows := uint32(len(cache) / hashWords)
|
||||||
|
|
||||||
@ -163,7 +198,7 @@ func generateDatasetItem(cache []uint32, index uint32, keccak512 crypto.Hasher)
|
|||||||
for i := 1; i < hashWords; i++ {
|
for i := 1; i < hashWords; i++ {
|
||||||
binary.LittleEndian.PutUint32(mix[i*4:], cache[(index%rows)*hashWords+uint32(i)])
|
binary.LittleEndian.PutUint32(mix[i*4:], cache[(index%rows)*hashWords+uint32(i)])
|
||||||
}
|
}
|
||||||
mix = keccak512(mix)
|
keccak512(mix, mix)
|
||||||
|
|
||||||
// Convert the mix to uint32s to avoid constant bit shifting
|
// Convert the mix to uint32s to avoid constant bit shifting
|
||||||
intMix := make([]uint32, hashWords)
|
intMix := make([]uint32, hashWords)
|
||||||
@ -179,22 +214,39 @@ func generateDatasetItem(cache []uint32, index uint32, keccak512 crypto.Hasher)
|
|||||||
for i, val := range intMix {
|
for i, val := range intMix {
|
||||||
binary.LittleEndian.PutUint32(mix[i*4:], val)
|
binary.LittleEndian.PutUint32(mix[i*4:], val)
|
||||||
}
|
}
|
||||||
return keccak512(mix)
|
keccak512(mix, mix)
|
||||||
|
return mix
|
||||||
}
|
}
|
||||||
|
|
||||||
// generateDataset generates the entire ethash dataset for mining.
|
// generateDataset generates the entire ethash dataset for mining.
|
||||||
func generateDataset(size uint64, cache []uint32) []byte {
|
//
|
||||||
|
// This method places the result into dest in machine byte order.
|
||||||
|
func generateDataset(dest []uint32, epoch uint64, cache []uint32) {
|
||||||
// Print some debug logs to allow analysis on low end devices
|
// Print some debug logs to allow analysis on low end devices
|
||||||
logger := log.New("size", size)
|
logger := log.New("epoch", epoch)
|
||||||
logger.Debug("Generating ethash dataset")
|
|
||||||
|
|
||||||
defer func(start time.Time) {
|
defer func(start time.Time) {
|
||||||
logger.Debug("Generated ethash dataset", "elapsed", common.PrettyDuration(time.Since(start)))
|
elapsed := time.Since(start)
|
||||||
|
|
||||||
|
logFn := logger.Debug
|
||||||
|
if elapsed > 3*time.Second {
|
||||||
|
logFn = logger.Info
|
||||||
|
}
|
||||||
|
logFn("Generated ethash verification cache", "elapsed", common.PrettyDuration(elapsed))
|
||||||
}(time.Now())
|
}(time.Now())
|
||||||
|
|
||||||
|
// Figure out whether the bytes need to be swapped for the machine
|
||||||
|
swapped := !isLittleEndian()
|
||||||
|
|
||||||
|
// Convert our destination slice to a byte buffer
|
||||||
|
header := *(*reflect.SliceHeader)(unsafe.Pointer(&dest))
|
||||||
|
header.Len *= 4
|
||||||
|
header.Cap *= 4
|
||||||
|
dataset := *(*[]byte)(unsafe.Pointer(&header))
|
||||||
|
|
||||||
// Generate the dataset on many goroutines since it takes a while
|
// Generate the dataset on many goroutines since it takes a while
|
||||||
dataset := make([]byte, size)
|
|
||||||
threads := runtime.NumCPU()
|
threads := runtime.NumCPU()
|
||||||
|
size := uint64(len(dataset))
|
||||||
|
|
||||||
var pend sync.WaitGroup
|
var pend sync.WaitGroup
|
||||||
pend.Add(threads)
|
pend.Add(threads)
|
||||||
@ -205,7 +257,7 @@ func generateDataset(size uint64, cache []uint32) []byte {
|
|||||||
defer pend.Done()
|
defer pend.Done()
|
||||||
|
|
||||||
// Create a hasher to reuse between invocations
|
// Create a hasher to reuse between invocations
|
||||||
keccak512 := crypto.Keccak512Hasher()
|
keccak512 := makeHasher(sha3.NewKeccak512())
|
||||||
|
|
||||||
// Calculate the data segment this thread should generate
|
// Calculate the data segment this thread should generate
|
||||||
batch := uint32(size / hashBytes / uint64(threads))
|
batch := uint32(size / hashBytes / uint64(threads))
|
||||||
@ -217,7 +269,11 @@ func generateDataset(size uint64, cache []uint32) []byte {
|
|||||||
// Calculate the dataset segment
|
// Calculate the dataset segment
|
||||||
percent := uint32(size / hashBytes / 100)
|
percent := uint32(size / hashBytes / 100)
|
||||||
for index := start; index < limit; index++ {
|
for index := start; index < limit; index++ {
|
||||||
copy(dataset[index*hashBytes:], generateDatasetItem(cache, index, keccak512))
|
item := generateDatasetItem(cache, index, keccak512)
|
||||||
|
if swapped {
|
||||||
|
swap(item)
|
||||||
|
}
|
||||||
|
copy(dataset[index*hashBytes:], item)
|
||||||
|
|
||||||
if status := atomic.AddUint32(&progress, 1); status%percent == 0 {
|
if status := atomic.AddUint32(&progress, 1); status%percent == 0 {
|
||||||
logger.Info("Generating DAG in progress", "percentage", uint64(status*100)/(size/hashBytes))
|
logger.Info("Generating DAG in progress", "percentage", uint64(status*100)/(size/hashBytes))
|
||||||
@ -227,8 +283,6 @@ func generateDataset(size uint64, cache []uint32) []byte {
|
|||||||
}
|
}
|
||||||
// Wait for all the generators to finish and return
|
// Wait for all the generators to finish and return
|
||||||
pend.Wait()
|
pend.Wait()
|
||||||
|
|
||||||
return dataset
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// hashimoto aggregates data from the full dataset in order to produce our final
|
// hashimoto aggregates data from the full dataset in order to produce our final
|
||||||
@ -277,7 +331,7 @@ func hashimoto(hash []byte, nonce uint64, size uint64, lookup func(index uint32)
|
|||||||
// in-memory cache) in order to produce our final value for a particular header
|
// in-memory cache) in order to produce our final value for a particular header
|
||||||
// hash and nonce.
|
// hash and nonce.
|
||||||
func hashimotoLight(size uint64, cache []uint32, hash []byte, nonce uint64) ([]byte, []byte) {
|
func hashimotoLight(size uint64, cache []uint32, hash []byte, nonce uint64) ([]byte, []byte) {
|
||||||
keccak512 := crypto.Keccak512Hasher()
|
keccak512 := makeHasher(sha3.NewKeccak512())
|
||||||
|
|
||||||
lookup := func(index uint32) []uint32 {
|
lookup := func(index uint32) []uint32 {
|
||||||
rawData := generateDatasetItem(cache, index, keccak512)
|
rawData := generateDatasetItem(cache, index, keccak512)
|
||||||
|
@ -18,21 +18,28 @@ package pow
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
|
"io/ioutil"
|
||||||
|
"math/big"
|
||||||
|
"os"
|
||||||
|
"reflect"
|
||||||
|
"sync"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/common/hexutil"
|
"github.com/ethereum/go-ethereum/common/hexutil"
|
||||||
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Tests that verification caches can be correctly generated.
|
// Tests that verification caches can be correctly generated.
|
||||||
func TestCacheGeneration(t *testing.T) {
|
func TestCacheGeneration(t *testing.T) {
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
size uint64
|
size uint64
|
||||||
seed []byte
|
epoch uint64
|
||||||
cache []byte
|
cache []byte
|
||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
size: 1024,
|
size: 1024,
|
||||||
seed: make([]byte, 32),
|
epoch: 0,
|
||||||
cache: hexutil.MustDecode("0x" +
|
cache: hexutil.MustDecode("0x" +
|
||||||
"7ce2991c951f7bf4c4c1bb119887ee07871eb5339d7b97b8588e85c742de90e5bafd5bbe6ce93a134fb6be9ad3e30db99d9528a2ea7846833f52e9ca119b6b54" +
|
"7ce2991c951f7bf4c4c1bb119887ee07871eb5339d7b97b8588e85c742de90e5bafd5bbe6ce93a134fb6be9ad3e30db99d9528a2ea7846833f52e9ca119b6b54" +
|
||||||
"8979480c46e19972bd0738779c932c1b43e665a2fd3122fc3ddb2691f353ceb0ed3e38b8f51fd55b6940290743563c9f8fa8822e611924657501a12aafab8a8d" +
|
"8979480c46e19972bd0738779c932c1b43e665a2fd3122fc3ddb2691f353ceb0ed3e38b8f51fd55b6940290743563c9f8fa8822e611924657501a12aafab8a8d" +
|
||||||
@ -52,8 +59,8 @@ func TestCacheGeneration(t *testing.T) {
|
|||||||
"845f64fd8324bb85312979dead74f764c9677aab89801ad4f927f1c00f12e28f22422bb44200d1969d9ab377dd6b099dc6dbc3222e9321b2c1e84f8e2f07731c"),
|
"845f64fd8324bb85312979dead74f764c9677aab89801ad4f927f1c00f12e28f22422bb44200d1969d9ab377dd6b099dc6dbc3222e9321b2c1e84f8e2f07731c"),
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
size: 1024,
|
size: 1024,
|
||||||
seed: hexutil.MustDecode("0x290decd9548b62a8d60345a988386fc84ba6bc95484008f6362f93160ef3e563"),
|
epoch: 1,
|
||||||
cache: hexutil.MustDecode("0x" +
|
cache: hexutil.MustDecode("0x" +
|
||||||
"1f56855d59cc5a085720899b4377a0198f1abe948d85fe5820dc0e346b7c0931b9cde8e541d751de3b2b3275d0aabfae316209d5879297d8bd99f8a033c9d4df" +
|
"1f56855d59cc5a085720899b4377a0198f1abe948d85fe5820dc0e346b7c0931b9cde8e541d751de3b2b3275d0aabfae316209d5879297d8bd99f8a033c9d4df" +
|
||||||
"35add1029f4e6404a022d504fb8023e42989aba985a65933b0109c7218854356f9284983c9e7de97de591828ae348b63d1fc78d8db58157344d4e06530ffd422" +
|
"35add1029f4e6404a022d504fb8023e42989aba985a65933b0109c7218854356f9284983c9e7de97de591828ae348b63d1fc78d8db58157344d4e06530ffd422" +
|
||||||
@ -74,22 +81,28 @@ func TestCacheGeneration(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
for i, tt := range tests {
|
for i, tt := range tests {
|
||||||
if cache := generateCache(tt.size, tt.seed); !bytes.Equal(cache, tt.cache) {
|
cache := make([]uint32, tt.size/4)
|
||||||
t.Errorf("cache %d: content mismatch: have %x, want %x", i, cache, tt.cache)
|
generateCache(cache, tt.epoch, seedHash(tt.epoch*epochLength+1))
|
||||||
|
|
||||||
|
want := make([]uint32, tt.size/4)
|
||||||
|
prepare(want, tt.cache)
|
||||||
|
|
||||||
|
if !reflect.DeepEqual(cache, want) {
|
||||||
|
t.Errorf("cache %d: content mismatch: have %x, want %x", i, cache, want)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestDatasetGeneration(t *testing.T) {
|
func TestDatasetGeneration(t *testing.T) {
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
|
epoch uint64
|
||||||
cacheSize uint64
|
cacheSize uint64
|
||||||
cacheSeed []byte
|
|
||||||
datasetSize uint64
|
datasetSize uint64
|
||||||
dataset []byte
|
dataset []byte
|
||||||
}{
|
}{
|
||||||
{
|
{
|
||||||
|
epoch: 0,
|
||||||
cacheSize: 1024,
|
cacheSize: 1024,
|
||||||
cacheSeed: make([]byte, 32),
|
|
||||||
datasetSize: 32 * 1024,
|
datasetSize: 32 * 1024,
|
||||||
dataset: hexutil.MustDecode("0x" +
|
dataset: hexutil.MustDecode("0x" +
|
||||||
"4bc09fbd530a041dd2ec296110a29e8f130f179c59d223f51ecce3126e8b0c74326abc2f32ccd9d7f976bd0944e3ccf8479db39343cbbffa467046ca97e2da63" +
|
"4bc09fbd530a041dd2ec296110a29e8f130f179c59d223f51ecce3126e8b0c74326abc2f32ccd9d7f976bd0944e3ccf8479db39343cbbffa467046ca97e2da63" +
|
||||||
@ -608,11 +621,17 @@ func TestDatasetGeneration(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
for i, tt := range tests {
|
for i, tt := range tests {
|
||||||
rawCache := generateCache(tt.cacheSize, tt.cacheSeed)
|
cache := make([]uint32, tt.cacheSize/4)
|
||||||
cache := prepare(uint64(len(rawCache)), bytes.NewReader(rawCache))
|
generateCache(cache, tt.epoch, seedHash(tt.epoch*epochLength+1))
|
||||||
|
|
||||||
if dataset := generateDataset(tt.datasetSize, cache); !bytes.Equal(dataset, tt.dataset) {
|
dataset := make([]uint32, tt.datasetSize/4)
|
||||||
t.Errorf("dataset %d: content mismatch: have %x, want %x", i, dataset, tt.dataset)
|
generateDataset(dataset, tt.epoch, cache)
|
||||||
|
|
||||||
|
want := make([]uint32, tt.datasetSize/4)
|
||||||
|
prepare(want, tt.dataset)
|
||||||
|
|
||||||
|
if !reflect.DeepEqual(dataset, want) {
|
||||||
|
t.Errorf("dataset %d: content mismatch: have %x, want %x", i, dataset, want)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -621,12 +640,12 @@ func TestDatasetGeneration(t *testing.T) {
|
|||||||
// datasets.
|
// datasets.
|
||||||
func TestHashimoto(t *testing.T) {
|
func TestHashimoto(t *testing.T) {
|
||||||
// Create the verification cache and mining dataset
|
// Create the verification cache and mining dataset
|
||||||
var (
|
cache := make([]uint32, 1024/4)
|
||||||
rawCache = generateCache(1024, make([]byte, 32))
|
generateCache(cache, 0, make([]byte, 32))
|
||||||
cache = prepare(uint64(len(rawCache)), bytes.NewReader(rawCache))
|
|
||||||
rawDataset = generateDataset(32*1024, cache)
|
dataset := make([]uint32, 32*1024/4)
|
||||||
dataset = prepare(uint64(len(rawDataset)), bytes.NewReader(rawDataset))
|
generateDataset(dataset, 0, cache)
|
||||||
)
|
|
||||||
// Create a block to verify
|
// Create a block to verify
|
||||||
hash := hexutil.MustDecode("0xc9149cc0386e689d789a1c2f3d5d169a61a6218ed30e74414dc736e442ef3d1f")
|
hash := hexutil.MustDecode("0xc9149cc0386e689d789a1c2f3d5d169a61a6218ed30e74414dc736e442ef3d1f")
|
||||||
nonce := uint64(0)
|
nonce := uint64(0)
|
||||||
@ -650,31 +669,77 @@ func TestHashimoto(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Tests that caches generated on disk may be done concurrently.
|
||||||
|
func TestConcurrentDiskCacheGeneration(t *testing.T) {
|
||||||
|
// Create a temp folder to generate the caches into
|
||||||
|
cachedir, err := ioutil.TempDir("", "")
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("Failed to create temporary cache dir: %v", err)
|
||||||
|
}
|
||||||
|
defer os.RemoveAll(cachedir)
|
||||||
|
|
||||||
|
// Define a heavy enough block, one from mainnet should do
|
||||||
|
block := types.NewBlockWithHeader(&types.Header{
|
||||||
|
Number: big.NewInt(3311058),
|
||||||
|
ParentHash: common.HexToHash("0xd783efa4d392943503f28438ad5830b2d5964696ffc285f338585e9fe0a37a05"),
|
||||||
|
UncleHash: common.HexToHash("0x1dcc4de8dec75d7aab85b567b6ccd41ad312451b948a7413f0a142fd40d49347"),
|
||||||
|
Coinbase: common.HexToAddress("0xc0ea08a2d404d3172d2add29a45be56da40e2949"),
|
||||||
|
Root: common.HexToHash("0x77d14e10470b5850332524f8cd6f69ad21f070ce92dca33ab2858300242ef2f1"),
|
||||||
|
TxHash: common.HexToHash("0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421"),
|
||||||
|
ReceiptHash: common.HexToHash("0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421"),
|
||||||
|
Difficulty: big.NewInt(167925187834220),
|
||||||
|
GasLimit: big.NewInt(4015682),
|
||||||
|
GasUsed: big.NewInt(0),
|
||||||
|
Time: big.NewInt(1488928920),
|
||||||
|
Extra: []byte("www.bw.com"),
|
||||||
|
MixDigest: common.HexToHash("0x3e140b0784516af5e5ec6730f2fb20cca22f32be399b9e4ad77d32541f798cd0"),
|
||||||
|
Nonce: types.EncodeNonce(0xf400cd0006070c49),
|
||||||
|
})
|
||||||
|
// Simulate multiple processes sharing the same datadir
|
||||||
|
var pend sync.WaitGroup
|
||||||
|
|
||||||
|
for i := 0; i < 3; i++ {
|
||||||
|
pend.Add(1)
|
||||||
|
|
||||||
|
go func(idx int) {
|
||||||
|
defer pend.Done()
|
||||||
|
|
||||||
|
ethash := NewFullEthash(cachedir, 0, 1, "", 0, 0)
|
||||||
|
if err := ethash.Verify(block); err != nil {
|
||||||
|
t.Errorf("proc %d: block verification failed: %v", idx, err)
|
||||||
|
}
|
||||||
|
}(i)
|
||||||
|
}
|
||||||
|
pend.Wait()
|
||||||
|
}
|
||||||
|
|
||||||
// Benchmarks the cache generation performance.
|
// Benchmarks the cache generation performance.
|
||||||
func BenchmarkCacheGeneration(b *testing.B) {
|
func BenchmarkCacheGeneration(b *testing.B) {
|
||||||
for i := 0; i < b.N; i++ {
|
for i := 0; i < b.N; i++ {
|
||||||
generateCache(cacheSize(1), make([]byte, 32))
|
cache := make([]uint32, cacheSize(1)/4)
|
||||||
|
generateCache(cache, 0, make([]byte, 32))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Benchmarks the dataset (small) generation performance.
|
// Benchmarks the dataset (small) generation performance.
|
||||||
func BenchmarkSmallDatasetGeneration(b *testing.B) {
|
func BenchmarkSmallDatasetGeneration(b *testing.B) {
|
||||||
rawCache := generateCache(65536, make([]byte, 32))
|
cache := make([]uint32, 65536/4)
|
||||||
cache := prepare(uint64(len(rawCache)), bytes.NewReader(rawCache))
|
generateCache(cache, 0, make([]byte, 32))
|
||||||
|
|
||||||
b.ResetTimer()
|
b.ResetTimer()
|
||||||
for i := 0; i < b.N; i++ {
|
for i := 0; i < b.N; i++ {
|
||||||
generateDataset(32*65536, cache)
|
dataset := make([]uint32, 32*65536/4)
|
||||||
|
generateDataset(dataset, 0, cache)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Benchmarks the light verification performance.
|
// Benchmarks the light verification performance.
|
||||||
func BenchmarkHashimotoLight(b *testing.B) {
|
func BenchmarkHashimotoLight(b *testing.B) {
|
||||||
var (
|
cache := make([]uint32, cacheSize(1)/4)
|
||||||
rawCache = generateCache(cacheSize(1), make([]byte, 32))
|
generateCache(cache, 0, make([]byte, 32))
|
||||||
cache = prepare(uint64(len(rawCache)), bytes.NewReader(rawCache))
|
|
||||||
hash = hexutil.MustDecode("0xc9149cc0386e689d789a1c2f3d5d169a61a6218ed30e74414dc736e442ef3d1f")
|
hash := hexutil.MustDecode("0xc9149cc0386e689d789a1c2f3d5d169a61a6218ed30e74414dc736e442ef3d1f")
|
||||||
)
|
|
||||||
b.ResetTimer()
|
b.ResetTimer()
|
||||||
for i := 0; i < b.N; i++ {
|
for i := 0; i < b.N; i++ {
|
||||||
hashimotoLight(datasetSize(1), cache, hash, 0)
|
hashimotoLight(datasetSize(1), cache, hash, 0)
|
||||||
@ -683,13 +748,14 @@ func BenchmarkHashimotoLight(b *testing.B) {
|
|||||||
|
|
||||||
// Benchmarks the full (small) verification performance.
|
// Benchmarks the full (small) verification performance.
|
||||||
func BenchmarkHashimotoFullSmall(b *testing.B) {
|
func BenchmarkHashimotoFullSmall(b *testing.B) {
|
||||||
var (
|
cache := make([]uint32, 65536/4)
|
||||||
rawCache = generateCache(65536, make([]byte, 32))
|
generateCache(cache, 0, make([]byte, 32))
|
||||||
cache = prepare(uint64(len(rawCache)), bytes.NewReader(rawCache))
|
|
||||||
rawDataset = generateDataset(32*65536, cache)
|
dataset := make([]uint32, 32*65536/4)
|
||||||
dataset = prepare(uint64(len(rawDataset)), bytes.NewReader(rawDataset))
|
generateDataset(dataset, 0, cache)
|
||||||
hash = hexutil.MustDecode("0xc9149cc0386e689d789a1c2f3d5d169a61a6218ed30e74414dc736e442ef3d1f")
|
|
||||||
)
|
hash := hexutil.MustDecode("0xc9149cc0386e689d789a1c2f3d5d169a61a6218ed30e74414dc736e442ef3d1f")
|
||||||
|
|
||||||
b.ResetTimer()
|
b.ResetTimer()
|
||||||
for i := 0; i < b.N; i++ {
|
for i := 0; i < b.N; i++ {
|
||||||
hashimotoFull(32*65536, dataset, hash, 0)
|
hashimotoFull(32*65536, dataset, hash, 0)
|
||||||
|
25
vendor/github.com/edsrzf/mmap-go/LICENSE
generated
vendored
Normal file
25
vendor/github.com/edsrzf/mmap-go/LICENSE
generated
vendored
Normal file
@ -0,0 +1,25 @@
|
|||||||
|
Copyright (c) 2011, Evan Shaw <edsrzf@gmail.com>
|
||||||
|
All rights reserved.
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are met:
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer in the
|
||||||
|
documentation and/or other materials provided with the distribution.
|
||||||
|
* Neither the name of the copyright holder nor the
|
||||||
|
names of its contributors may be used to endorse or promote products
|
||||||
|
derived from this software without specific prior written permission.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
|
||||||
|
ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
||||||
|
WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||||
|
DISCLAIMED. IN NO EVENT SHALL <COPYRIGHT HOLDER> BE LIABLE FOR ANY
|
||||||
|
DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
|
||||||
|
(INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
|
||||||
|
LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
|
||||||
|
ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
|
||||||
|
SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
12
vendor/github.com/edsrzf/mmap-go/README.md
generated
vendored
Normal file
12
vendor/github.com/edsrzf/mmap-go/README.md
generated
vendored
Normal file
@ -0,0 +1,12 @@
|
|||||||
|
mmap-go
|
||||||
|
=======
|
||||||
|
|
||||||
|
mmap-go is a portable mmap package for the [Go programming language](http://golang.org).
|
||||||
|
It has been tested on Linux (386, amd64), OS X, and Windows (386). It should also
|
||||||
|
work on other Unix-like platforms, but hasn't been tested with them. I'm interested
|
||||||
|
to hear about the results.
|
||||||
|
|
||||||
|
I haven't been able to add more features without adding significant complexity,
|
||||||
|
so mmap-go doesn't support mprotect, mincore, and maybe a few other things.
|
||||||
|
If you're running on a Unix-like platform and need some of these features,
|
||||||
|
I suggest Gustavo Niemeyer's [gommap](http://labix.org/gommap).
|
112
vendor/github.com/edsrzf/mmap-go/mmap.go
generated
vendored
Normal file
112
vendor/github.com/edsrzf/mmap-go/mmap.go
generated
vendored
Normal file
@ -0,0 +1,112 @@
|
|||||||
|
// Copyright 2011 Evan Shaw. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
// This file defines the common package interface and contains a little bit of
|
||||||
|
// factored out logic.
|
||||||
|
|
||||||
|
// Package mmap allows mapping files into memory. It tries to provide a simple, reasonably portable interface,
|
||||||
|
// but doesn't go out of its way to abstract away every little platform detail.
|
||||||
|
// This specifically means:
|
||||||
|
// * forked processes may or may not inherit mappings
|
||||||
|
// * a file's timestamp may or may not be updated by writes through mappings
|
||||||
|
// * specifying a size larger than the file's actual size can increase the file's size
|
||||||
|
// * If the mapped file is being modified by another process while your program's running, don't expect consistent results between platforms
|
||||||
|
package mmap
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"os"
|
||||||
|
"reflect"
|
||||||
|
"unsafe"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// RDONLY maps the memory read-only.
|
||||||
|
// Attempts to write to the MMap object will result in undefined behavior.
|
||||||
|
RDONLY = 0
|
||||||
|
// RDWR maps the memory as read-write. Writes to the MMap object will update the
|
||||||
|
// underlying file.
|
||||||
|
RDWR = 1 << iota
|
||||||
|
// COPY maps the memory as copy-on-write. Writes to the MMap object will affect
|
||||||
|
// memory, but the underlying file will remain unchanged.
|
||||||
|
COPY
|
||||||
|
// If EXEC is set, the mapped memory is marked as executable.
|
||||||
|
EXEC
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// If the ANON flag is set, the mapped memory will not be backed by a file.
|
||||||
|
ANON = 1 << iota
|
||||||
|
)
|
||||||
|
|
||||||
|
// MMap represents a file mapped into memory.
|
||||||
|
type MMap []byte
|
||||||
|
|
||||||
|
// Map maps an entire file into memory.
|
||||||
|
// If ANON is set in flags, f is ignored.
|
||||||
|
func Map(f *os.File, prot, flags int) (MMap, error) {
|
||||||
|
return MapRegion(f, -1, prot, flags, 0)
|
||||||
|
}
|
||||||
|
|
||||||
|
// MapRegion maps part of a file into memory.
|
||||||
|
// The offset parameter must be a multiple of the system's page size.
|
||||||
|
// If length < 0, the entire file will be mapped.
|
||||||
|
// If ANON is set in flags, f is ignored.
|
||||||
|
func MapRegion(f *os.File, length int, prot, flags int, offset int64) (MMap, error) {
|
||||||
|
var fd uintptr
|
||||||
|
if flags&ANON == 0 {
|
||||||
|
fd = uintptr(f.Fd())
|
||||||
|
if length < 0 {
|
||||||
|
fi, err := f.Stat()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
length = int(fi.Size())
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if length <= 0 {
|
||||||
|
return nil, errors.New("anonymous mapping requires non-zero length")
|
||||||
|
}
|
||||||
|
fd = ^uintptr(0)
|
||||||
|
}
|
||||||
|
return mmap(length, uintptr(prot), uintptr(flags), fd, offset)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *MMap) header() *reflect.SliceHeader {
|
||||||
|
return (*reflect.SliceHeader)(unsafe.Pointer(m))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Lock keeps the mapped region in physical memory, ensuring that it will not be
|
||||||
|
// swapped out.
|
||||||
|
func (m MMap) Lock() error {
|
||||||
|
dh := m.header()
|
||||||
|
return lock(dh.Data, uintptr(dh.Len))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Unlock reverses the effect of Lock, allowing the mapped region to potentially
|
||||||
|
// be swapped out.
|
||||||
|
// If m is already unlocked, aan error will result.
|
||||||
|
func (m MMap) Unlock() error {
|
||||||
|
dh := m.header()
|
||||||
|
return unlock(dh.Data, uintptr(dh.Len))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Flush synchronizes the mapping's contents to the file's contents on disk.
|
||||||
|
func (m MMap) Flush() error {
|
||||||
|
dh := m.header()
|
||||||
|
return flush(dh.Data, uintptr(dh.Len))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Unmap deletes the memory mapped region, flushes any remaining changes, and sets
|
||||||
|
// m to nil.
|
||||||
|
// Trying to read or write any remaining references to m after Unmap is called will
|
||||||
|
// result in undefined behavior.
|
||||||
|
// Unmap should only be called on the slice value that was originally returned from
|
||||||
|
// a call to Map. Calling Unmap on a derived slice may cause errors.
|
||||||
|
func (m *MMap) Unmap() error {
|
||||||
|
dh := m.header()
|
||||||
|
err := unmap(dh.Data, uintptr(dh.Len))
|
||||||
|
*m = nil
|
||||||
|
return err
|
||||||
|
}
|
67
vendor/github.com/edsrzf/mmap-go/mmap_unix.go
generated
vendored
Normal file
67
vendor/github.com/edsrzf/mmap-go/mmap_unix.go
generated
vendored
Normal file
@ -0,0 +1,67 @@
|
|||||||
|
// Copyright 2011 Evan Shaw. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
// +build darwin dragonfly freebsd linux openbsd solaris netbsd
|
||||||
|
|
||||||
|
package mmap
|
||||||
|
|
||||||
|
import (
|
||||||
|
"syscall"
|
||||||
|
)
|
||||||
|
|
||||||
|
func mmap(len int, inprot, inflags, fd uintptr, off int64) ([]byte, error) {
|
||||||
|
flags := syscall.MAP_SHARED
|
||||||
|
prot := syscall.PROT_READ
|
||||||
|
switch {
|
||||||
|
case inprot© != 0:
|
||||||
|
prot |= syscall.PROT_WRITE
|
||||||
|
flags = syscall.MAP_PRIVATE
|
||||||
|
case inprot&RDWR != 0:
|
||||||
|
prot |= syscall.PROT_WRITE
|
||||||
|
}
|
||||||
|
if inprot&EXEC != 0 {
|
||||||
|
prot |= syscall.PROT_EXEC
|
||||||
|
}
|
||||||
|
if inflags&ANON != 0 {
|
||||||
|
flags |= syscall.MAP_ANON
|
||||||
|
}
|
||||||
|
|
||||||
|
b, err := syscall.Mmap(int(fd), off, len, prot, flags)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return b, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func flush(addr, len uintptr) error {
|
||||||
|
_, _, errno := syscall.Syscall(_SYS_MSYNC, addr, len, _MS_SYNC)
|
||||||
|
if errno != 0 {
|
||||||
|
return syscall.Errno(errno)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func lock(addr, len uintptr) error {
|
||||||
|
_, _, errno := syscall.Syscall(syscall.SYS_MLOCK, addr, len, 0)
|
||||||
|
if errno != 0 {
|
||||||
|
return syscall.Errno(errno)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func unlock(addr, len uintptr) error {
|
||||||
|
_, _, errno := syscall.Syscall(syscall.SYS_MUNLOCK, addr, len, 0)
|
||||||
|
if errno != 0 {
|
||||||
|
return syscall.Errno(errno)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func unmap(addr, len uintptr) error {
|
||||||
|
_, _, errno := syscall.Syscall(syscall.SYS_MUNMAP, addr, len, 0)
|
||||||
|
if errno != 0 {
|
||||||
|
return syscall.Errno(errno)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
125
vendor/github.com/edsrzf/mmap-go/mmap_windows.go
generated
vendored
Normal file
125
vendor/github.com/edsrzf/mmap-go/mmap_windows.go
generated
vendored
Normal file
@ -0,0 +1,125 @@
|
|||||||
|
// Copyright 2011 Evan Shaw. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
package mmap
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"os"
|
||||||
|
"sync"
|
||||||
|
"syscall"
|
||||||
|
)
|
||||||
|
|
||||||
|
// mmap on Windows is a two-step process.
|
||||||
|
// First, we call CreateFileMapping to get a handle.
|
||||||
|
// Then, we call MapviewToFile to get an actual pointer into memory.
|
||||||
|
// Because we want to emulate a POSIX-style mmap, we don't want to expose
|
||||||
|
// the handle -- only the pointer. We also want to return only a byte slice,
|
||||||
|
// not a struct, so it's convenient to manipulate.
|
||||||
|
|
||||||
|
// We keep this map so that we can get back the original handle from the memory address.
|
||||||
|
var handleLock sync.Mutex
|
||||||
|
var handleMap = map[uintptr]syscall.Handle{}
|
||||||
|
|
||||||
|
func mmap(len int, prot, flags, hfile uintptr, off int64) ([]byte, error) {
|
||||||
|
flProtect := uint32(syscall.PAGE_READONLY)
|
||||||
|
dwDesiredAccess := uint32(syscall.FILE_MAP_READ)
|
||||||
|
switch {
|
||||||
|
case prot© != 0:
|
||||||
|
flProtect = syscall.PAGE_WRITECOPY
|
||||||
|
dwDesiredAccess = syscall.FILE_MAP_COPY
|
||||||
|
case prot&RDWR != 0:
|
||||||
|
flProtect = syscall.PAGE_READWRITE
|
||||||
|
dwDesiredAccess = syscall.FILE_MAP_WRITE
|
||||||
|
}
|
||||||
|
if prot&EXEC != 0 {
|
||||||
|
flProtect <<= 4
|
||||||
|
dwDesiredAccess |= syscall.FILE_MAP_EXECUTE
|
||||||
|
}
|
||||||
|
|
||||||
|
// The maximum size is the area of the file, starting from 0,
|
||||||
|
// that we wish to allow to be mappable. It is the sum of
|
||||||
|
// the length the user requested, plus the offset where that length
|
||||||
|
// is starting from. This does not map the data into memory.
|
||||||
|
maxSizeHigh := uint32((off + int64(len)) >> 32)
|
||||||
|
maxSizeLow := uint32((off + int64(len)) & 0xFFFFFFFF)
|
||||||
|
// TODO: Do we need to set some security attributes? It might help portability.
|
||||||
|
h, errno := syscall.CreateFileMapping(syscall.Handle(hfile), nil, flProtect, maxSizeHigh, maxSizeLow, nil)
|
||||||
|
if h == 0 {
|
||||||
|
return nil, os.NewSyscallError("CreateFileMapping", errno)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Actually map a view of the data into memory. The view's size
|
||||||
|
// is the length the user requested.
|
||||||
|
fileOffsetHigh := uint32(off >> 32)
|
||||||
|
fileOffsetLow := uint32(off & 0xFFFFFFFF)
|
||||||
|
addr, errno := syscall.MapViewOfFile(h, dwDesiredAccess, fileOffsetHigh, fileOffsetLow, uintptr(len))
|
||||||
|
if addr == 0 {
|
||||||
|
return nil, os.NewSyscallError("MapViewOfFile", errno)
|
||||||
|
}
|
||||||
|
handleLock.Lock()
|
||||||
|
handleMap[addr] = h
|
||||||
|
handleLock.Unlock()
|
||||||
|
|
||||||
|
m := MMap{}
|
||||||
|
dh := m.header()
|
||||||
|
dh.Data = addr
|
||||||
|
dh.Len = len
|
||||||
|
dh.Cap = dh.Len
|
||||||
|
|
||||||
|
return m, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func flush(addr, len uintptr) error {
|
||||||
|
errno := syscall.FlushViewOfFile(addr, len)
|
||||||
|
if errno != nil {
|
||||||
|
return os.NewSyscallError("FlushViewOfFile", errno)
|
||||||
|
}
|
||||||
|
|
||||||
|
handleLock.Lock()
|
||||||
|
defer handleLock.Unlock()
|
||||||
|
handle, ok := handleMap[addr]
|
||||||
|
if !ok {
|
||||||
|
// should be impossible; we would've errored above
|
||||||
|
return errors.New("unknown base address")
|
||||||
|
}
|
||||||
|
|
||||||
|
errno = syscall.FlushFileBuffers(handle)
|
||||||
|
return os.NewSyscallError("FlushFileBuffers", errno)
|
||||||
|
}
|
||||||
|
|
||||||
|
func lock(addr, len uintptr) error {
|
||||||
|
errno := syscall.VirtualLock(addr, len)
|
||||||
|
return os.NewSyscallError("VirtualLock", errno)
|
||||||
|
}
|
||||||
|
|
||||||
|
func unlock(addr, len uintptr) error {
|
||||||
|
errno := syscall.VirtualUnlock(addr, len)
|
||||||
|
return os.NewSyscallError("VirtualUnlock", errno)
|
||||||
|
}
|
||||||
|
|
||||||
|
func unmap(addr, len uintptr) error {
|
||||||
|
flush(addr, len)
|
||||||
|
// Lock the UnmapViewOfFile along with the handleMap deletion.
|
||||||
|
// As soon as we unmap the view, the OS is free to give the
|
||||||
|
// same addr to another new map. We don't want another goroutine
|
||||||
|
// to insert and remove the same addr into handleMap while
|
||||||
|
// we're trying to remove our old addr/handle pair.
|
||||||
|
handleLock.Lock()
|
||||||
|
defer handleLock.Unlock()
|
||||||
|
err := syscall.UnmapViewOfFile(addr)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
handle, ok := handleMap[addr]
|
||||||
|
if !ok {
|
||||||
|
// should be impossible; we would've errored above
|
||||||
|
return errors.New("unknown base address")
|
||||||
|
}
|
||||||
|
delete(handleMap, addr)
|
||||||
|
|
||||||
|
e := syscall.CloseHandle(syscall.Handle(handle))
|
||||||
|
return os.NewSyscallError("CloseHandle", e)
|
||||||
|
}
|
8
vendor/github.com/edsrzf/mmap-go/msync_netbsd.go
generated
vendored
Normal file
8
vendor/github.com/edsrzf/mmap-go/msync_netbsd.go
generated
vendored
Normal file
@ -0,0 +1,8 @@
|
|||||||
|
// Copyright 2011 Evan Shaw. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
package mmap
|
||||||
|
|
||||||
|
const _SYS_MSYNC = 277
|
||||||
|
const _MS_SYNC = 0x04
|
14
vendor/github.com/edsrzf/mmap-go/msync_unix.go
generated
vendored
Normal file
14
vendor/github.com/edsrzf/mmap-go/msync_unix.go
generated
vendored
Normal file
@ -0,0 +1,14 @@
|
|||||||
|
// Copyright 2011 Evan Shaw. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
// +build darwin dragonfly freebsd linux openbsd solaris
|
||||||
|
|
||||||
|
package mmap
|
||||||
|
|
||||||
|
import (
|
||||||
|
"syscall"
|
||||||
|
)
|
||||||
|
|
||||||
|
const _SYS_MSYNC = syscall.SYS_MSYNC
|
||||||
|
const _MS_SYNC = syscall.MS_SYNC
|
6
vendor/vendor.json
vendored
6
vendor/vendor.json
vendored
@ -34,6 +34,12 @@
|
|||||||
"revision": "346938d642f2ec3594ed81d874461961cd0faa76",
|
"revision": "346938d642f2ec3594ed81d874461961cd0faa76",
|
||||||
"revisionTime": "2016-10-29T20:57:26Z"
|
"revisionTime": "2016-10-29T20:57:26Z"
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"checksumSHA1": "zYnPsNAVm1/ViwCkN++dX2JQhBo=",
|
||||||
|
"path": "github.com/edsrzf/mmap-go",
|
||||||
|
"revision": "935e0e8a636ca4ba70b713f3e38a19e1b77739e8",
|
||||||
|
"revisionTime": "2016-05-12T03:30:02Z"
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"checksumSHA1": "7oFpbmDfGobwKsFLIf6wMUvVoKw=",
|
"checksumSHA1": "7oFpbmDfGobwKsFLIf6wMUvVoKw=",
|
||||||
"path": "github.com/fatih/color",
|
"path": "github.com/fatih/color",
|
||||||
|
Loading…
Reference in New Issue
Block a user