2018-06-20 15:06:27 +03:00
|
|
|
// Copyright 2016 The go-ethereum Authors
|
|
|
|
// This file is part of the go-ethereum library.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Lesser General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
package storage
|
|
|
|
|
|
|
|
import (
|
2018-07-09 15:11:49 +03:00
|
|
|
"context"
|
2018-06-20 15:06:27 +03:00
|
|
|
"io"
|
2019-02-06 14:16:43 +03:00
|
|
|
"sort"
|
2019-02-13 02:17:44 +03:00
|
|
|
"sync"
|
2018-06-20 15:06:27 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
/*
|
|
|
|
FileStore provides the client API entrypoints Store and Retrieve to store and retrieve
|
|
|
|
It can store anything that has a byte slice representation, so files or serialised objects etc.
|
|
|
|
|
|
|
|
Storage: FileStore calls the Chunker to segment the input datastream of any size to a merkle hashed tree of chunks. The key of the root block is returned to the client.
|
|
|
|
|
|
|
|
Retrieval: given the key of the root block, the FileStore retrieves the block chunks and reconstructs the original data and passes it back as a lazy reader. A lazy reader is a reader with on-demand delayed processing, i.e. the chunks needed to reconstruct a large file are only fetched and processed if that particular part of the document is actually read.
|
|
|
|
|
|
|
|
As the chunker produces chunks, FileStore dispatches them to its own chunk store
|
|
|
|
implementation for storage or retrieval.
|
|
|
|
*/
|
|
|
|
|
|
|
|
const (
|
|
|
|
defaultLDBCapacity = 5000000 // capacity for LevelDB, by default 5*10^6*4096 bytes == 20GB
|
|
|
|
defaultCacheCapacity = 10000 // capacity for in-memory chunks' cache
|
|
|
|
defaultChunkRequestsCacheCapacity = 5000000 // capacity for container holding outgoing requests for chunks. should be set to LevelDB capacity
|
|
|
|
)
|
|
|
|
|
|
|
|
type FileStore struct {
|
|
|
|
ChunkStore
|
|
|
|
hashFunc SwarmHasher
|
|
|
|
}
|
|
|
|
|
|
|
|
type FileStoreParams struct {
|
|
|
|
Hash string
|
|
|
|
}
|
|
|
|
|
|
|
|
func NewFileStoreParams() *FileStoreParams {
|
|
|
|
return &FileStoreParams{
|
|
|
|
Hash: DefaultHash,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// for testing locally
|
|
|
|
func NewLocalFileStore(datadir string, basekey []byte) (*FileStore, error) {
|
|
|
|
params := NewDefaultLocalStoreParams()
|
|
|
|
params.Init(datadir)
|
|
|
|
localStore, err := NewLocalStore(params, nil)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
localStore.Validators = append(localStore.Validators, NewContentAddressValidator(MakeHashFunc(DefaultHash)))
|
|
|
|
return NewFileStore(localStore, NewFileStoreParams()), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func NewFileStore(store ChunkStore, params *FileStoreParams) *FileStore {
|
|
|
|
hashFunc := MakeHashFunc(params.Hash)
|
|
|
|
return &FileStore{
|
|
|
|
ChunkStore: store,
|
|
|
|
hashFunc: hashFunc,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-02-07 15:51:24 +03:00
|
|
|
// Retrieve is a public API. Main entry point for document retrieval directly. Used by the
|
2018-06-20 15:06:27 +03:00
|
|
|
// FS-aware API and httpaccess
|
|
|
|
// Chunk retrieval blocks on netStore requests with a timeout so reader will
|
|
|
|
// report error if retrieval of chunks within requested range time out.
|
|
|
|
// It returns a reader with the chunk data and whether the content was encrypted
|
2018-07-09 15:11:49 +03:00
|
|
|
func (f *FileStore) Retrieve(ctx context.Context, addr Address) (reader *LazyChunkReader, isEncrypted bool) {
|
2018-06-20 15:06:27 +03:00
|
|
|
isEncrypted = len(addr) > f.hashFunc().Size()
|
|
|
|
getter := NewHasherStore(f.ChunkStore, f.hashFunc, isEncrypted)
|
2018-07-09 15:11:49 +03:00
|
|
|
reader = TreeJoin(ctx, addr, getter, 0)
|
2018-06-20 15:06:27 +03:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2019-02-07 15:51:24 +03:00
|
|
|
// Store is a public API. Main entry point for document storage directly. Used by the
|
2018-06-20 15:06:27 +03:00
|
|
|
// FS-aware API and httpaccess
|
2018-07-09 15:11:49 +03:00
|
|
|
func (f *FileStore) Store(ctx context.Context, data io.Reader, size int64, toEncrypt bool) (addr Address, wait func(context.Context) error, err error) {
|
2018-06-20 15:06:27 +03:00
|
|
|
putter := NewHasherStore(f.ChunkStore, f.hashFunc, toEncrypt)
|
2018-07-09 15:11:49 +03:00
|
|
|
return PyramidSplit(ctx, data, putter, putter)
|
2018-06-20 15:06:27 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
func (f *FileStore) HashSize() int {
|
|
|
|
return f.hashFunc().Size()
|
|
|
|
}
|
2019-02-06 14:16:43 +03:00
|
|
|
|
2019-02-07 15:51:24 +03:00
|
|
|
// GetAllReferences is a public API. This endpoint returns all chunk hashes (only) for a given file
|
2019-02-06 14:16:43 +03:00
|
|
|
func (f *FileStore) GetAllReferences(ctx context.Context, data io.Reader, toEncrypt bool) (addrs AddressCollection, err error) {
|
|
|
|
// create a special kind of putter, which only will store the references
|
2019-02-13 02:17:44 +03:00
|
|
|
putter := &hashExplorer{
|
2019-02-06 14:16:43 +03:00
|
|
|
hasherStore: NewHasherStore(f.ChunkStore, f.hashFunc, toEncrypt),
|
|
|
|
}
|
|
|
|
// do the actual splitting anyway, no way around it
|
2019-02-13 02:17:44 +03:00
|
|
|
_, wait, err := PyramidSplit(ctx, data, putter, putter)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
// wait for splitting to be complete and all chunks processed
|
|
|
|
err = wait(ctx)
|
2019-02-06 14:16:43 +03:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
// collect all references
|
|
|
|
addrs = NewAddressCollection(0)
|
2019-02-13 02:17:44 +03:00
|
|
|
for _, ref := range putter.references {
|
2019-02-06 14:16:43 +03:00
|
|
|
addrs = append(addrs, Address(ref))
|
|
|
|
}
|
|
|
|
sort.Sort(addrs)
|
|
|
|
return addrs, nil
|
|
|
|
}
|
|
|
|
|
2019-02-13 02:17:44 +03:00
|
|
|
// hashExplorer is a special kind of putter which will only store chunk references
|
|
|
|
type hashExplorer struct {
|
2019-02-06 14:16:43 +03:00
|
|
|
*hasherStore
|
2019-02-13 02:17:44 +03:00
|
|
|
references []Reference
|
|
|
|
lock sync.Mutex
|
2019-02-06 14:16:43 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// HashExplorer's Put will add just the chunk hashes to its `References`
|
2019-02-13 02:17:44 +03:00
|
|
|
func (he *hashExplorer) Put(ctx context.Context, chunkData ChunkData) (Reference, error) {
|
2019-02-06 14:16:43 +03:00
|
|
|
// Need to do the actual Put, which returns the references
|
|
|
|
ref, err := he.hasherStore.Put(ctx, chunkData)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
// internally store the reference
|
2019-02-13 02:17:44 +03:00
|
|
|
he.lock.Lock()
|
|
|
|
he.references = append(he.references, ref)
|
|
|
|
he.lock.Unlock()
|
2019-02-06 14:16:43 +03:00
|
|
|
return ref, nil
|
|
|
|
}
|