mirror of
https://github.com/minio/minio.git
synced 2025-04-09 06:00:12 -04:00
parent
6d84e84b3c
commit
b38b9fea79
@ -23,7 +23,7 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
// ReadFile - decoded erasure coded file.
|
// ReadFile - decoded erasure coded file.
|
||||||
func (e erasure) ReadFile(volume, path string, startOffset int64) (io.ReadCloser, error) {
|
func (e erasure) ReadFile(volume, path string, startOffset int64, totalSize int64) (io.ReadCloser, error) {
|
||||||
// Input validation.
|
// Input validation.
|
||||||
if !isValidVolname(volume) {
|
if !isValidVolname(volume) {
|
||||||
return nil, errInvalidArgument
|
return nil, errInvalidArgument
|
||||||
@ -32,13 +32,13 @@ func (e erasure) ReadFile(volume, path string, startOffset int64) (io.ReadCloser
|
|||||||
return nil, errInvalidArgument
|
return nil, errInvalidArgument
|
||||||
}
|
}
|
||||||
|
|
||||||
var wg = &sync.WaitGroup{}
|
var rwg = &sync.WaitGroup{}
|
||||||
|
|
||||||
readers := make([]io.ReadCloser, len(e.storageDisks))
|
readers := make([]io.ReadCloser, len(e.storageDisks))
|
||||||
for index, disk := range e.storageDisks {
|
for index, disk := range e.storageDisks {
|
||||||
wg.Add(1)
|
rwg.Add(1)
|
||||||
go func(index int, disk StorageAPI) {
|
go func(index int, disk StorageAPI) {
|
||||||
defer wg.Done()
|
defer rwg.Done()
|
||||||
// If disk.ReadFile returns error and we don't have read
|
// If disk.ReadFile returns error and we don't have read
|
||||||
// quorum it will be taken care as ReedSolomon.Reconstruct()
|
// quorum it will be taken care as ReedSolomon.Reconstruct()
|
||||||
// will fail later.
|
// will fail later.
|
||||||
@ -49,18 +49,28 @@ func (e erasure) ReadFile(volume, path string, startOffset int64) (io.ReadCloser
|
|||||||
}(index, disk)
|
}(index, disk)
|
||||||
}
|
}
|
||||||
|
|
||||||
wg.Wait()
|
// Wait for all readers.
|
||||||
|
rwg.Wait()
|
||||||
|
|
||||||
// Initialize pipe.
|
// Initialize pipe.
|
||||||
pipeReader, pipeWriter := io.Pipe()
|
pipeReader, pipeWriter := io.Pipe()
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
|
var totalLeft = totalSize
|
||||||
// Read until EOF.
|
// Read until EOF.
|
||||||
for {
|
for totalLeft > 0 {
|
||||||
|
// Figure out the right blockSize as it was encoded
|
||||||
|
// before.
|
||||||
|
var curBlockSize int64
|
||||||
|
if erasureBlockSize < totalLeft {
|
||||||
|
curBlockSize = erasureBlockSize
|
||||||
|
} else {
|
||||||
|
curBlockSize = totalLeft
|
||||||
|
}
|
||||||
// Calculate the current encoded block size.
|
// Calculate the current encoded block size.
|
||||||
curEncBlockSize := getEncodedBlockLen(erasureBlockSize, e.DataBlocks)
|
curEncBlockSize := getEncodedBlockLen(curBlockSize, e.DataBlocks)
|
||||||
enBlocks := make([][]byte, len(e.storageDisks))
|
enBlocks := make([][]byte, len(e.storageDisks))
|
||||||
// Loop through all readers and read.
|
// Read all the readers.
|
||||||
for index, reader := range readers {
|
for index, reader := range readers {
|
||||||
// Initialize shard slice and fill the data from each parts.
|
// Initialize shard slice and fill the data from each parts.
|
||||||
enBlocks[index] = make([]byte, curEncBlockSize)
|
enBlocks[index] = make([]byte, curEncBlockSize)
|
||||||
@ -68,26 +78,11 @@ func (e erasure) ReadFile(volume, path string, startOffset int64) (io.ReadCloser
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
// Read the necessary blocks.
|
// Read the necessary blocks.
|
||||||
n, rErr := io.ReadFull(reader, enBlocks[index])
|
_, rErr := io.ReadFull(reader, enBlocks[index])
|
||||||
if rErr == io.EOF {
|
|
||||||
// Close the pipe.
|
|
||||||
pipeWriter.Close()
|
|
||||||
|
|
||||||
// Cleanly close all the underlying data readers.
|
|
||||||
for _, reader := range readers {
|
|
||||||
if reader == nil {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
reader.Close()
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if rErr != nil && rErr != io.ErrUnexpectedEOF {
|
if rErr != nil && rErr != io.ErrUnexpectedEOF {
|
||||||
readers[index].Close()
|
readers[index].Close()
|
||||||
readers[index] = nil
|
readers[index] = nil
|
||||||
continue
|
|
||||||
}
|
}
|
||||||
enBlocks[index] = enBlocks[index][:n]
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check blocks if they are all zero in length.
|
// Check blocks if they are all zero in length.
|
||||||
@ -131,17 +126,18 @@ func (e erasure) ReadFile(volume, path string, startOffset int64) (io.ReadCloser
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Get all the data blocks.
|
// Get all the data blocks.
|
||||||
dataBlocks := getDataBlocks(enBlocks, e.DataBlocks)
|
dataBlocks := getDataBlocks(enBlocks, e.DataBlocks, int(curBlockSize))
|
||||||
|
|
||||||
// Verify if the offset is right for the block, if not move to
|
// Verify if the offset is right for the block, if not move to the next block.
|
||||||
// the next block.
|
|
||||||
if startOffset > 0 {
|
if startOffset > 0 {
|
||||||
startOffset = startOffset - int64(len(dataBlocks))
|
startOffset = startOffset - int64(len(dataBlocks))
|
||||||
// Start offset is greater than or equal to zero, skip the dataBlocks.
|
// Start offset is greater than or equal to zero, skip the dataBlocks.
|
||||||
if startOffset >= 0 {
|
if startOffset >= 0 {
|
||||||
|
totalLeft = totalLeft - erasureBlockSize
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
// Now get back the remaining offset if startOffset is negative.
|
// Now get back the remaining offset if startOffset is
|
||||||
|
// negative.
|
||||||
startOffset = startOffset + int64(len(dataBlocks))
|
startOffset = startOffset + int64(len(dataBlocks))
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -154,6 +150,20 @@ func (e erasure) ReadFile(volume, path string, startOffset int64) (io.ReadCloser
|
|||||||
|
|
||||||
// Reset offset to '0' to read rest of the blocks.
|
// Reset offset to '0' to read rest of the blocks.
|
||||||
startOffset = int64(0)
|
startOffset = int64(0)
|
||||||
|
|
||||||
|
// Save what's left after reading erasureBlockSize.
|
||||||
|
totalLeft = totalLeft - erasureBlockSize
|
||||||
|
}
|
||||||
|
|
||||||
|
// Cleanly end the pipe after a successful decoding.
|
||||||
|
pipeWriter.Close()
|
||||||
|
|
||||||
|
// Cleanly close all the underlying data readers.
|
||||||
|
for _, reader := range readers {
|
||||||
|
if reader == nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
reader.Close()
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|
||||||
|
@ -17,19 +17,12 @@
|
|||||||
package main
|
package main
|
||||||
|
|
||||||
// getDataBlocks - fetches the data block only part of the input encoded blocks.
|
// getDataBlocks - fetches the data block only part of the input encoded blocks.
|
||||||
func getDataBlocks(enBlocks [][]byte, dataBlocks int) []byte {
|
func getDataBlocks(enBlocks [][]byte, dataBlocks int, curBlockSize int) []byte {
|
||||||
var data []byte
|
var data []byte
|
||||||
for _, block := range enBlocks[:dataBlocks] {
|
for _, block := range enBlocks[:dataBlocks] {
|
||||||
var newBlock []byte
|
data = append(data, block...)
|
||||||
// FIXME: Find a better way to skip the padding zeros.
|
|
||||||
for _, b := range block {
|
|
||||||
if b == 0 {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
newBlock = append(newBlock, b)
|
|
||||||
}
|
|
||||||
data = append(data, newBlock...)
|
|
||||||
}
|
}
|
||||||
|
data = data[:curBlockSize]
|
||||||
return data
|
return data
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -37,6 +37,8 @@ func (xl xlObjects) GetObject(bucket, object string, startOffset int64) (io.Read
|
|||||||
return nil, toObjectErr(err, bucket, object)
|
return nil, toObjectErr(err, bucket, object)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
totalObjectSize := xlMeta.Stat.Size // Total object size.
|
||||||
|
|
||||||
// Hold a read lock once more which can be released after the following go-routine ends.
|
// Hold a read lock once more which can be released after the following go-routine ends.
|
||||||
// We hold RLock once more because the current function would return before the go routine below
|
// We hold RLock once more because the current function would return before the go routine below
|
||||||
// executes and hence releasing the read lock (because of defer'ed nsMutex.RUnlock() call).
|
// executes and hence releasing the read lock (because of defer'ed nsMutex.RUnlock() call).
|
||||||
@ -45,7 +47,7 @@ func (xl xlObjects) GetObject(bucket, object string, startOffset int64) (io.Read
|
|||||||
defer nsMutex.RUnlock(bucket, object)
|
defer nsMutex.RUnlock(bucket, object)
|
||||||
for ; partIndex < len(xlMeta.Parts); partIndex++ {
|
for ; partIndex < len(xlMeta.Parts); partIndex++ {
|
||||||
part := xlMeta.Parts[partIndex]
|
part := xlMeta.Parts[partIndex]
|
||||||
r, err := xl.erasureDisk.ReadFile(bucket, pathJoin(object, part.Name), offset)
|
r, err := xl.erasureDisk.ReadFile(bucket, pathJoin(object, part.Name), offset, totalObjectSize)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
fileWriter.CloseWithError(err)
|
fileWriter.CloseWithError(err)
|
||||||
return
|
return
|
||||||
|
Loading…
x
Reference in New Issue
Block a user