mirror of
https://github.com/minio/minio.git
synced 2024-12-26 07:05:55 -05:00
f8e13fb00e
Fixes #2971 - honors ignore-disks option properly. Fixes #2969 - change the net.Dial to have a timeout of 3secs.
263 lines
6.7 KiB
Go
263 lines
6.7 KiB
Go
/*
|
|
* Minio Cloud Storage, (C) 2016 Minio, Inc.
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
package cmd
|
|
|
|
import (
|
|
"net"
|
|
"strings"
|
|
"sync"
|
|
)
|
|
|
|
const (
|
|
// Block size used for all internal operations version 1.
|
|
blockSizeV1 = 10 * 1024 * 1024 // 10MiB.
|
|
|
|
// Staging buffer read size for all internal operations version 1.
|
|
readSizeV1 = 128 * 1024 // 128KiB.
|
|
|
|
// Buckets meta prefix.
|
|
bucketMetaPrefix = "buckets"
|
|
)
|
|
|
|
// Global object layer mutex, used for safely updating object layer.
|
|
var globalObjLayerMutex *sync.Mutex
|
|
|
|
// Global object layer, only accessed by newObjectLayerFn().
|
|
var globalObjectAPI ObjectLayer
|
|
|
|
func init() {
|
|
// Initialize this once per server initialization.
|
|
globalObjLayerMutex = &sync.Mutex{}
|
|
}
|
|
|
|
// isErrIgnored should we ignore this error?, takes a list of errors which can be ignored.
|
|
func isErrIgnored(err error, ignoredErrs []error) bool {
|
|
err = errorCause(err)
|
|
for _, ignoredErr := range ignoredErrs {
|
|
if ignoredErr == err {
|
|
return true
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
// House keeping code needed for FS.
|
|
func fsHouseKeeping(storageDisk StorageAPI) error {
|
|
// Cleanup all temp entries upon start.
|
|
err := cleanupDir(storageDisk, minioMetaBucket, tmpMetaPrefix)
|
|
if err != nil {
|
|
return toObjectErr(err, minioMetaBucket, tmpMetaPrefix)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Check if a network path is local to this node.
|
|
func isLocalStorage(networkPath string) bool {
|
|
if idx := strings.LastIndex(networkPath, ":"); idx != -1 {
|
|
// e.g 10.0.0.1:/mnt/networkPath
|
|
netAddr, _, err := splitNetPath(networkPath)
|
|
if err != nil {
|
|
errorIf(err, "Splitting into ip and path failed")
|
|
return false
|
|
}
|
|
// netAddr will only be set if this is not a local path.
|
|
if netAddr == "" {
|
|
return true
|
|
}
|
|
// Resolve host to address to check if the IP is loopback.
|
|
// If address resolution fails, assume it's a non-local host.
|
|
addrs, err := net.LookupHost(netAddr)
|
|
if err != nil {
|
|
errorIf(err, "Failed to lookup host")
|
|
return false
|
|
}
|
|
for _, addr := range addrs {
|
|
if ip := net.ParseIP(addr); ip.IsLoopback() {
|
|
return true
|
|
}
|
|
}
|
|
iaddrs, err := net.InterfaceAddrs()
|
|
if err != nil {
|
|
errorIf(err, "Unable to list interface addresses")
|
|
return false
|
|
}
|
|
for _, addr := range addrs {
|
|
for _, iaddr := range iaddrs {
|
|
ip, _, err := net.ParseCIDR(iaddr.String())
|
|
if err != nil {
|
|
errorIf(err, "Unable to parse CIDR")
|
|
return false
|
|
}
|
|
if ip.String() == addr {
|
|
return true
|
|
}
|
|
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
return true
|
|
}
|
|
|
|
// Depending on the disk type network or local, initialize storage API.
|
|
func newStorageAPI(disk string) (storage StorageAPI, err error) {
|
|
if isLocalStorage(disk) {
|
|
return newPosix(disk)
|
|
}
|
|
return newRPCClient(disk)
|
|
}
|
|
|
|
// Initializes meta volume on all input storage disks.
|
|
func initMetaVolume(storageDisks []StorageAPI) error {
|
|
// This happens for the first time, but keep this here since this
|
|
// is the only place where it can be made expensive optimizing all
|
|
// other calls. Create minio meta volume, if it doesn't exist yet.
|
|
var wg = &sync.WaitGroup{}
|
|
|
|
// Initialize errs to collect errors inside go-routine.
|
|
var errs = make([]error, len(storageDisks))
|
|
|
|
// Initialize all disks in parallel.
|
|
for index, disk := range storageDisks {
|
|
if disk == nil {
|
|
// Ignore create meta volume on disks which are not found.
|
|
continue
|
|
}
|
|
wg.Add(1)
|
|
go func(index int, disk StorageAPI) {
|
|
// Indicate this wait group is done.
|
|
defer wg.Done()
|
|
|
|
// Attempt to create `.minio.sys`.
|
|
err := disk.MakeVol(minioMetaBucket)
|
|
if err != nil {
|
|
switch err {
|
|
// Ignored errors.
|
|
case errVolumeExists, errDiskNotFound, errFaultyDisk:
|
|
default:
|
|
errs[index] = err
|
|
}
|
|
}
|
|
}(index, disk)
|
|
}
|
|
|
|
// Wait for all cleanup to finish.
|
|
wg.Wait()
|
|
|
|
// Return upon first error.
|
|
for _, err := range errs {
|
|
if err == nil {
|
|
continue
|
|
}
|
|
return toObjectErr(err, minioMetaBucket)
|
|
}
|
|
|
|
// Return success here.
|
|
return nil
|
|
}
|
|
|
|
// House keeping code needed for XL.
|
|
func xlHouseKeeping(storageDisks []StorageAPI) error {
|
|
// This happens for the first time, but keep this here since this
|
|
// is the only place where it can be made expensive optimizing all
|
|
// other calls. Create metavolume.
|
|
var wg = &sync.WaitGroup{}
|
|
|
|
// Initialize errs to collect errors inside go-routine.
|
|
var errs = make([]error, len(storageDisks))
|
|
|
|
// Initialize all disks in parallel.
|
|
for index, disk := range storageDisks {
|
|
if disk == nil {
|
|
continue
|
|
}
|
|
wg.Add(1)
|
|
go func(index int, disk StorageAPI) {
|
|
// Indicate this wait group is done.
|
|
defer wg.Done()
|
|
|
|
// Cleanup all temp entries upon start.
|
|
err := cleanupDir(disk, minioMetaBucket, tmpMetaPrefix)
|
|
if err != nil {
|
|
switch errorCause(err) {
|
|
case errDiskNotFound, errVolumeNotFound, errFileNotFound:
|
|
default:
|
|
errs[index] = err
|
|
}
|
|
}
|
|
}(index, disk)
|
|
}
|
|
|
|
// Wait for all cleanup to finish.
|
|
wg.Wait()
|
|
|
|
// Return upon first error.
|
|
for _, err := range errs {
|
|
if err == nil {
|
|
continue
|
|
}
|
|
return toObjectErr(err, minioMetaBucket, tmpMetaPrefix)
|
|
}
|
|
|
|
// Return success here.
|
|
return nil
|
|
}
|
|
|
|
// Cleanup a directory recursively.
|
|
func cleanupDir(storage StorageAPI, volume, dirPath string) error {
|
|
var delFunc func(string) error
|
|
// Function to delete entries recursively.
|
|
delFunc = func(entryPath string) error {
|
|
if !strings.HasSuffix(entryPath, slashSeparator) {
|
|
// Delete the file entry.
|
|
err := storage.DeleteFile(volume, entryPath)
|
|
return traceError(err)
|
|
}
|
|
|
|
// If it's a directory, list and call delFunc() for each entry.
|
|
entries, err := storage.ListDir(volume, entryPath)
|
|
// If entryPath prefix never existed, safe to ignore.
|
|
if err == errFileNotFound {
|
|
return nil
|
|
} else if err != nil { // For any other errors fail.
|
|
return traceError(err)
|
|
} // else on success..
|
|
|
|
// Recurse and delete all other entries.
|
|
for _, entry := range entries {
|
|
if err = delFunc(pathJoin(entryPath, entry)); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
err := delFunc(retainSlash(pathJoin(dirPath)))
|
|
return err
|
|
}
|
|
|
|
// Checks whether bucket exists.
|
|
func isBucketExist(bucket string, obj ObjectLayer) error {
|
|
if !IsValidBucketName(bucket) {
|
|
return BucketNameInvalid{Bucket: bucket}
|
|
}
|
|
_, err := obj.GetBucketInfo(bucket)
|
|
if err != nil {
|
|
return BucketNotFound{Bucket: bucket}
|
|
}
|
|
return nil
|
|
}
|