mirror of
				https://github.com/minio/minio.git
				synced 2025-10-30 15:51:47 +01:00 
			
		
		
		
	
		
			
				
	
	
		
			335 lines
		
	
	
		
			9.4 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			335 lines
		
	
	
		
			9.4 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| /*
 | |
|  * Minio Cloud Storage, (C) 2016 Minio, Inc.
 | |
|  *
 | |
|  * Licensed under the Apache License, Version 2.0 (the "License");
 | |
|  * you may not use this file except in compliance with the License.
 | |
|  * You may obtain a copy of the License at
 | |
|  *
 | |
|  *     http://www.apache.org/licenses/LICENSE-2.0
 | |
|  *
 | |
|  * Unless required by applicable law or agreed to in writing, software
 | |
|  * distributed under the License is distributed on an "AS IS" BASIS,
 | |
|  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
|  * See the License for the specific language governing permissions and
 | |
|  * limitations under the License.
 | |
|  */
 | |
| 
 | |
| package cmd
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"reflect"
 | |
| 	"sort"
 | |
| 	"sync"
 | |
| 
 | |
| 	"github.com/minio/minio-go/pkg/policy"
 | |
| 	"github.com/minio/minio/cmd/logger"
 | |
| )
 | |
| 
 | |
| // list all errors that can be ignore in a bucket operation.
 | |
| var bucketOpIgnoredErrs = append(baseIgnoredErrs, errDiskAccessDenied)
 | |
| 
 | |
| // list all errors that can be ignored in a bucket metadata operation.
 | |
| var bucketMetadataOpIgnoredErrs = append(bucketOpIgnoredErrs, errVolumeNotFound)
 | |
| 
 | |
| /// Bucket operations
 | |
| 
 | |
| // MakeBucket - make a bucket.
 | |
| func (xl xlObjects) MakeBucketWithLocation(ctx context.Context, bucket, location string) error {
 | |
| 	// Verify if bucket is valid.
 | |
| 	if !IsValidBucketName(bucket) {
 | |
| 		logger.LogIf(ctx, BucketNameInvalid{Bucket: bucket})
 | |
| 		return BucketNameInvalid{Bucket: bucket}
 | |
| 	}
 | |
| 
 | |
| 	// Initialize sync waitgroup.
 | |
| 	var wg = &sync.WaitGroup{}
 | |
| 
 | |
| 	// Initialize list of errors.
 | |
| 	var dErrs = make([]error, len(xl.getDisks()))
 | |
| 
 | |
| 	// Make a volume entry on all underlying storage disks.
 | |
| 	for index, disk := range xl.getDisks() {
 | |
| 		if disk == nil {
 | |
| 			logger.LogIf(ctx, errDiskNotFound)
 | |
| 			dErrs[index] = errDiskNotFound
 | |
| 			continue
 | |
| 		}
 | |
| 		wg.Add(1)
 | |
| 		// Make a volume inside a go-routine.
 | |
| 		go func(index int, disk StorageAPI) {
 | |
| 			defer wg.Done()
 | |
| 			err := disk.MakeVol(bucket)
 | |
| 			if err != nil {
 | |
| 				if err != errVolumeExists {
 | |
| 					logger.LogIf(ctx, err)
 | |
| 				}
 | |
| 				dErrs[index] = err
 | |
| 			}
 | |
| 		}(index, disk)
 | |
| 	}
 | |
| 
 | |
| 	// Wait for all make vol to finish.
 | |
| 	wg.Wait()
 | |
| 
 | |
| 	writeQuorum := len(xl.getDisks())/2 + 1
 | |
| 	err := reduceWriteQuorumErrs(ctx, dErrs, bucketOpIgnoredErrs, writeQuorum)
 | |
| 	if err == errXLWriteQuorum {
 | |
| 		// Purge successfully created buckets if we don't have writeQuorum.
 | |
| 		undoMakeBucket(xl.getDisks(), bucket)
 | |
| 	}
 | |
| 	return toObjectErr(err, bucket)
 | |
| }
 | |
| 
 | |
| func (xl xlObjects) undoDeleteBucket(bucket string) {
 | |
| 	// Initialize sync waitgroup.
 | |
| 	var wg = &sync.WaitGroup{}
 | |
| 	// Undo previous make bucket entry on all underlying storage disks.
 | |
| 	for index, disk := range xl.getDisks() {
 | |
| 		if disk == nil {
 | |
| 			continue
 | |
| 		}
 | |
| 		wg.Add(1)
 | |
| 		// Delete a bucket inside a go-routine.
 | |
| 		go func(index int, disk StorageAPI) {
 | |
| 			defer wg.Done()
 | |
| 			_ = disk.MakeVol(bucket)
 | |
| 		}(index, disk)
 | |
| 	}
 | |
| 
 | |
| 	// Wait for all make vol to finish.
 | |
| 	wg.Wait()
 | |
| }
 | |
| 
 | |
| // undo make bucket operation upon quorum failure.
 | |
| func undoMakeBucket(storageDisks []StorageAPI, bucket string) {
 | |
| 	// Initialize sync waitgroup.
 | |
| 	var wg = &sync.WaitGroup{}
 | |
| 	// Undo previous make bucket entry on all underlying storage disks.
 | |
| 	for index, disk := range storageDisks {
 | |
| 		if disk == nil {
 | |
| 			continue
 | |
| 		}
 | |
| 		wg.Add(1)
 | |
| 		// Delete a bucket inside a go-routine.
 | |
| 		go func(index int, disk StorageAPI) {
 | |
| 			defer wg.Done()
 | |
| 			_ = disk.DeleteVol(bucket)
 | |
| 		}(index, disk)
 | |
| 	}
 | |
| 
 | |
| 	// Wait for all make vol to finish.
 | |
| 	wg.Wait()
 | |
| }
 | |
| 
 | |
| // getBucketInfo - returns the BucketInfo from one of the load balanced disks.
 | |
| func (xl xlObjects) getBucketInfo(ctx context.Context, bucketName string) (bucketInfo BucketInfo, err error) {
 | |
| 	var bucketErrs []error
 | |
| 	for _, disk := range xl.getLoadBalancedDisks() {
 | |
| 		if disk == nil {
 | |
| 			bucketErrs = append(bucketErrs, errDiskNotFound)
 | |
| 			continue
 | |
| 		}
 | |
| 		volInfo, serr := disk.StatVol(bucketName)
 | |
| 		if serr == nil {
 | |
| 			bucketInfo = BucketInfo{
 | |
| 				Name:    volInfo.Name,
 | |
| 				Created: volInfo.Created,
 | |
| 			}
 | |
| 			return bucketInfo, nil
 | |
| 		}
 | |
| 		logger.LogIf(ctx, serr)
 | |
| 		err = serr
 | |
| 		// For any reason disk went offline continue and pick the next one.
 | |
| 		if IsErrIgnored(err, bucketMetadataOpIgnoredErrs...) {
 | |
| 			bucketErrs = append(bucketErrs, err)
 | |
| 			continue
 | |
| 		}
 | |
| 		// Any error which cannot be ignored, we return quickly.
 | |
| 		return BucketInfo{}, err
 | |
| 	}
 | |
| 	// If all our errors were ignored, then we try to
 | |
| 	// reduce to one error based on read quorum.
 | |
| 	// `nil` is deliberately passed for ignoredErrs
 | |
| 	// because these errors were already ignored.
 | |
| 	readQuorum := len(xl.getDisks()) / 2
 | |
| 	return BucketInfo{}, reduceReadQuorumErrs(ctx, bucketErrs, nil, readQuorum)
 | |
| }
 | |
| 
 | |
| // GetBucketInfo - returns BucketInfo for a bucket.
 | |
| func (xl xlObjects) GetBucketInfo(ctx context.Context, bucket string) (bi BucketInfo, e error) {
 | |
| 	bucketLock := xl.nsMutex.NewNSLock(bucket, "")
 | |
| 	if e := bucketLock.GetRLock(globalObjectTimeout); e != nil {
 | |
| 		return bi, e
 | |
| 	}
 | |
| 	defer bucketLock.RUnlock()
 | |
| 	// Verify if bucket is valid.
 | |
| 	if !IsValidBucketName(bucket) {
 | |
| 		return bi, BucketNameInvalid{Bucket: bucket}
 | |
| 	}
 | |
| 
 | |
| 	bucketInfo, err := xl.getBucketInfo(ctx, bucket)
 | |
| 	if err != nil {
 | |
| 		return bi, toObjectErr(err, bucket)
 | |
| 	}
 | |
| 	return bucketInfo, nil
 | |
| }
 | |
| 
 | |
| // listBuckets - returns list of all buckets from a disk picked at random.
 | |
| func (xl xlObjects) listBuckets(ctx context.Context) (bucketsInfo []BucketInfo, err error) {
 | |
| 	for _, disk := range xl.getLoadBalancedDisks() {
 | |
| 		if disk == nil {
 | |
| 			continue
 | |
| 		}
 | |
| 		var volsInfo []VolInfo
 | |
| 		volsInfo, err = disk.ListVols()
 | |
| 		if err == nil {
 | |
| 			// NOTE: The assumption here is that volumes across all disks in
 | |
| 			// readQuorum have consistent view i.e they all have same number
 | |
| 			// of buckets. This is essentially not verified since healing
 | |
| 			// should take care of this.
 | |
| 			var bucketsInfo []BucketInfo
 | |
| 			for _, volInfo := range volsInfo {
 | |
| 				if isReservedOrInvalidBucket(volInfo.Name) {
 | |
| 					continue
 | |
| 				}
 | |
| 				bucketsInfo = append(bucketsInfo, BucketInfo{
 | |
| 					Name:    volInfo.Name,
 | |
| 					Created: volInfo.Created,
 | |
| 				})
 | |
| 			}
 | |
| 			// For buckets info empty, loop once again to check
 | |
| 			// if we have, can happen if disks were down.
 | |
| 			if len(bucketsInfo) == 0 {
 | |
| 				continue
 | |
| 			}
 | |
| 			return bucketsInfo, nil
 | |
| 		}
 | |
| 		logger.LogIf(ctx, err)
 | |
| 		// Ignore any disks not found.
 | |
| 		if IsErrIgnored(err, bucketMetadataOpIgnoredErrs...) {
 | |
| 			continue
 | |
| 		}
 | |
| 		break
 | |
| 	}
 | |
| 	return nil, err
 | |
| }
 | |
| 
 | |
| // ListBuckets - lists all the buckets, sorted by its name.
 | |
| func (xl xlObjects) ListBuckets(ctx context.Context) ([]BucketInfo, error) {
 | |
| 	bucketInfos, err := xl.listBuckets(ctx)
 | |
| 	if err != nil {
 | |
| 		return nil, toObjectErr(err)
 | |
| 	}
 | |
| 	// Sort by bucket name before returning.
 | |
| 	sort.Sort(byBucketName(bucketInfos))
 | |
| 	return bucketInfos, nil
 | |
| }
 | |
| 
 | |
| // DeleteBucket - deletes a bucket.
 | |
| func (xl xlObjects) DeleteBucket(ctx context.Context, bucket string) error {
 | |
| 
 | |
| 	bucketLock := xl.nsMutex.NewNSLock(bucket, "")
 | |
| 	if err := bucketLock.GetLock(globalObjectTimeout); err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	defer bucketLock.Unlock()
 | |
| 
 | |
| 	// Verify if bucket is valid.
 | |
| 	if !IsValidBucketName(bucket) {
 | |
| 		return BucketNameInvalid{Bucket: bucket}
 | |
| 	}
 | |
| 
 | |
| 	// Collect if all disks report volume not found.
 | |
| 	var wg = &sync.WaitGroup{}
 | |
| 	var dErrs = make([]error, len(xl.getDisks()))
 | |
| 
 | |
| 	// Remove a volume entry on all underlying storage disks.
 | |
| 	for index, disk := range xl.getDisks() {
 | |
| 		if disk == nil {
 | |
| 			logger.LogIf(ctx, errDiskNotFound)
 | |
| 			dErrs[index] = errDiskNotFound
 | |
| 			continue
 | |
| 		}
 | |
| 		wg.Add(1)
 | |
| 		// Delete volume inside a go-routine.
 | |
| 		go func(index int, disk StorageAPI) {
 | |
| 			defer wg.Done()
 | |
| 			// Attempt to delete bucket.
 | |
| 			err := disk.DeleteVol(bucket)
 | |
| 
 | |
| 			if err != nil {
 | |
| 				logger.LogIf(ctx, err)
 | |
| 				dErrs[index] = err
 | |
| 				return
 | |
| 			}
 | |
| 			// Cleanup all the previously incomplete multiparts.
 | |
| 			err = cleanupDir(ctx, disk, minioMetaMultipartBucket, bucket)
 | |
| 
 | |
| 			if err != nil {
 | |
| 				if err == errVolumeNotFound {
 | |
| 					return
 | |
| 				}
 | |
| 				dErrs[index] = err
 | |
| 			}
 | |
| 		}(index, disk)
 | |
| 	}
 | |
| 
 | |
| 	// Wait for all the delete vols to finish.
 | |
| 	wg.Wait()
 | |
| 
 | |
| 	writeQuorum := len(xl.getDisks())/2 + 1
 | |
| 	err := reduceWriteQuorumErrs(ctx, dErrs, bucketOpIgnoredErrs, writeQuorum)
 | |
| 	if err == errXLWriteQuorum {
 | |
| 		xl.undoDeleteBucket(bucket)
 | |
| 	}
 | |
| 	if err != nil {
 | |
| 		return toObjectErr(err, bucket)
 | |
| 	}
 | |
| 
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // SetBucketPolicy sets policy on bucket
 | |
| func (xl xlObjects) SetBucketPolicy(ctx context.Context, bucket string, policy policy.BucketAccessPolicy) error {
 | |
| 	return persistAndNotifyBucketPolicyChange(ctx, bucket, false, policy, xl)
 | |
| }
 | |
| 
 | |
| // GetBucketPolicy will get policy on bucket
 | |
| func (xl xlObjects) GetBucketPolicy(ctx context.Context, bucket string) (policy.BucketAccessPolicy, error) {
 | |
| 	// fetch bucket policy from cache.
 | |
| 	bpolicy := xl.bucketPolicies.GetBucketPolicy(bucket)
 | |
| 	if reflect.DeepEqual(bpolicy, emptyBucketPolicy) {
 | |
| 		return ReadBucketPolicy(bucket, xl)
 | |
| 	}
 | |
| 	return bpolicy, nil
 | |
| }
 | |
| 
 | |
| // DeleteBucketPolicy deletes all policies on bucket
 | |
| func (xl xlObjects) DeleteBucketPolicy(ctx context.Context, bucket string) error {
 | |
| 	return persistAndNotifyBucketPolicyChange(ctx, bucket, true, emptyBucketPolicy, xl)
 | |
| }
 | |
| 
 | |
| // RefreshBucketPolicy refreshes policy cache from disk
 | |
| func (xl xlObjects) RefreshBucketPolicy(ctx context.Context, bucket string) error {
 | |
| 	policy, err := ReadBucketPolicy(bucket, xl)
 | |
| 
 | |
| 	if err != nil {
 | |
| 		if reflect.DeepEqual(policy, emptyBucketPolicy) {
 | |
| 			return xl.bucketPolicies.DeleteBucketPolicy(bucket)
 | |
| 		}
 | |
| 		return err
 | |
| 	}
 | |
| 	return xl.bucketPolicies.SetBucketPolicy(bucket, policy)
 | |
| }
 | |
| 
 | |
| // IsNotificationSupported returns whether bucket notification is applicable for this layer.
 | |
| func (xl xlObjects) IsNotificationSupported() bool {
 | |
| 	return true
 | |
| }
 | |
| 
 | |
| // IsEncryptionSupported returns whether server side encryption is applicable for this layer.
 | |
| func (xl xlObjects) IsEncryptionSupported() bool {
 | |
| 	return true
 | |
| }
 |