mirror of
				https://github.com/minio/minio.git
				synced 2025-10-31 08:11:19 +01:00 
			
		
		
		
	- using miniogo.ObjectInfo.UserMetadata is not correct - using UserTags from Map->String() can change order - ContentType comparison needs to be removed. - Compare both lowercase and uppercase key names. - do not silently error out constructing PutObjectOptions if tag parsing fails - avoid notification for empty object info, failed operations should rely on valid objInfo for notification in all situations - optimize copyObject implementation, also introduce a new replication event - clone ObjectInfo() before scheduling for replication - add additional headers for comparison - remove strings.EqualFold comparison avoid unexpected bugs - fix pool based proxying with multiple pools - compare only specific metadata Co-authored-by: Poorna Krishnamoorthy <poornas@users.noreply.github.com>
		
			
				
	
	
		
			1276 lines
		
	
	
		
			42 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			1276 lines
		
	
	
		
			42 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| /*
 | |
|  * MinIO Cloud Storage, (C) 2016-2020 MinIO, Inc.
 | |
|  *
 | |
|  * Licensed under the Apache License, Version 2.0 (the "License");
 | |
|  * you may not use this file except in compliance with the License.
 | |
|  * You may obtain a copy of the License at
 | |
|  *
 | |
|  *     http://www.apache.org/licenses/LICENSE-2.0
 | |
|  *
 | |
|  * Unless required by applicable law or agreed to in writing, software
 | |
|  * distributed under the License is distributed on an "AS IS" BASIS,
 | |
|  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
|  * See the License for the specific language governing permissions and
 | |
|  * limitations under the License.
 | |
|  */
 | |
| 
 | |
| package cmd
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"errors"
 | |
| 	"fmt"
 | |
| 	"io"
 | |
| 	"net/http"
 | |
| 	"path"
 | |
| 	"strings"
 | |
| 	"sync"
 | |
| 	"time"
 | |
| 
 | |
| 	"github.com/minio/minio-go/v7/pkg/tags"
 | |
| 	xhttp "github.com/minio/minio/cmd/http"
 | |
| 	"github.com/minio/minio/cmd/logger"
 | |
| 	"github.com/minio/minio/pkg/bucket/lifecycle"
 | |
| 	"github.com/minio/minio/pkg/bucket/replication"
 | |
| 	"github.com/minio/minio/pkg/madmin"
 | |
| 	"github.com/minio/minio/pkg/mimedb"
 | |
| 	"github.com/minio/minio/pkg/sync/errgroup"
 | |
| )
 | |
| 
 | |
| // list all errors which can be ignored in object operations.
 | |
| var objectOpIgnoredErrs = append(baseIgnoredErrs, errDiskAccessDenied, errUnformattedDisk)
 | |
| 
 | |
| /// Object Operations
 | |
| 
 | |
| // CopyObject - copy object source object to destination object.
 | |
| // if source object and destination object are same we only
 | |
| // update metadata.
 | |
| func (er erasureObjects) CopyObject(ctx context.Context, srcBucket, srcObject, dstBucket, dstObject string, srcInfo ObjectInfo, srcOpts, dstOpts ObjectOptions) (oi ObjectInfo, e error) {
 | |
| 	// This call shouldn't be used for anything other than metadata updates or adding self referential versions.
 | |
| 	if !srcInfo.metadataOnly {
 | |
| 		return oi, NotImplemented{}
 | |
| 	}
 | |
| 
 | |
| 	defer ObjectPathUpdated(pathJoin(dstBucket, dstObject))
 | |
| 
 | |
| 	lk := er.NewNSLock(dstBucket, dstObject)
 | |
| 	if err := lk.GetLock(ctx, globalOperationTimeout); err != nil {
 | |
| 		return oi, err
 | |
| 	}
 | |
| 	defer lk.Unlock()
 | |
| 
 | |
| 	// Read metadata associated with the object from all disks.
 | |
| 	storageDisks := er.getDisks()
 | |
| 	metaArr, errs := readAllFileInfo(ctx, storageDisks, srcBucket, srcObject, srcOpts.VersionID, false)
 | |
| 
 | |
| 	// get Quorum for this object
 | |
| 	readQuorum, writeQuorum, err := objectQuorumFromMeta(ctx, metaArr, errs, er.defaultParityCount)
 | |
| 	if err != nil {
 | |
| 		return oi, toObjectErr(err, srcBucket, srcObject)
 | |
| 	}
 | |
| 
 | |
| 	// List all online disks.
 | |
| 	onlineDisks, modTime := listOnlineDisks(storageDisks, metaArr, errs)
 | |
| 
 | |
| 	// Pick latest valid metadata.
 | |
| 	fi, err := pickValidFileInfo(ctx, metaArr, modTime, readQuorum)
 | |
| 	if err != nil {
 | |
| 		return oi, toObjectErr(err, srcBucket, srcObject)
 | |
| 	}
 | |
| 	if fi.Deleted {
 | |
| 		if srcOpts.VersionID == "" {
 | |
| 			return oi, toObjectErr(errFileNotFound, srcBucket, srcObject)
 | |
| 		}
 | |
| 		return fi.ToObjectInfo(srcBucket, srcObject), toObjectErr(errMethodNotAllowed, srcBucket, srcObject)
 | |
| 	}
 | |
| 
 | |
| 	onlineDisks, metaArr = shuffleDisksAndPartsMetadataByIndex(onlineDisks, metaArr, fi.Erasure.Distribution)
 | |
| 
 | |
| 	versionID := srcInfo.VersionID
 | |
| 	if srcInfo.versionOnly {
 | |
| 		versionID = dstOpts.VersionID
 | |
| 		// preserve destination versionId if specified.
 | |
| 		if versionID == "" {
 | |
| 			versionID = mustGetUUID()
 | |
| 		}
 | |
| 		modTime = UTCNow()
 | |
| 	}
 | |
| 	fi.VersionID = versionID // set any new versionID we might have created
 | |
| 	fi.ModTime = modTime     // set modTime for the new versionID
 | |
| 	if !dstOpts.MTime.IsZero() {
 | |
| 		modTime = dstOpts.MTime
 | |
| 		fi.ModTime = dstOpts.MTime
 | |
| 	}
 | |
| 	fi.Metadata = srcInfo.UserDefined
 | |
| 	srcInfo.UserDefined["etag"] = srcInfo.ETag
 | |
| 
 | |
| 	// Update `xl.meta` content on each disks.
 | |
| 	for index := range metaArr {
 | |
| 		if metaArr[index].IsValid() {
 | |
| 			metaArr[index].ModTime = modTime
 | |
| 			metaArr[index].VersionID = versionID
 | |
| 			metaArr[index].Metadata = srcInfo.UserDefined
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	tempObj := mustGetUUID()
 | |
| 
 | |
| 	// Cleanup in case of xl.meta writing failure
 | |
| 	defer er.deleteObject(context.Background(), minioMetaTmpBucket, tempObj, writeQuorum)
 | |
| 
 | |
| 	// Write unique `xl.meta` for each disk.
 | |
| 	if onlineDisks, err = writeUniqueFileInfo(ctx, onlineDisks, minioMetaTmpBucket, tempObj, metaArr, writeQuorum); err != nil {
 | |
| 		return oi, toObjectErr(err, srcBucket, srcObject)
 | |
| 	}
 | |
| 
 | |
| 	// Rename atomically `xl.meta` from tmp location to destination for each disk.
 | |
| 	if _, err = renameFileInfo(ctx, onlineDisks, minioMetaTmpBucket, tempObj, srcBucket, srcObject, writeQuorum); err != nil {
 | |
| 		return oi, toObjectErr(err, srcBucket, srcObject)
 | |
| 	}
 | |
| 
 | |
| 	return fi.ToObjectInfo(srcBucket, srcObject), nil
 | |
| }
 | |
| 
 | |
| // GetObjectNInfo - returns object info and an object
 | |
| // Read(Closer). When err != nil, the returned reader is always nil.
 | |
| func (er erasureObjects) GetObjectNInfo(ctx context.Context, bucket, object string, rs *HTTPRangeSpec, h http.Header, lockType LockType, opts ObjectOptions) (gr *GetObjectReader, err error) {
 | |
| 	var unlockOnDefer bool
 | |
| 	var nsUnlocker = func() {}
 | |
| 	defer func() {
 | |
| 		if unlockOnDefer {
 | |
| 			nsUnlocker()
 | |
| 		}
 | |
| 	}()
 | |
| 
 | |
| 	// Acquire lock
 | |
| 	if lockType != noLock {
 | |
| 		lock := er.NewNSLock(bucket, object)
 | |
| 		switch lockType {
 | |
| 		case writeLock:
 | |
| 			if err = lock.GetLock(ctx, globalOperationTimeout); err != nil {
 | |
| 				return nil, err
 | |
| 			}
 | |
| 			nsUnlocker = lock.Unlock
 | |
| 		case readLock:
 | |
| 			if err = lock.GetRLock(ctx, globalOperationTimeout); err != nil {
 | |
| 				return nil, err
 | |
| 			}
 | |
| 			nsUnlocker = lock.RUnlock
 | |
| 		}
 | |
| 		unlockOnDefer = true
 | |
| 	}
 | |
| 
 | |
| 	fi, metaArr, onlineDisks, err := er.getObjectFileInfo(ctx, bucket, object, opts, true)
 | |
| 	if err != nil {
 | |
| 		return nil, toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	objInfo := fi.ToObjectInfo(bucket, object)
 | |
| 	if objInfo.DeleteMarker {
 | |
| 		if opts.VersionID == "" {
 | |
| 			return &GetObjectReader{
 | |
| 				ObjInfo: objInfo,
 | |
| 			}, toObjectErr(errFileNotFound, bucket, object)
 | |
| 		}
 | |
| 		// Make sure to return object info to provide extra information.
 | |
| 		return &GetObjectReader{
 | |
| 			ObjInfo: objInfo,
 | |
| 		}, toObjectErr(errMethodNotAllowed, bucket, object)
 | |
| 	}
 | |
| 	if objInfo.TransitionStatus == lifecycle.TransitionComplete {
 | |
| 		// If transitioned, stream from transition tier unless object is restored locally or restore date is past.
 | |
| 		restoreHdr, ok := objInfo.UserDefined[xhttp.AmzRestore]
 | |
| 		if !ok || !strings.HasPrefix(restoreHdr, "ongoing-request=false") || (!objInfo.RestoreExpires.IsZero() && time.Now().After(objInfo.RestoreExpires)) {
 | |
| 			return getTransitionedObjectReader(ctx, bucket, object, rs, h, objInfo, opts)
 | |
| 		}
 | |
| 	}
 | |
| 	unlockOnDefer = false
 | |
| 	fn, off, length, nErr := NewGetObjectReader(rs, objInfo, opts, nsUnlocker)
 | |
| 	if nErr != nil {
 | |
| 		return nil, nErr
 | |
| 	}
 | |
| 	pr, pw := io.Pipe()
 | |
| 	go func() {
 | |
| 		err := er.getObjectWithFileInfo(ctx, bucket, object, off, length, pw, fi, metaArr, onlineDisks)
 | |
| 		pw.CloseWithError(err)
 | |
| 	}()
 | |
| 
 | |
| 	// Cleanup function to cause the go routine above to exit, in
 | |
| 	// case of incomplete read.
 | |
| 	pipeCloser := func() { pr.Close() }
 | |
| 
 | |
| 	return fn(pr, h, opts.CheckPrecondFn, pipeCloser)
 | |
| }
 | |
| 
 | |
| // GetObject - reads an object erasured coded across multiple
 | |
| // disks. Supports additional parameters like offset and length
 | |
| // which are synonymous with HTTP Range requests.
 | |
| //
 | |
| // startOffset indicates the starting read location of the object.
 | |
| // length indicates the total length of the object.
 | |
| func (er erasureObjects) GetObject(ctx context.Context, bucket, object string, startOffset int64, length int64, writer io.Writer, etag string, opts ObjectOptions) error {
 | |
| 	// Lock the object before reading.
 | |
| 	lk := er.NewNSLock(bucket, object)
 | |
| 	if err := lk.GetRLock(ctx, globalOperationTimeout); err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	defer lk.RUnlock()
 | |
| 
 | |
| 	// Start offset cannot be negative.
 | |
| 	if startOffset < 0 {
 | |
| 		logger.LogIf(ctx, errUnexpected, logger.Application)
 | |
| 		return errUnexpected
 | |
| 	}
 | |
| 
 | |
| 	// Writer cannot be nil.
 | |
| 	if writer == nil {
 | |
| 		logger.LogIf(ctx, errUnexpected)
 | |
| 		return errUnexpected
 | |
| 	}
 | |
| 
 | |
| 	return er.getObject(ctx, bucket, object, startOffset, length, writer, opts)
 | |
| }
 | |
| 
 | |
| func (er erasureObjects) getObjectWithFileInfo(ctx context.Context, bucket, object string, startOffset int64, length int64, writer io.Writer, fi FileInfo, metaArr []FileInfo, onlineDisks []StorageAPI) error {
 | |
| 	// Reorder online disks based on erasure distribution order.
 | |
| 	// Reorder parts metadata based on erasure distribution order.
 | |
| 	onlineDisks, metaArr = shuffleDisksAndPartsMetadataByIndex(onlineDisks, metaArr, fi.Erasure.Distribution)
 | |
| 
 | |
| 	// For negative length read everything.
 | |
| 	if length < 0 {
 | |
| 		length = fi.Size - startOffset
 | |
| 	}
 | |
| 
 | |
| 	// Reply back invalid range if the input offset and length fall out of range.
 | |
| 	if startOffset > fi.Size || startOffset+length > fi.Size {
 | |
| 		logger.LogIf(ctx, InvalidRange{startOffset, length, fi.Size}, logger.Application)
 | |
| 		return InvalidRange{startOffset, length, fi.Size}
 | |
| 	}
 | |
| 
 | |
| 	// Get start part index and offset.
 | |
| 	partIndex, partOffset, err := fi.ObjectToPartOffset(ctx, startOffset)
 | |
| 	if err != nil {
 | |
| 		return InvalidRange{startOffset, length, fi.Size}
 | |
| 	}
 | |
| 
 | |
| 	// Calculate endOffset according to length
 | |
| 	endOffset := startOffset
 | |
| 	if length > 0 {
 | |
| 		endOffset += length - 1
 | |
| 	}
 | |
| 
 | |
| 	// Get last part index to read given length.
 | |
| 	lastPartIndex, _, err := fi.ObjectToPartOffset(ctx, endOffset)
 | |
| 	if err != nil {
 | |
| 		return InvalidRange{startOffset, length, fi.Size}
 | |
| 	}
 | |
| 
 | |
| 	var totalBytesRead int64
 | |
| 	erasure, err := NewErasure(ctx, fi.Erasure.DataBlocks, fi.Erasure.ParityBlocks, fi.Erasure.BlockSize)
 | |
| 	if err != nil {
 | |
| 		return toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 	var healOnce sync.Once
 | |
| 
 | |
| 	for ; partIndex <= lastPartIndex; partIndex++ {
 | |
| 		if length == totalBytesRead {
 | |
| 			break
 | |
| 		}
 | |
| 
 | |
| 		partNumber := fi.Parts[partIndex].Number
 | |
| 
 | |
| 		// Save the current part name and size.
 | |
| 		partSize := fi.Parts[partIndex].Size
 | |
| 
 | |
| 		partLength := partSize - partOffset
 | |
| 		// partLength should be adjusted so that we don't write more data than what was requested.
 | |
| 		if partLength > (length - totalBytesRead) {
 | |
| 			partLength = length - totalBytesRead
 | |
| 		}
 | |
| 
 | |
| 		tillOffset := erasure.ShardFileOffset(partOffset, partLength, partSize)
 | |
| 		// Get the checksums of the current part.
 | |
| 		readers := make([]io.ReaderAt, len(onlineDisks))
 | |
| 		prefer := make([]bool, len(onlineDisks))
 | |
| 		for index, disk := range onlineDisks {
 | |
| 			if disk == OfflineDisk {
 | |
| 				continue
 | |
| 			}
 | |
| 			if !metaArr[index].IsValid() {
 | |
| 				continue
 | |
| 			}
 | |
| 			checksumInfo := metaArr[index].Erasure.GetChecksumInfo(partNumber)
 | |
| 			partPath := pathJoin(object, metaArr[index].DataDir, fmt.Sprintf("part.%d", partNumber))
 | |
| 			data := metaArr[index].Data
 | |
| 			readers[index] = newBitrotReader(disk, data, bucket, partPath, tillOffset,
 | |
| 				checksumInfo.Algorithm, checksumInfo.Hash, erasure.ShardSize())
 | |
| 
 | |
| 			// Prefer local disks
 | |
| 			prefer[index] = disk.Hostname() == ""
 | |
| 		}
 | |
| 
 | |
| 		written, err := erasure.Decode(ctx, writer, readers, partOffset, partLength, partSize, prefer)
 | |
| 		// Note: we should not be defer'ing the following closeBitrotReaders() call as
 | |
| 		// we are inside a for loop i.e if we use defer, we would accumulate a lot of open files by the time
 | |
| 		// we return from this function.
 | |
| 		closeBitrotReaders(readers)
 | |
| 		if err != nil {
 | |
| 			// If we have successfully written all the content that was asked
 | |
| 			// by the client, but we still see an error - this would mean
 | |
| 			// that we have some parts or data blocks missing or corrupted
 | |
| 			// - attempt a heal to successfully heal them for future calls.
 | |
| 			if written == partLength {
 | |
| 				var scan madmin.HealScanMode
 | |
| 				if errors.Is(err, errFileNotFound) {
 | |
| 					scan = madmin.HealNormalScan
 | |
| 				} else if errors.Is(err, errFileCorrupt) {
 | |
| 					scan = madmin.HealDeepScan
 | |
| 				}
 | |
| 				if scan != madmin.HealUnknownScan {
 | |
| 					healOnce.Do(func() {
 | |
| 						go healObject(bucket, object, fi.VersionID, scan)
 | |
| 					})
 | |
| 				}
 | |
| 			}
 | |
| 			if err != nil {
 | |
| 				return toObjectErr(err, bucket, object)
 | |
| 			}
 | |
| 		}
 | |
| 		for i, r := range readers {
 | |
| 			if r == nil {
 | |
| 				onlineDisks[i] = OfflineDisk
 | |
| 			}
 | |
| 		}
 | |
| 		// Track total bytes read from disk and written to the client.
 | |
| 		totalBytesRead += partLength
 | |
| 		// partOffset will be valid only for the first part, hence reset it to 0 for
 | |
| 		// the remaining parts.
 | |
| 		partOffset = 0
 | |
| 	} // End of read all parts loop.
 | |
| 	// Return success.
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // getObject wrapper for erasure GetObject
 | |
| func (er erasureObjects) getObject(ctx context.Context, bucket, object string, startOffset, length int64, writer io.Writer, opts ObjectOptions) error {
 | |
| 	fi, metaArr, onlineDisks, err := er.getObjectFileInfo(ctx, bucket, object, opts, true)
 | |
| 	if err != nil {
 | |
| 		return toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 	if fi.Deleted {
 | |
| 		if opts.VersionID == "" {
 | |
| 			return toObjectErr(errFileNotFound, bucket, object)
 | |
| 		}
 | |
| 		// Make sure to return object info to provide extra information.
 | |
| 		return toObjectErr(errMethodNotAllowed, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	return er.getObjectWithFileInfo(ctx, bucket, object, startOffset, length, writer, fi, metaArr, onlineDisks)
 | |
| }
 | |
| 
 | |
| // GetObjectInfo - reads object metadata and replies back ObjectInfo.
 | |
| func (er erasureObjects) GetObjectInfo(ctx context.Context, bucket, object string, opts ObjectOptions) (info ObjectInfo, err error) {
 | |
| 	// Lock the object before reading.
 | |
| 	lk := er.NewNSLock(bucket, object)
 | |
| 	if err := lk.GetRLock(ctx, globalOperationTimeout); err != nil {
 | |
| 		return ObjectInfo{}, err
 | |
| 	}
 | |
| 	defer lk.RUnlock()
 | |
| 
 | |
| 	return er.getObjectInfo(ctx, bucket, object, opts)
 | |
| }
 | |
| 
 | |
| func (er erasureObjects) getObjectFileInfo(ctx context.Context, bucket, object string, opts ObjectOptions, readData bool) (fi FileInfo, metaArr []FileInfo, onlineDisks []StorageAPI, err error) {
 | |
| 	disks := er.getDisks()
 | |
| 
 | |
| 	// Read metadata associated with the object from all disks.
 | |
| 	metaArr, errs := readAllFileInfo(ctx, disks, bucket, object, opts.VersionID, readData)
 | |
| 
 | |
| 	readQuorum, _, err := objectQuorumFromMeta(ctx, metaArr, errs, er.defaultParityCount)
 | |
| 	if err != nil {
 | |
| 		return fi, nil, nil, err
 | |
| 	}
 | |
| 
 | |
| 	if reducedErr := reduceReadQuorumErrs(ctx, errs, objectOpIgnoredErrs, readQuorum); reducedErr != nil {
 | |
| 		if reducedErr == errErasureReadQuorum && bucket != minioMetaBucket {
 | |
| 			if _, ok := isObjectDangling(metaArr, errs, nil); ok {
 | |
| 				reducedErr = errFileNotFound
 | |
| 				if opts.VersionID != "" {
 | |
| 					reducedErr = errFileVersionNotFound
 | |
| 				}
 | |
| 				// Remove the dangling object only when:
 | |
| 				//  - This is a non versioned bucket
 | |
| 				//  - This is a versioned bucket and the version ID is passed, the reason
 | |
| 				//    is that we cannot fetch the ID of the latest version when we don't trust xl.meta
 | |
| 				if !opts.Versioned || opts.VersionID != "" {
 | |
| 					er.deleteObjectVersion(ctx, bucket, object, 1, FileInfo{
 | |
| 						Name:      object,
 | |
| 						VersionID: opts.VersionID,
 | |
| 					}, false)
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 		return fi, nil, nil, toObjectErr(reducedErr, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	// List all online disks.
 | |
| 	onlineDisks, modTime := listOnlineDisks(disks, metaArr, errs)
 | |
| 
 | |
| 	// Pick latest valid metadata.
 | |
| 	fi, err = pickValidFileInfo(ctx, metaArr, modTime, readQuorum)
 | |
| 	if err != nil {
 | |
| 		return fi, nil, nil, err
 | |
| 	}
 | |
| 
 | |
| 	var missingBlocks int
 | |
| 	for i, err := range errs {
 | |
| 		if err != nil && errors.Is(err, errFileNotFound) {
 | |
| 			missingBlocks++
 | |
| 			continue
 | |
| 		}
 | |
| 		if metaArr[i].IsValid() && metaArr[i].ModTime.Equal(fi.ModTime) {
 | |
| 			continue
 | |
| 		}
 | |
| 		missingBlocks++
 | |
| 	}
 | |
| 
 | |
| 	// if missing metadata can be reconstructed, attempt to reconstruct.
 | |
| 	if missingBlocks > 0 && missingBlocks < readQuorum {
 | |
| 		go healObject(bucket, object, fi.VersionID, madmin.HealNormalScan)
 | |
| 	}
 | |
| 
 | |
| 	return fi, metaArr, onlineDisks, nil
 | |
| }
 | |
| 
 | |
| // getObjectInfo - wrapper for reading object metadata and constructs ObjectInfo.
 | |
| func (er erasureObjects) getObjectInfo(ctx context.Context, bucket, object string, opts ObjectOptions) (objInfo ObjectInfo, err error) {
 | |
| 	fi, _, _, err := er.getObjectFileInfo(ctx, bucket, object, opts, false)
 | |
| 	if err != nil {
 | |
| 		return objInfo, toObjectErr(err, bucket, object)
 | |
| 
 | |
| 	}
 | |
| 	objInfo = fi.ToObjectInfo(bucket, object)
 | |
| 	if objInfo.TransitionStatus == lifecycle.TransitionComplete {
 | |
| 		// overlay storage class for transitioned objects with transition tier SC Label
 | |
| 		if sc := transitionSC(ctx, bucket); sc != "" {
 | |
| 			objInfo.StorageClass = sc
 | |
| 		}
 | |
| 	}
 | |
| 	if !fi.VersionPurgeStatus.Empty() {
 | |
| 		// Make sure to return object info to provide extra information.
 | |
| 		return objInfo, toObjectErr(errMethodNotAllowed, bucket, object)
 | |
| 	}
 | |
| 	if fi.Deleted {
 | |
| 		if opts.VersionID == "" || opts.DeleteMarker {
 | |
| 			return objInfo, toObjectErr(errFileNotFound, bucket, object)
 | |
| 		}
 | |
| 		// Make sure to return object info to provide extra information.
 | |
| 		return objInfo, toObjectErr(errMethodNotAllowed, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	return objInfo, nil
 | |
| }
 | |
| 
 | |
| func undoRename(disks []StorageAPI, srcBucket, srcEntry, dstBucket, dstEntry string, isDir bool, errs []error) {
 | |
| 	// Undo rename object on disks where RenameFile succeeded.
 | |
| 
 | |
| 	// If srcEntry/dstEntry are objects then add a trailing slash to copy
 | |
| 	// over all the parts inside the object directory
 | |
| 	if isDir {
 | |
| 		srcEntry = retainSlash(srcEntry)
 | |
| 		dstEntry = retainSlash(dstEntry)
 | |
| 	}
 | |
| 	g := errgroup.WithNErrs(len(disks))
 | |
| 	for index, disk := range disks {
 | |
| 		if disk == nil {
 | |
| 			continue
 | |
| 		}
 | |
| 		index := index
 | |
| 		g.Go(func() error {
 | |
| 			if errs[index] == nil {
 | |
| 				_ = disks[index].RenameFile(context.TODO(), dstBucket, dstEntry, srcBucket, srcEntry)
 | |
| 			}
 | |
| 			return nil
 | |
| 		}, index)
 | |
| 	}
 | |
| 	g.Wait()
 | |
| }
 | |
| 
 | |
| // Similar to rename but renames data from srcEntry to dstEntry at dataDir
 | |
| func renameData(ctx context.Context, disks []StorageAPI, srcBucket, srcEntry, dataDir, dstBucket, dstEntry string, writeQuorum int, ignoredErr []error) ([]StorageAPI, error) {
 | |
| 	dataDir = retainSlash(dataDir)
 | |
| 	defer ObjectPathUpdated(pathJoin(srcBucket, srcEntry))
 | |
| 	defer ObjectPathUpdated(pathJoin(dstBucket, dstEntry))
 | |
| 
 | |
| 	g := errgroup.WithNErrs(len(disks))
 | |
| 
 | |
| 	// Rename file on all underlying storage disks.
 | |
| 	for index := range disks {
 | |
| 		index := index
 | |
| 		g.Go(func() error {
 | |
| 			if disks[index] == nil {
 | |
| 				return errDiskNotFound
 | |
| 			}
 | |
| 			if err := disks[index].RenameData(ctx, srcBucket, srcEntry, dataDir, dstBucket, dstEntry); err != nil {
 | |
| 				if !IsErrIgnored(err, ignoredErr...) {
 | |
| 					return err
 | |
| 				}
 | |
| 			}
 | |
| 			return nil
 | |
| 		}, index)
 | |
| 	}
 | |
| 
 | |
| 	// Wait for all renames to finish.
 | |
| 	errs := g.Wait()
 | |
| 
 | |
| 	// We can safely allow RenameFile errors up to len(er.getDisks()) - writeQuorum
 | |
| 	// otherwise return failure. Cleanup successful renames.
 | |
| 	err := reduceWriteQuorumErrs(ctx, errs, objectOpIgnoredErrs, writeQuorum)
 | |
| 	if err == errErasureWriteQuorum {
 | |
| 		ug := errgroup.WithNErrs(len(disks))
 | |
| 		for index, disk := range disks {
 | |
| 			if disk == nil {
 | |
| 				continue
 | |
| 			}
 | |
| 			index := index
 | |
| 			ug.Go(func() error {
 | |
| 				// Undo all the partial rename operations.
 | |
| 				if errs[index] == nil {
 | |
| 					_ = disks[index].RenameData(context.Background(), dstBucket, dstEntry, dataDir, srcBucket, srcEntry)
 | |
| 				}
 | |
| 				return nil
 | |
| 			}, index)
 | |
| 		}
 | |
| 		ug.Wait()
 | |
| 	}
 | |
| 	return evalDisks(disks, errs), err
 | |
| }
 | |
| 
 | |
| // rename - common function that renamePart and renameObject use to rename
 | |
| // the respective underlying storage layer representations.
 | |
| func rename(ctx context.Context, disks []StorageAPI, srcBucket, srcEntry, dstBucket, dstEntry string, isDir bool, writeQuorum int, ignoredErr []error) ([]StorageAPI, error) {
 | |
| 	if isDir {
 | |
| 		dstEntry = retainSlash(dstEntry)
 | |
| 		srcEntry = retainSlash(srcEntry)
 | |
| 	}
 | |
| 	defer ObjectPathUpdated(pathJoin(srcBucket, srcEntry))
 | |
| 	defer ObjectPathUpdated(pathJoin(dstBucket, dstEntry))
 | |
| 
 | |
| 	g := errgroup.WithNErrs(len(disks))
 | |
| 
 | |
| 	// Rename file on all underlying storage disks.
 | |
| 	for index := range disks {
 | |
| 		index := index
 | |
| 		g.Go(func() error {
 | |
| 			if disks[index] == nil {
 | |
| 				return errDiskNotFound
 | |
| 			}
 | |
| 			if err := disks[index].RenameFile(ctx, srcBucket, srcEntry, dstBucket, dstEntry); err != nil {
 | |
| 				if !IsErrIgnored(err, ignoredErr...) {
 | |
| 					return err
 | |
| 				}
 | |
| 			}
 | |
| 			return nil
 | |
| 		}, index)
 | |
| 	}
 | |
| 
 | |
| 	// Wait for all renames to finish.
 | |
| 	errs := g.Wait()
 | |
| 
 | |
| 	// We can safely allow RenameFile errors up to len(er.getDisks()) - writeQuorum
 | |
| 	// otherwise return failure. Cleanup successful renames.
 | |
| 	err := reduceWriteQuorumErrs(ctx, errs, objectOpIgnoredErrs, writeQuorum)
 | |
| 	if err == errErasureWriteQuorum {
 | |
| 		// Undo all the partial rename operations.
 | |
| 		undoRename(disks, srcBucket, srcEntry, dstBucket, dstEntry, isDir, errs)
 | |
| 	}
 | |
| 	return evalDisks(disks, errs), err
 | |
| }
 | |
| 
 | |
| // PutObject - creates an object upon reading from the input stream
 | |
| // until EOF, erasure codes the data across all disk and additionally
 | |
| // writes `xl.meta` which carries the necessary metadata for future
 | |
| // object operations.
 | |
| func (er erasureObjects) PutObject(ctx context.Context, bucket string, object string, data *PutObjReader, opts ObjectOptions) (objInfo ObjectInfo, err error) {
 | |
| 	return er.putObject(ctx, bucket, object, data, opts)
 | |
| }
 | |
| 
 | |
| // putObject wrapper for erasureObjects PutObject
 | |
| func (er erasureObjects) putObject(ctx context.Context, bucket string, object string, r *PutObjReader, opts ObjectOptions) (objInfo ObjectInfo, err error) {
 | |
| 	defer func() {
 | |
| 		ObjectPathUpdated(pathJoin(bucket, object))
 | |
| 	}()
 | |
| 
 | |
| 	data := r.Reader
 | |
| 
 | |
| 	uniqueID := mustGetUUID()
 | |
| 	tempObj := uniqueID
 | |
| 	// No metadata is set, allocate a new one.
 | |
| 	if opts.UserDefined == nil {
 | |
| 		opts.UserDefined = make(map[string]string)
 | |
| 	}
 | |
| 
 | |
| 	storageDisks := er.getDisks()
 | |
| 
 | |
| 	// Get parity and data drive count based on storage class metadata
 | |
| 	parityDrives := globalStorageClass.GetParityForSC(opts.UserDefined[xhttp.AmzStorageClass])
 | |
| 	if parityDrives <= 0 {
 | |
| 		parityDrives = er.defaultParityCount
 | |
| 	}
 | |
| 	dataDrives := len(storageDisks) - parityDrives
 | |
| 
 | |
| 	// we now know the number of blocks this object needs for data and parity.
 | |
| 	// writeQuorum is dataBlocks + 1
 | |
| 	writeQuorum := dataDrives
 | |
| 	if dataDrives == parityDrives {
 | |
| 		writeQuorum++
 | |
| 	}
 | |
| 
 | |
| 	// Delete temporary object in the event of failure.
 | |
| 	// If PutObject succeeded there would be no temporary
 | |
| 	// object to delete.
 | |
| 	defer er.deleteObject(context.Background(), minioMetaTmpBucket, tempObj, writeQuorum)
 | |
| 
 | |
| 	// Validate input data size and it can never be less than zero.
 | |
| 	if data.Size() < -1 {
 | |
| 		logger.LogIf(ctx, errInvalidArgument, logger.Application)
 | |
| 		return ObjectInfo{}, toObjectErr(errInvalidArgument)
 | |
| 	}
 | |
| 
 | |
| 	// Check if an object is present as one of the parent dir.
 | |
| 	// -- FIXME. (needs a new kind of lock).
 | |
| 	// -- FIXME (this also causes performance issue when disks are down).
 | |
| 	if opts.ParentIsObject != nil && opts.ParentIsObject(ctx, bucket, path.Dir(object)) {
 | |
| 		return ObjectInfo{}, toObjectErr(errFileParentIsFile, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	// Initialize parts metadata
 | |
| 	partsMetadata := make([]FileInfo, len(storageDisks))
 | |
| 
 | |
| 	fi := newFileInfo(object, dataDrives, parityDrives)
 | |
| 
 | |
| 	if opts.Versioned {
 | |
| 		fi.VersionID = opts.VersionID
 | |
| 		if fi.VersionID == "" {
 | |
| 			fi.VersionID = mustGetUUID()
 | |
| 		}
 | |
| 	}
 | |
| 	fi.DataDir = mustGetUUID()
 | |
| 
 | |
| 	// Initialize erasure metadata.
 | |
| 	for index := range partsMetadata {
 | |
| 		partsMetadata[index] = fi
 | |
| 	}
 | |
| 
 | |
| 	// Order disks according to erasure distribution
 | |
| 	var onlineDisks []StorageAPI
 | |
| 	onlineDisks, partsMetadata = shuffleDisksAndPartsMetadata(storageDisks, partsMetadata, fi.Erasure.Distribution)
 | |
| 
 | |
| 	erasure, err := NewErasure(ctx, fi.Erasure.DataBlocks, fi.Erasure.ParityBlocks, fi.Erasure.BlockSize)
 | |
| 	if err != nil {
 | |
| 		return ObjectInfo{}, toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	// Fetch buffer for I/O, returns from the pool if not allocates a new one and returns.
 | |
| 	var buffer []byte
 | |
| 	switch size := data.Size(); {
 | |
| 	case size == 0:
 | |
| 		buffer = make([]byte, 1) // Allocate atleast a byte to reach EOF
 | |
| 	case size == -1 || size >= fi.Erasure.BlockSize:
 | |
| 		buffer = er.bp.Get()
 | |
| 		defer er.bp.Put(buffer)
 | |
| 	case size < fi.Erasure.BlockSize:
 | |
| 		// No need to allocate fully blockSizeV1 buffer if the incoming data is smaller.
 | |
| 		buffer = make([]byte, size, 2*size+int64(fi.Erasure.ParityBlocks+fi.Erasure.DataBlocks-1))
 | |
| 	}
 | |
| 
 | |
| 	if len(buffer) > int(fi.Erasure.BlockSize) {
 | |
| 		buffer = buffer[:fi.Erasure.BlockSize]
 | |
| 	}
 | |
| 
 | |
| 	partName := "part.1"
 | |
| 	tempErasureObj := pathJoin(uniqueID, fi.DataDir, partName)
 | |
| 
 | |
| 	writers := make([]io.Writer, len(onlineDisks))
 | |
| 	for i, disk := range onlineDisks {
 | |
| 		if disk == nil {
 | |
| 			continue
 | |
| 		}
 | |
| 		writers[i] = newBitrotWriter(disk, minioMetaTmpBucket, tempErasureObj, erasure.ShardFileSize(data.Size()), DefaultBitrotAlgorithm, erasure.ShardSize())
 | |
| 	}
 | |
| 
 | |
| 	n, erasureErr := erasure.Encode(ctx, data, writers, buffer, writeQuorum)
 | |
| 	closeBitrotWriters(writers)
 | |
| 	if erasureErr != nil {
 | |
| 		return ObjectInfo{}, toObjectErr(erasureErr, minioMetaTmpBucket, tempErasureObj)
 | |
| 	}
 | |
| 
 | |
| 	// Should return IncompleteBody{} error when reader has fewer bytes
 | |
| 	// than specified in request header.
 | |
| 	if n < data.Size() {
 | |
| 		return ObjectInfo{}, IncompleteBody{Bucket: bucket, Object: object}
 | |
| 	}
 | |
| 
 | |
| 	if !opts.NoLock {
 | |
| 		lk := er.NewNSLock(bucket, object)
 | |
| 		if err := lk.GetLock(ctx, globalOperationTimeout); err != nil {
 | |
| 			return ObjectInfo{}, err
 | |
| 		}
 | |
| 		defer lk.Unlock()
 | |
| 	}
 | |
| 
 | |
| 	for i, w := range writers {
 | |
| 		if w == nil {
 | |
| 			onlineDisks[i] = nil
 | |
| 			continue
 | |
| 		}
 | |
| 		partsMetadata[i].AddObjectPart(1, "", n, data.ActualSize())
 | |
| 		partsMetadata[i].Erasure.AddChecksumInfo(ChecksumInfo{
 | |
| 			PartNumber: 1,
 | |
| 			Algorithm:  DefaultBitrotAlgorithm,
 | |
| 			Hash:       bitrotWriterSum(w),
 | |
| 		})
 | |
| 	}
 | |
| 	if opts.UserDefined["etag"] == "" {
 | |
| 		opts.UserDefined["etag"] = r.MD5CurrentHexString()
 | |
| 	}
 | |
| 
 | |
| 	// Guess content-type from the extension if possible.
 | |
| 	if opts.UserDefined["content-type"] == "" {
 | |
| 		opts.UserDefined["content-type"] = mimedb.TypeByExtension(path.Ext(object))
 | |
| 	}
 | |
| 
 | |
| 	modTime := opts.MTime
 | |
| 	if opts.MTime.IsZero() {
 | |
| 		modTime = UTCNow()
 | |
| 	}
 | |
| 
 | |
| 	// Fill all the necessary metadata.
 | |
| 	// Update `xl.meta` content on each disks.
 | |
| 	for index := range partsMetadata {
 | |
| 		partsMetadata[index].Metadata = opts.UserDefined
 | |
| 		partsMetadata[index].Size = n
 | |
| 		partsMetadata[index].ModTime = modTime
 | |
| 	}
 | |
| 
 | |
| 	// Write unique `xl.meta` for each disk.
 | |
| 	if onlineDisks, err = writeUniqueFileInfo(ctx, onlineDisks, minioMetaTmpBucket, tempObj, partsMetadata, writeQuorum); err != nil {
 | |
| 		return ObjectInfo{}, toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	// Rename the successfully written temporary object to final location.
 | |
| 	if onlineDisks, err = renameData(ctx, onlineDisks, minioMetaTmpBucket, tempObj, fi.DataDir, bucket, object, writeQuorum, nil); err != nil {
 | |
| 		return ObjectInfo{}, toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	// Whether a disk was initially or becomes offline
 | |
| 	// during this upload, send it to the MRF list.
 | |
| 	for i := 0; i < len(onlineDisks); i++ {
 | |
| 		if onlineDisks[i] != nil && onlineDisks[i].IsOnline() {
 | |
| 			continue
 | |
| 		}
 | |
| 		er.addPartial(bucket, object, fi.VersionID)
 | |
| 		break
 | |
| 	}
 | |
| 
 | |
| 	for i := 0; i < len(onlineDisks); i++ {
 | |
| 		if onlineDisks[i] != nil && onlineDisks[i].IsOnline() {
 | |
| 			// Object info is the same in all disks, so we can pick
 | |
| 			// the first meta from online disk
 | |
| 			fi = partsMetadata[i]
 | |
| 			break
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return fi.ToObjectInfo(bucket, object), nil
 | |
| }
 | |
| 
 | |
| func (er erasureObjects) deleteObjectVersion(ctx context.Context, bucket, object string, writeQuorum int, fi FileInfo, forceDelMarker bool) error {
 | |
| 	defer ObjectPathUpdated(pathJoin(bucket, object))
 | |
| 	disks := er.getDisks()
 | |
| 	g := errgroup.WithNErrs(len(disks))
 | |
| 	for index := range disks {
 | |
| 		index := index
 | |
| 		g.Go(func() error {
 | |
| 			if disks[index] == nil {
 | |
| 				return errDiskNotFound
 | |
| 			}
 | |
| 			return disks[index].DeleteVersion(ctx, bucket, object, fi, forceDelMarker)
 | |
| 		}, index)
 | |
| 	}
 | |
| 	// return errors if any during deletion
 | |
| 	return reduceWriteQuorumErrs(ctx, g.Wait(), objectOpIgnoredErrs, writeQuorum)
 | |
| }
 | |
| 
 | |
| // deleteEmptyDir knows only how to remove an empty directory (not the empty object with a
 | |
| // trailing slash), this is called for the healing code to remove such directories.
 | |
| func (er erasureObjects) deleteEmptyDir(ctx context.Context, bucket, object string) error {
 | |
| 	defer ObjectPathUpdated(pathJoin(bucket, object))
 | |
| 
 | |
| 	if bucket == minioMetaTmpBucket {
 | |
| 		return nil
 | |
| 	}
 | |
| 
 | |
| 	disks := er.getDisks()
 | |
| 	g := errgroup.WithNErrs(len(disks))
 | |
| 	for index := range disks {
 | |
| 		index := index
 | |
| 		g.Go(func() error {
 | |
| 			if disks[index] == nil {
 | |
| 				return errDiskNotFound
 | |
| 			}
 | |
| 			return disks[index].Delete(ctx, bucket, object, false)
 | |
| 		}, index)
 | |
| 	}
 | |
| 
 | |
| 	// return errors if any during deletion
 | |
| 	return reduceWriteQuorumErrs(ctx, g.Wait(), objectOpIgnoredErrs, len(disks)/2+1)
 | |
| }
 | |
| 
 | |
| // deleteObject - wrapper for delete object, deletes an object from
 | |
| // all the disks in parallel, including `xl.meta` associated with the
 | |
| // object.
 | |
| func (er erasureObjects) deleteObject(ctx context.Context, bucket, object string, writeQuorum int) error {
 | |
| 	var err error
 | |
| 	defer ObjectPathUpdated(pathJoin(bucket, object))
 | |
| 
 | |
| 	tmpObj := mustGetUUID()
 | |
| 	disks := er.getDisks()
 | |
| 	if bucket == minioMetaTmpBucket {
 | |
| 		tmpObj = object
 | |
| 	} else {
 | |
| 		// Rename the current object while requiring write quorum, but also consider
 | |
| 		// that a non found object in a given disk as a success since it already
 | |
| 		// confirms that the object doesn't have a part in that disk (already removed)
 | |
| 		disks, err = rename(ctx, disks, bucket, object, minioMetaTmpBucket, tmpObj, true, writeQuorum,
 | |
| 			[]error{errFileNotFound})
 | |
| 		if err != nil {
 | |
| 			return toObjectErr(err, bucket, object)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	g := errgroup.WithNErrs(len(disks))
 | |
| 	for index := range disks {
 | |
| 		index := index
 | |
| 		g.Go(func() error {
 | |
| 			if disks[index] == nil {
 | |
| 				return errDiskNotFound
 | |
| 			}
 | |
| 			return cleanupDir(ctx, disks[index], minioMetaTmpBucket, tmpObj)
 | |
| 		}, index)
 | |
| 	}
 | |
| 
 | |
| 	// return errors if any during deletion
 | |
| 	return reduceWriteQuorumErrs(ctx, g.Wait(), objectOpIgnoredErrs, writeQuorum)
 | |
| }
 | |
| 
 | |
| // DeleteObjects deletes objects/versions in bulk, this function will still automatically split objects list
 | |
| // into smaller bulks if some object names are found to be duplicated in the delete list, splitting
 | |
| // into smaller bulks will avoid holding twice the write lock of the duplicated object names.
 | |
| func (er erasureObjects) DeleteObjects(ctx context.Context, bucket string, objects []ObjectToDelete, opts ObjectOptions) ([]DeletedObject, []error) {
 | |
| 	errs := make([]error, len(objects))
 | |
| 	dobjects := make([]DeletedObject, len(objects))
 | |
| 	writeQuorums := make([]int, len(objects))
 | |
| 
 | |
| 	storageDisks := er.getDisks()
 | |
| 
 | |
| 	for i := range objects {
 | |
| 		// Assume (N/2 + 1) quorums for all objects
 | |
| 		// this is a theoretical assumption such that
 | |
| 		// for delete's we do not need to honor storage
 | |
| 		// class for objects which have reduced quorum
 | |
| 		// storage class only needs to be honored for
 | |
| 		// Read() requests alone which we already do.
 | |
| 		writeQuorums[i] = getWriteQuorum(len(storageDisks))
 | |
| 	}
 | |
| 
 | |
| 	versions := make([]FileInfo, len(objects))
 | |
| 	for i := range objects {
 | |
| 		if objects[i].VersionID == "" {
 | |
| 			modTime := opts.MTime
 | |
| 			if opts.MTime.IsZero() {
 | |
| 				modTime = UTCNow()
 | |
| 			}
 | |
| 			uuid := opts.VersionID
 | |
| 			if uuid == "" {
 | |
| 				uuid = mustGetUUID()
 | |
| 			}
 | |
| 			if opts.Versioned || opts.VersionSuspended {
 | |
| 				versions[i] = FileInfo{
 | |
| 					Name:                          objects[i].ObjectName,
 | |
| 					ModTime:                       modTime,
 | |
| 					Deleted:                       true, // delete marker
 | |
| 					DeleteMarkerReplicationStatus: objects[i].DeleteMarkerReplicationStatus,
 | |
| 					VersionPurgeStatus:            objects[i].VersionPurgeStatus,
 | |
| 				}
 | |
| 				if opts.Versioned {
 | |
| 					versions[i].VersionID = uuid
 | |
| 				}
 | |
| 				continue
 | |
| 			}
 | |
| 		}
 | |
| 		versions[i] = FileInfo{
 | |
| 			Name:                          objects[i].ObjectName,
 | |
| 			VersionID:                     objects[i].VersionID,
 | |
| 			DeleteMarkerReplicationStatus: objects[i].DeleteMarkerReplicationStatus,
 | |
| 			VersionPurgeStatus:            objects[i].VersionPurgeStatus,
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// Initialize list of errors.
 | |
| 	var delObjErrs = make([][]error, len(storageDisks))
 | |
| 
 | |
| 	var wg sync.WaitGroup
 | |
| 	// Remove versions in bulk for each disk
 | |
| 	for index, disk := range storageDisks {
 | |
| 		wg.Add(1)
 | |
| 		go func(index int, disk StorageAPI) {
 | |
| 			defer wg.Done()
 | |
| 			if disk == nil {
 | |
| 				delObjErrs[index] = make([]error, len(versions))
 | |
| 				for i := range versions {
 | |
| 					delObjErrs[index][i] = errDiskNotFound
 | |
| 				}
 | |
| 				return
 | |
| 			}
 | |
| 			delObjErrs[index] = disk.DeleteVersions(ctx, bucket, versions)
 | |
| 		}(index, disk)
 | |
| 	}
 | |
| 
 | |
| 	wg.Wait()
 | |
| 
 | |
| 	// Reduce errors for each object
 | |
| 	for objIndex := range objects {
 | |
| 		diskErrs := make([]error, len(storageDisks))
 | |
| 		// Iterate over disks to fetch the error
 | |
| 		// of deleting of the current object
 | |
| 		for i := range delObjErrs {
 | |
| 			// delObjErrs[i] is not nil when disks[i] is also not nil
 | |
| 			if delObjErrs[i] != nil {
 | |
| 				diskErrs[i] = delObjErrs[i][objIndex]
 | |
| 			}
 | |
| 		}
 | |
| 		err := reduceWriteQuorumErrs(ctx, diskErrs, objectOpIgnoredErrs, writeQuorums[objIndex])
 | |
| 		if objects[objIndex].VersionID != "" {
 | |
| 			errs[objIndex] = toObjectErr(err, bucket, objects[objIndex].ObjectName, objects[objIndex].VersionID)
 | |
| 		} else {
 | |
| 			errs[objIndex] = toObjectErr(err, bucket, objects[objIndex].ObjectName)
 | |
| 		}
 | |
| 
 | |
| 		if errs[objIndex] == nil {
 | |
| 			ObjectPathUpdated(pathJoin(bucket, objects[objIndex].ObjectName))
 | |
| 		}
 | |
| 
 | |
| 		if versions[objIndex].Deleted {
 | |
| 			dobjects[objIndex] = DeletedObject{
 | |
| 				DeleteMarker:                  versions[objIndex].Deleted,
 | |
| 				DeleteMarkerVersionID:         versions[objIndex].VersionID,
 | |
| 				DeleteMarkerMTime:             DeleteMarkerMTime{versions[objIndex].ModTime},
 | |
| 				DeleteMarkerReplicationStatus: versions[objIndex].DeleteMarkerReplicationStatus,
 | |
| 				ObjectName:                    versions[objIndex].Name,
 | |
| 				VersionPurgeStatus:            versions[objIndex].VersionPurgeStatus,
 | |
| 				PurgeTransitioned:             objects[objIndex].PurgeTransitioned,
 | |
| 			}
 | |
| 		} else {
 | |
| 			dobjects[objIndex] = DeletedObject{
 | |
| 				ObjectName:                    versions[objIndex].Name,
 | |
| 				VersionID:                     versions[objIndex].VersionID,
 | |
| 				VersionPurgeStatus:            versions[objIndex].VersionPurgeStatus,
 | |
| 				DeleteMarkerReplicationStatus: versions[objIndex].DeleteMarkerReplicationStatus,
 | |
| 				PurgeTransitioned:             objects[objIndex].PurgeTransitioned,
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// Check failed deletes across multiple objects
 | |
| 	for _, version := range versions {
 | |
| 		// Check if there is any offline disk and add it to the MRF list
 | |
| 		for _, disk := range storageDisks {
 | |
| 			if disk != nil && disk.IsOnline() {
 | |
| 				// Skip attempted heal on online disks.
 | |
| 				continue
 | |
| 			}
 | |
| 
 | |
| 			// all other direct versionId references we should
 | |
| 			// ensure no dangling file is left over.
 | |
| 			er.addPartial(bucket, version.Name, version.VersionID)
 | |
| 			break
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return dobjects, errs
 | |
| }
 | |
| 
 | |
| // DeleteObject - deletes an object, this call doesn't necessary reply
 | |
| // any error as it is not necessary for the handler to reply back a
 | |
| // response to the client request.
 | |
| func (er erasureObjects) DeleteObject(ctx context.Context, bucket, object string, opts ObjectOptions) (objInfo ObjectInfo, err error) {
 | |
| 	versionFound := true
 | |
| 	objInfo = ObjectInfo{VersionID: opts.VersionID} // version id needed in Delete API response.
 | |
| 	goi, gerr := er.GetObjectInfo(ctx, bucket, object, opts)
 | |
| 	if gerr != nil && goi.Name == "" {
 | |
| 		switch gerr.(type) {
 | |
| 		case InsufficientReadQuorum:
 | |
| 			return objInfo, InsufficientWriteQuorum{}
 | |
| 		}
 | |
| 		// For delete marker replication, versionID being replicated will not exist on disk
 | |
| 		if opts.DeleteMarker {
 | |
| 			versionFound = false
 | |
| 		} else {
 | |
| 			return objInfo, gerr
 | |
| 		}
 | |
| 	}
 | |
| 	// Acquire a write lock before deleting the object.
 | |
| 	lk := er.NewNSLock(bucket, object)
 | |
| 	if err = lk.GetLock(ctx, globalDeleteOperationTimeout); err != nil {
 | |
| 		return ObjectInfo{}, err
 | |
| 	}
 | |
| 	defer lk.Unlock()
 | |
| 
 | |
| 	storageDisks := er.getDisks()
 | |
| 	writeQuorum := len(storageDisks)/2 + 1
 | |
| 	var markDelete bool
 | |
| 	// Determine whether to mark object deleted for replication
 | |
| 	if goi.VersionID != "" {
 | |
| 		markDelete = true
 | |
| 	}
 | |
| 	// Default deleteMarker to true if object is under versioning
 | |
| 	deleteMarker := true
 | |
| 	if gerr == nil {
 | |
| 		deleteMarker = goi.VersionID != ""
 | |
| 	}
 | |
| 	if opts.VersionID != "" {
 | |
| 		// case where replica version needs to be deleted on target cluster
 | |
| 		if versionFound && opts.DeleteMarkerReplicationStatus == replication.Replica.String() {
 | |
| 			markDelete = false
 | |
| 		}
 | |
| 		if opts.VersionPurgeStatus.Empty() && opts.DeleteMarkerReplicationStatus == "" {
 | |
| 			markDelete = false
 | |
| 		}
 | |
| 		if opts.DeleteMarker && opts.VersionPurgeStatus == Complete {
 | |
| 			markDelete = false
 | |
| 		}
 | |
| 		// determine if the version represents an object delete
 | |
| 		// deleteMarker = true
 | |
| 		if versionFound && !goi.DeleteMarker { // implies a versioned delete of object
 | |
| 			deleteMarker = false
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	modTime := opts.MTime
 | |
| 	if opts.MTime.IsZero() {
 | |
| 		modTime = UTCNow()
 | |
| 	}
 | |
| 	if markDelete {
 | |
| 		if opts.Versioned || opts.VersionSuspended {
 | |
| 			fi := FileInfo{
 | |
| 				Name:                          object,
 | |
| 				Deleted:                       deleteMarker,
 | |
| 				MarkDeleted:                   markDelete,
 | |
| 				ModTime:                       modTime,
 | |
| 				DeleteMarkerReplicationStatus: opts.DeleteMarkerReplicationStatus,
 | |
| 				VersionPurgeStatus:            opts.VersionPurgeStatus,
 | |
| 			}
 | |
| 			if opts.Versioned {
 | |
| 				fi.VersionID = mustGetUUID()
 | |
| 				if opts.VersionID != "" {
 | |
| 					fi.VersionID = opts.VersionID
 | |
| 				}
 | |
| 			}
 | |
| 			fi.TransitionStatus = opts.TransitionStatus
 | |
| 
 | |
| 			// versioning suspended means we add `null`
 | |
| 			// version as delete marker
 | |
| 			// Add delete marker, since we don't have any version specified explicitly.
 | |
| 			// Or if a particular version id needs to be replicated.
 | |
| 			if err = er.deleteObjectVersion(ctx, bucket, object, writeQuorum, fi, opts.DeleteMarker); err != nil {
 | |
| 				return objInfo, toObjectErr(err, bucket, object)
 | |
| 			}
 | |
| 			return fi.ToObjectInfo(bucket, object), nil
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// Delete the object version on all disks.
 | |
| 	if err = er.deleteObjectVersion(ctx, bucket, object, writeQuorum, FileInfo{
 | |
| 		Name:                          object,
 | |
| 		VersionID:                     opts.VersionID,
 | |
| 		MarkDeleted:                   markDelete,
 | |
| 		Deleted:                       deleteMarker,
 | |
| 		ModTime:                       modTime,
 | |
| 		DeleteMarkerReplicationStatus: opts.DeleteMarkerReplicationStatus,
 | |
| 		VersionPurgeStatus:            opts.VersionPurgeStatus,
 | |
| 		TransitionStatus:              opts.TransitionStatus,
 | |
| 	}, opts.DeleteMarker); err != nil {
 | |
| 		return objInfo, toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	for _, disk := range storageDisks {
 | |
| 		if disk != nil && disk.IsOnline() {
 | |
| 			continue
 | |
| 		}
 | |
| 		er.addPartial(bucket, object, opts.VersionID)
 | |
| 		break
 | |
| 	}
 | |
| 
 | |
| 	return ObjectInfo{
 | |
| 		Bucket:             bucket,
 | |
| 		Name:               object,
 | |
| 		VersionID:          opts.VersionID,
 | |
| 		VersionPurgeStatus: opts.VersionPurgeStatus,
 | |
| 		ReplicationStatus:  replication.StatusType(opts.DeleteMarkerReplicationStatus),
 | |
| 	}, nil
 | |
| }
 | |
| 
 | |
| // Send the successful but partial upload/delete, however ignore
 | |
| // if the channel is blocked by other items.
 | |
| func (er erasureObjects) addPartial(bucket, object, versionID string) {
 | |
| 	select {
 | |
| 	case er.mrfOpCh <- partialOperation{bucket: bucket, object: object, versionID: versionID}:
 | |
| 	default:
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // PutObjectTags - replace or add tags to an existing object
 | |
| func (er erasureObjects) PutObjectTags(ctx context.Context, bucket, object string, tags string, opts ObjectOptions) (ObjectInfo, error) {
 | |
| 	// Lock the object before updating tags.
 | |
| 	lk := er.NewNSLock(bucket, object)
 | |
| 	if err := lk.GetLock(ctx, globalOperationTimeout); err != nil {
 | |
| 		return ObjectInfo{}, err
 | |
| 	}
 | |
| 	defer lk.Unlock()
 | |
| 
 | |
| 	disks := er.getDisks()
 | |
| 
 | |
| 	// Read metadata associated with the object from all disks.
 | |
| 	metaArr, errs := readAllFileInfo(ctx, disks, bucket, object, opts.VersionID, false)
 | |
| 
 | |
| 	readQuorum, writeQuorum, err := objectQuorumFromMeta(ctx, metaArr, errs, er.defaultParityCount)
 | |
| 	if err != nil {
 | |
| 		return ObjectInfo{}, toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	// List all online disks.
 | |
| 	onlineDisks, modTime := listOnlineDisks(disks, metaArr, errs)
 | |
| 
 | |
| 	// Pick latest valid metadata.
 | |
| 	fi, err := pickValidFileInfo(ctx, metaArr, modTime, readQuorum)
 | |
| 	if err != nil {
 | |
| 		return ObjectInfo{}, toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 	if fi.Deleted {
 | |
| 		if opts.VersionID == "" {
 | |
| 			return ObjectInfo{}, toObjectErr(errFileNotFound, bucket, object)
 | |
| 		}
 | |
| 		return ObjectInfo{}, toObjectErr(errMethodNotAllowed, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	onlineDisks, metaArr = shuffleDisksAndPartsMetadataByIndex(onlineDisks, metaArr, fi.Erasure.Distribution)
 | |
| 	for i, metaFi := range metaArr {
 | |
| 		if metaFi.IsValid() {
 | |
| 			// clean fi.Meta of tag key, before updating the new tags
 | |
| 			delete(metaFi.Metadata, xhttp.AmzObjectTagging)
 | |
| 			// Don't update for empty tags
 | |
| 			if tags != "" {
 | |
| 				metaFi.Metadata[xhttp.AmzObjectTagging] = tags
 | |
| 			}
 | |
| 			for k, v := range opts.UserDefined {
 | |
| 				metaFi.Metadata[k] = v
 | |
| 			}
 | |
| 			metaArr[i].Metadata = metaFi.Metadata
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	tempObj := mustGetUUID()
 | |
| 
 | |
| 	// Write unique `xl.meta` for each disk.
 | |
| 	if onlineDisks, err = writeUniqueFileInfo(ctx, onlineDisks, minioMetaTmpBucket, tempObj, metaArr, writeQuorum); err != nil {
 | |
| 		return ObjectInfo{}, toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	// Atomically rename metadata from tmp location to destination for each disk.
 | |
| 	if _, err = renameFileInfo(ctx, onlineDisks, minioMetaTmpBucket, tempObj, bucket, object, writeQuorum); err != nil {
 | |
| 		return ObjectInfo{}, toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	objInfo := fi.ToObjectInfo(bucket, object)
 | |
| 	objInfo.UserTags = tags
 | |
| 
 | |
| 	return objInfo, nil
 | |
| }
 | |
| 
 | |
| // updateObjectMeta will update the metadata of a file.
 | |
| func (er erasureObjects) updateObjectMeta(ctx context.Context, bucket, object string, meta map[string]string, opts ObjectOptions) error {
 | |
| 	if len(meta) == 0 {
 | |
| 		return nil
 | |
| 	}
 | |
| 	disks := er.getDisks()
 | |
| 
 | |
| 	// Read metadata associated with the object from all disks.
 | |
| 	metaArr, errs := readAllFileInfo(ctx, disks, bucket, object, opts.VersionID, false)
 | |
| 
 | |
| 	readQuorum, writeQuorum, err := objectQuorumFromMeta(ctx, metaArr, errs, er.defaultParityCount)
 | |
| 	if err != nil {
 | |
| 		return toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	// List all online disks.
 | |
| 	_, modTime := listOnlineDisks(disks, metaArr, errs)
 | |
| 
 | |
| 	// Pick latest valid metadata.
 | |
| 	fi, err := pickValidFileInfo(ctx, metaArr, modTime, readQuorum)
 | |
| 	if err != nil {
 | |
| 		return toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	// Update metadata
 | |
| 	for k, v := range meta {
 | |
| 		fi.Metadata[k] = v
 | |
| 	}
 | |
| 
 | |
| 	if fi.Deleted {
 | |
| 		if opts.VersionID == "" {
 | |
| 			return toObjectErr(errFileNotFound, bucket, object)
 | |
| 		}
 | |
| 		return toObjectErr(errMethodNotAllowed, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	for i := range metaArr {
 | |
| 		if errs[i] != nil {
 | |
| 			// Avoid disks where loading metadata fail
 | |
| 			continue
 | |
| 		}
 | |
| 
 | |
| 		metaArr[i].Metadata = fi.Metadata
 | |
| 	}
 | |
| 
 | |
| 	tempObj := mustGetUUID()
 | |
| 
 | |
| 	// Write unique `xl.meta` for each disk.
 | |
| 	if disks, err = writeUniqueFileInfo(ctx, disks, minioMetaTmpBucket, tempObj, metaArr, writeQuorum); err != nil {
 | |
| 		return toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	// Atomically rename metadata from tmp location to destination for each disk.
 | |
| 	if _, err = renameFileInfo(ctx, disks, minioMetaTmpBucket, tempObj, bucket, object, writeQuorum); err != nil {
 | |
| 		return toObjectErr(err, bucket, object)
 | |
| 	}
 | |
| 
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // DeleteObjectTags - delete object tags from an existing object
 | |
| func (er erasureObjects) DeleteObjectTags(ctx context.Context, bucket, object string, opts ObjectOptions) (ObjectInfo, error) {
 | |
| 	return er.PutObjectTags(ctx, bucket, object, "", opts)
 | |
| }
 | |
| 
 | |
| // GetObjectTags - get object tags from an existing object
 | |
| func (er erasureObjects) GetObjectTags(ctx context.Context, bucket, object string, opts ObjectOptions) (*tags.Tags, error) {
 | |
| 	// GetObjectInfo will return tag value as well
 | |
| 	oi, err := er.GetObjectInfo(ctx, bucket, object, opts)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	return tags.ParseObjectTags(oi.UserTags)
 | |
| }
 |