mirror of
				https://github.com/minio/minio.git
				synced 2025-10-31 00:01:27 +01:00 
			
		
		
		
	Rewrite parentIsObject() function. Currently if a client uploads a/b/c/d, we always check if c, b, a are actual objects or not. The new code will check with the reverse order and quickly quit if the segment doesn't exist. So if a, b, c in 'a/b/c' does not exist in the first place, then returns false quickly.
		
			
				
	
	
		
			257 lines
		
	
	
		
			6.6 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			257 lines
		
	
	
		
			6.6 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| /*
 | |
|  * MinIO Cloud Storage, (C) 2016-2020 MinIO, Inc.
 | |
|  *
 | |
|  * Licensed under the Apache License, Version 2.0 (the "License");
 | |
|  * you may not use this file except in compliance with the License.
 | |
|  * You may obtain a copy of the License at
 | |
|  *
 | |
|  *     http://www.apache.org/licenses/LICENSE-2.0
 | |
|  *
 | |
|  * Unless required by applicable law or agreed to in writing, software
 | |
|  * distributed under the License is distributed on an "AS IS" BASIS,
 | |
|  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
|  * See the License for the specific language governing permissions and
 | |
|  * limitations under the License.
 | |
|  */
 | |
| 
 | |
| package cmd
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"sort"
 | |
| 	"strings"
 | |
| 	"sync"
 | |
| 
 | |
| 	"github.com/minio/minio/pkg/sync/errgroup"
 | |
| )
 | |
| 
 | |
| func (er erasureObjects) getLoadBalancedLocalDisks() (newDisks []StorageAPI) {
 | |
| 	disks := er.getDisks()
 | |
| 	// Based on the random shuffling return back randomized disks.
 | |
| 	for _, i := range hashOrder(UTCNow().String(), len(disks)) {
 | |
| 		if disks[i-1] != nil && disks[i-1].IsLocal() {
 | |
| 			if !disks[i-1].Healing() && disks[i-1].IsOnline() {
 | |
| 				newDisks = append(newDisks, disks[i-1])
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	return newDisks
 | |
| }
 | |
| 
 | |
| type sortSlices struct {
 | |
| 	disks []StorageAPI
 | |
| 	infos []DiskInfo
 | |
| }
 | |
| 
 | |
| type sortByOther sortSlices
 | |
| 
 | |
| func (sbo sortByOther) Len() int {
 | |
| 	return len(sbo.disks)
 | |
| }
 | |
| 
 | |
| func (sbo sortByOther) Swap(i, j int) {
 | |
| 	sbo.disks[i], sbo.disks[j] = sbo.disks[j], sbo.disks[i]
 | |
| 	sbo.infos[i], sbo.infos[j] = sbo.infos[j], sbo.infos[i]
 | |
| }
 | |
| 
 | |
| func (sbo sortByOther) Less(i, j int) bool {
 | |
| 	return sbo.infos[i].UsedInodes < sbo.infos[j].UsedInodes
 | |
| }
 | |
| 
 | |
| func (er erasureObjects) getOnlineDisksSortedByUsedInodes() (newDisks []StorageAPI) {
 | |
| 	disks := er.getDisks()
 | |
| 	var wg sync.WaitGroup
 | |
| 	var mu sync.Mutex
 | |
| 	var infos []DiskInfo
 | |
| 	for _, i := range hashOrder(UTCNow().String(), len(disks)) {
 | |
| 		i := i
 | |
| 		wg.Add(1)
 | |
| 		go func() {
 | |
| 			defer wg.Done()
 | |
| 			if disks[i-1] == nil {
 | |
| 				return
 | |
| 			}
 | |
| 			di, err := disks[i-1].DiskInfo(context.Background())
 | |
| 			if err != nil || di.Healing {
 | |
| 
 | |
| 				// - Do not consume disks which are not reachable
 | |
| 				//   unformatted or simply not accessible for some reason.
 | |
| 				//
 | |
| 				// - Do not consume disks which are being healed
 | |
| 				//
 | |
| 				// - Future: skip busy disks
 | |
| 				return
 | |
| 			}
 | |
| 
 | |
| 			mu.Lock()
 | |
| 			newDisks = append(newDisks, disks[i-1])
 | |
| 			infos = append(infos, di)
 | |
| 			mu.Unlock()
 | |
| 		}()
 | |
| 	}
 | |
| 	wg.Wait()
 | |
| 
 | |
| 	slices := sortSlices{newDisks, infos}
 | |
| 	sort.Sort(sortByOther(slices))
 | |
| 
 | |
| 	return slices.disks
 | |
| }
 | |
| 
 | |
| func (er erasureObjects) getOnlineDisks() (newDisks []StorageAPI) {
 | |
| 	disks := er.getDisks()
 | |
| 	var wg sync.WaitGroup
 | |
| 	var mu sync.Mutex
 | |
| 	for _, i := range hashOrder(UTCNow().String(), len(disks)) {
 | |
| 		i := i
 | |
| 		wg.Add(1)
 | |
| 		go func() {
 | |
| 			defer wg.Done()
 | |
| 			if disks[i-1] == nil {
 | |
| 				return
 | |
| 			}
 | |
| 			di, err := disks[i-1].DiskInfo(context.Background())
 | |
| 			if err != nil || di.Healing {
 | |
| 
 | |
| 				// - Do not consume disks which are not reachable
 | |
| 				//   unformatted or simply not accessible for some reason.
 | |
| 				//
 | |
| 				// - Do not consume disks which are being healed
 | |
| 				//
 | |
| 				// - Future: skip busy disks
 | |
| 				return
 | |
| 			}
 | |
| 
 | |
| 			mu.Lock()
 | |
| 			newDisks = append(newDisks, disks[i-1])
 | |
| 			mu.Unlock()
 | |
| 		}()
 | |
| 	}
 | |
| 	wg.Wait()
 | |
| 	return newDisks
 | |
| }
 | |
| 
 | |
| // getLoadBalancedNDisks - fetches load balanced (sufficiently randomized) disk slice
 | |
| // with N disks online. If ndisks is zero or negative, then it will returns all disks,
 | |
| // same if ndisks is greater than the number of all disks.
 | |
| func (er erasureObjects) getLoadBalancedNDisks(ndisks int) (newDisks []StorageAPI) {
 | |
| 	disks := er.getLoadBalancedDisks(ndisks != -1)
 | |
| 	for _, disk := range disks {
 | |
| 		if disk == nil {
 | |
| 			continue
 | |
| 		}
 | |
| 		newDisks = append(newDisks, disk)
 | |
| 		ndisks--
 | |
| 		if ndisks == 0 {
 | |
| 			break
 | |
| 		}
 | |
| 	}
 | |
| 	return
 | |
| }
 | |
| 
 | |
| // getLoadBalancedDisks - fetches load balanced (sufficiently randomized) disk slice.
 | |
| // ensures to skip disks if they are not healing and online.
 | |
| func (er erasureObjects) getLoadBalancedDisks(optimized bool) []StorageAPI {
 | |
| 	disks := er.getDisks()
 | |
| 
 | |
| 	if !optimized {
 | |
| 		var newDisks []StorageAPI
 | |
| 		for _, i := range hashOrder(UTCNow().String(), len(disks)) {
 | |
| 			newDisks = append(newDisks, disks[i-1])
 | |
| 		}
 | |
| 		return newDisks
 | |
| 	}
 | |
| 
 | |
| 	var wg sync.WaitGroup
 | |
| 	var mu sync.Mutex
 | |
| 	var newDisks = map[uint64][]StorageAPI{}
 | |
| 	// Based on the random shuffling return back randomized disks.
 | |
| 	for _, i := range hashOrder(UTCNow().String(), len(disks)) {
 | |
| 		i := i
 | |
| 		wg.Add(1)
 | |
| 		go func() {
 | |
| 			defer wg.Done()
 | |
| 			if disks[i-1] == nil {
 | |
| 				return
 | |
| 			}
 | |
| 			di, err := disks[i-1].DiskInfo(context.Background())
 | |
| 			if err != nil || di.Healing {
 | |
| 				// - Do not consume disks which are not reachable
 | |
| 				//   unformatted or simply not accessible for some reason.
 | |
| 				//
 | |
| 				// - Do not consume disks which are being healed
 | |
| 				//
 | |
| 				// - Future: skip busy disks
 | |
| 				return
 | |
| 			}
 | |
| 
 | |
| 			mu.Lock()
 | |
| 			// Capture disks usage wise upto resolution of MiB
 | |
| 			newDisks[di.Used/1024/1024] = append(newDisks[di.Used/1024/1024], disks[i-1])
 | |
| 			mu.Unlock()
 | |
| 		}()
 | |
| 	}
 | |
| 	wg.Wait()
 | |
| 
 | |
| 	var max uint64
 | |
| 	for k := range newDisks {
 | |
| 		if k > max {
 | |
| 			max = k
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// Return disks which have maximum disk usage common.
 | |
| 	return newDisks[max]
 | |
| }
 | |
| 
 | |
| // This function does the following check, suppose
 | |
| // object is "a/b/c/d", stat makes sure that objects ""a/b/c""
 | |
| // "a/b" and "a" do not exist.
 | |
| func (er erasureObjects) parentDirIsObject(ctx context.Context, bucket, parent string) bool {
 | |
| 	path := ""
 | |
| 	segments := strings.Split(parent, slashSeparator)
 | |
| 	for _, s := range segments {
 | |
| 		if s == "" {
 | |
| 			break
 | |
| 		}
 | |
| 		path += s
 | |
| 		isObject, pathNotExist := er.isObject(ctx, bucket, path)
 | |
| 		if pathNotExist {
 | |
| 			return false
 | |
| 		}
 | |
| 		if isObject {
 | |
| 			// If there is already a file at prefix "p", return true.
 | |
| 			return true
 | |
| 		}
 | |
| 		path += slashSeparator
 | |
| 	}
 | |
| 	return false
 | |
| }
 | |
| 
 | |
| // isObject - returns `true` if the prefix is an object i.e if
 | |
| // `xl.meta` exists at the leaf, false otherwise.
 | |
| func (er erasureObjects) isObject(ctx context.Context, bucket, prefix string) (ok, pathDoesNotExist bool) {
 | |
| 	storageDisks := er.getDisks()
 | |
| 
 | |
| 	g := errgroup.WithNErrs(len(storageDisks))
 | |
| 
 | |
| 	for index := range storageDisks {
 | |
| 		index := index
 | |
| 		g.Go(func() error {
 | |
| 			if storageDisks[index] == nil {
 | |
| 				return errDiskNotFound
 | |
| 			}
 | |
| 			// Check if 'prefix' is an object on this 'disk', else continue the check the next disk
 | |
| 			return storageDisks[index].CheckFile(ctx, bucket, prefix)
 | |
| 		}, index)
 | |
| 	}
 | |
| 
 | |
| 	// NOTE: Observe we are not trying to read `xl.meta` and figure out the actual
 | |
| 	// quorum intentionally, but rely on the default case scenario. Actual quorum
 | |
| 	// verification will happen by top layer by using getObjectInfo() and will be
 | |
| 	// ignored if necessary.
 | |
| 	readQuorum := getReadQuorum(len(storageDisks))
 | |
| 
 | |
| 	err := reduceReadQuorumErrs(ctx, g.Wait(), objectOpIgnoredErrs, readQuorum)
 | |
| 	return err == nil, err == errPathNotFound
 | |
| }
 |