mirror of
				https://github.com/minio/minio.git
				synced 2025-10-31 00:01:27 +01:00 
			
		
		
		
	* Provide information on *actively* healing, buckets healed/queued, objects healed/failed. * Add concurrent healing of multiple sets (typically on startup). * Add bucket level resume, so restarts will only heal non-healed buckets. * Print summary after healing a disk is done.
		
			
				
	
	
		
			185 lines
		
	
	
		
			4.9 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			185 lines
		
	
	
		
			4.9 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| /*
 | |
|  * MinIO Cloud Storage, (C) 2016-2020 MinIO, Inc.
 | |
|  *
 | |
|  * Licensed under the Apache License, Version 2.0 (the "License");
 | |
|  * you may not use this file except in compliance with the License.
 | |
|  * You may obtain a copy of the License at
 | |
|  *
 | |
|  *     http://www.apache.org/licenses/LICENSE-2.0
 | |
|  *
 | |
|  * Unless required by applicable law or agreed to in writing, software
 | |
|  * distributed under the License is distributed on an "AS IS" BASIS,
 | |
|  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
|  * See the License for the specific language governing permissions and
 | |
|  * limitations under the License.
 | |
|  */
 | |
| 
 | |
| package cmd
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"sync"
 | |
| 
 | |
| 	"github.com/minio/minio/pkg/sync/errgroup"
 | |
| )
 | |
| 
 | |
| func (er erasureObjects) getLocalDisks() (localDisks []StorageAPI) {
 | |
| 	disks := er.getDisks()
 | |
| 	for _, disk := range disks {
 | |
| 		if disk != nil && disk.IsLocal() {
 | |
| 			localDisks = append(localDisks, disk)
 | |
| 		}
 | |
| 	}
 | |
| 	return localDisks
 | |
| }
 | |
| 
 | |
| func (er erasureObjects) getLoadBalancedLocalDisks() (newDisks []StorageAPI) {
 | |
| 	disks := er.getDisks()
 | |
| 	// Based on the random shuffling return back randomized disks.
 | |
| 	for _, i := range hashOrder(UTCNow().String(), len(disks)) {
 | |
| 		if disks[i-1] != nil && disks[i-1].IsLocal() {
 | |
| 			if disks[i-1].Healing() == nil && disks[i-1].IsOnline() {
 | |
| 				newDisks = append(newDisks, disks[i-1])
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	return newDisks
 | |
| }
 | |
| 
 | |
| func (er erasureObjects) getOnlineDisks() (newDisks []StorageAPI) {
 | |
| 	disks := er.getDisks()
 | |
| 	var wg sync.WaitGroup
 | |
| 	var mu sync.Mutex
 | |
| 	for _, i := range hashOrder(UTCNow().String(), len(disks)) {
 | |
| 		i := i
 | |
| 		wg.Add(1)
 | |
| 		go func() {
 | |
| 			defer wg.Done()
 | |
| 			if disks[i-1] == nil {
 | |
| 				return
 | |
| 			}
 | |
| 			di, err := disks[i-1].DiskInfo(context.Background())
 | |
| 			if err != nil || di.Healing {
 | |
| 
 | |
| 				// - Do not consume disks which are not reachable
 | |
| 				//   unformatted or simply not accessible for some reason.
 | |
| 				//
 | |
| 				// - Do not consume disks which are being healed
 | |
| 				//
 | |
| 				// - Future: skip busy disks
 | |
| 				return
 | |
| 			}
 | |
| 
 | |
| 			mu.Lock()
 | |
| 			newDisks = append(newDisks, disks[i-1])
 | |
| 			mu.Unlock()
 | |
| 		}()
 | |
| 	}
 | |
| 	wg.Wait()
 | |
| 	return newDisks
 | |
| }
 | |
| 
 | |
| // getLoadBalancedNDisks - fetches load balanced (sufficiently randomized) disk slice
 | |
| // with N disks online. If ndisks is zero or negative, then it will returns all disks,
 | |
| // same if ndisks is greater than the number of all disks.
 | |
| func (er erasureObjects) getLoadBalancedNDisks(ndisks int) (newDisks []StorageAPI) {
 | |
| 	disks := er.getLoadBalancedDisks(ndisks != -1)
 | |
| 	for _, disk := range disks {
 | |
| 		if disk == nil {
 | |
| 			continue
 | |
| 		}
 | |
| 		newDisks = append(newDisks, disk)
 | |
| 		ndisks--
 | |
| 		if ndisks == 0 {
 | |
| 			break
 | |
| 		}
 | |
| 	}
 | |
| 	return
 | |
| }
 | |
| 
 | |
| // getLoadBalancedDisks - fetches load balanced (sufficiently randomized) disk slice.
 | |
| // ensures to skip disks if they are not healing and online.
 | |
| func (er erasureObjects) getLoadBalancedDisks(optimized bool) []StorageAPI {
 | |
| 	disks := er.getDisks()
 | |
| 
 | |
| 	if !optimized {
 | |
| 		var newDisks []StorageAPI
 | |
| 		for _, i := range hashOrder(UTCNow().String(), len(disks)) {
 | |
| 			newDisks = append(newDisks, disks[i-1])
 | |
| 		}
 | |
| 		return newDisks
 | |
| 	}
 | |
| 
 | |
| 	var wg sync.WaitGroup
 | |
| 	var mu sync.Mutex
 | |
| 	var newDisks = map[uint64][]StorageAPI{}
 | |
| 	// Based on the random shuffling return back randomized disks.
 | |
| 	for _, i := range hashOrder(UTCNow().String(), len(disks)) {
 | |
| 		i := i
 | |
| 		wg.Add(1)
 | |
| 		go func() {
 | |
| 			defer wg.Done()
 | |
| 			if disks[i-1] == nil {
 | |
| 				return
 | |
| 			}
 | |
| 			di, err := disks[i-1].DiskInfo(context.Background())
 | |
| 			if err != nil || di.Healing {
 | |
| 				// - Do not consume disks which are not reachable
 | |
| 				//   unformatted or simply not accessible for some reason.
 | |
| 				//
 | |
| 				// - Do not consume disks which are being healed
 | |
| 				//
 | |
| 				// - Future: skip busy disks
 | |
| 				return
 | |
| 			}
 | |
| 
 | |
| 			mu.Lock()
 | |
| 			// Capture disks usage wise upto resolution of MiB
 | |
| 			newDisks[di.Used/1024/1024] = append(newDisks[di.Used/1024/1024], disks[i-1])
 | |
| 			mu.Unlock()
 | |
| 		}()
 | |
| 	}
 | |
| 	wg.Wait()
 | |
| 
 | |
| 	var max uint64
 | |
| 	for k := range newDisks {
 | |
| 		if k > max {
 | |
| 			max = k
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// Return disks which have maximum disk usage common.
 | |
| 	return newDisks[max]
 | |
| }
 | |
| 
 | |
| // This function does the following check, suppose
 | |
| // object is "a/b/c/d", stat makes sure that objects
 | |
| // - "a/b/c"
 | |
| // - "a/b"
 | |
| // - "a"
 | |
| // do not exist on the namespace.
 | |
| func (er erasureObjects) parentDirIsObject(ctx context.Context, bucket, parent string) bool {
 | |
| 	storageDisks := er.getDisks()
 | |
| 
 | |
| 	g := errgroup.WithNErrs(len(storageDisks))
 | |
| 
 | |
| 	for index := range storageDisks {
 | |
| 		index := index
 | |
| 		g.Go(func() error {
 | |
| 			if storageDisks[index] == nil {
 | |
| 				return errDiskNotFound
 | |
| 			}
 | |
| 			// Check if 'prefix' is an object on this 'disk', else continue the check the next disk
 | |
| 			return storageDisks[index].CheckFile(ctx, bucket, parent)
 | |
| 		}, index)
 | |
| 	}
 | |
| 
 | |
| 	// NOTE: Observe we are not trying to read `xl.meta` and figure out the actual
 | |
| 	// quorum intentionally, but rely on the default case scenario. Actual quorum
 | |
| 	// verification will happen by top layer by using getObjectInfo() and will be
 | |
| 	// ignored if necessary.
 | |
| 	readQuorum := getReadQuorum(len(storageDisks))
 | |
| 
 | |
| 	return reduceReadQuorumErrs(ctx, g.Wait(), objectOpIgnoredErrs, readQuorum) == nil
 | |
| }
 |