mirror of
				https://github.com/minio/minio.git
				synced 2025-10-31 00:01:27 +01:00 
			
		
		
		
	- optimize writing part.N.meta by writing both part.N
  and its meta in sequence without network component.
- remove part.N.meta, part.N which were partially success
  ful, in quorum loss situations during renamePart()
- allow for strict read quorum check arbitrated via ETag
  for the given part number, this makes it double safer
  upon final commit.
- return an appropriate error when read quorum is missing,
  instead of returning InvalidPart{}, which is non-retryable
  error. This kind of situation can happen when many
  nodes are going offline in rotation, an example of such
  a restart() behavior is statefulset updates in k8s.
fixes #20091
		
	
			
		
			
				
	
	
		
			86 lines
		
	
	
		
			2.3 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			86 lines
		
	
	
		
			2.3 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // Copyright (c) 2015-2021 MinIO, Inc.
 | |
| //
 | |
| // This file is part of MinIO Object Storage stack
 | |
| //
 | |
| // This program is free software: you can redistribute it and/or modify
 | |
| // it under the terms of the GNU Affero General Public License as published by
 | |
| // the Free Software Foundation, either version 3 of the License, or
 | |
| // (at your option) any later version.
 | |
| //
 | |
| // This program is distributed in the hope that it will be useful
 | |
| // but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
| // GNU Affero General Public License for more details.
 | |
| //
 | |
| // You should have received a copy of the GNU Affero General Public License
 | |
| // along with this program.  If not, see <http://www.gnu.org/licenses/>.
 | |
| 
 | |
| package cmd
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"math/rand"
 | |
| 	"sync"
 | |
| 	"time"
 | |
| )
 | |
| 
 | |
| func (er erasureObjects) getOnlineDisks() (newDisks []StorageAPI) {
 | |
| 	disks := er.getDisks()
 | |
| 	var wg sync.WaitGroup
 | |
| 	var mu sync.Mutex
 | |
| 	r := rand.New(rand.NewSource(time.Now().UnixNano()))
 | |
| 	for _, i := range r.Perm(len(disks)) {
 | |
| 		i := i
 | |
| 		wg.Add(1)
 | |
| 		go func() {
 | |
| 			defer wg.Done()
 | |
| 			if disks[i] == nil {
 | |
| 				return
 | |
| 			}
 | |
| 			di, err := disks[i].DiskInfo(context.Background(), DiskInfoOptions{})
 | |
| 			if err != nil || di.Healing {
 | |
| 				// - Do not consume disks which are not reachable
 | |
| 				//   unformatted or simply not accessible for some reason.
 | |
| 				//
 | |
| 				// - Do not consume disks which are being healed
 | |
| 				//
 | |
| 				// - Future: skip busy disks
 | |
| 				return
 | |
| 			}
 | |
| 
 | |
| 			mu.Lock()
 | |
| 			newDisks = append(newDisks, disks[i])
 | |
| 			mu.Unlock()
 | |
| 		}()
 | |
| 	}
 | |
| 	wg.Wait()
 | |
| 	return newDisks
 | |
| }
 | |
| 
 | |
| func (er erasureObjects) getOnlineLocalDisks() (newDisks []StorageAPI) {
 | |
| 	disks := er.getOnlineDisks()
 | |
| 
 | |
| 	// Based on the random shuffling return back randomized disks.
 | |
| 	r := rand.New(rand.NewSource(time.Now().UnixNano()))
 | |
| 
 | |
| 	for _, i := range r.Perm(len(disks)) {
 | |
| 		if disks[i] != nil && disks[i].IsLocal() {
 | |
| 			newDisks = append(newDisks, disks[i])
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return newDisks
 | |
| }
 | |
| 
 | |
| func (er erasureObjects) getLocalDisks() (newDisks []StorageAPI) {
 | |
| 	disks := er.getDisks()
 | |
| 	// Based on the random shuffling return back randomized disks.
 | |
| 	r := rand.New(rand.NewSource(time.Now().UnixNano()))
 | |
| 	for _, i := range r.Perm(len(disks)) {
 | |
| 		if disks[i] != nil && disks[i].IsLocal() {
 | |
| 			newDisks = append(newDisks, disks[i])
 | |
| 		}
 | |
| 	}
 | |
| 	return newDisks
 | |
| }
 |