mirror of
				https://github.com/minio/minio.git
				synced 2025-11-04 02:01:05 +01:00 
			
		
		
		
	With this change, MinIO's ILM supports transitioning objects to a remote tier. This change includes support for Azure Blob Storage, AWS S3 compatible object storage incl. MinIO and Google Cloud Storage as remote tier storage backends. Some new additions include: - Admin APIs remote tier configuration management - Simple journal to track remote objects to be 'collected' This is used by object API handlers which 'mutate' object versions by overwriting/replacing content (Put/CopyObject) or removing the version itself (e.g DeleteObjectVersion). - Rework of previous ILM transition to fit the new model In the new model, a storage class (a.k.a remote tier) is defined by the 'remote' object storage type (one of s3, azure, GCS), bucket name and a prefix. * Fixed bugs, review comments, and more unit-tests - Leverage inline small object feature - Migrate legacy objects to the latest object format before transitioning - Fix restore to particular version if specified - Extend SharedDataDirCount to handle transitioned and restored objects - Restore-object should accept version-id for version-suspended bucket (#12091) - Check if remote tier creds have sufficient permissions - Bonus minor fixes to existing error messages Co-authored-by: Poorna Krishnamoorthy <poorna@minio.io> Co-authored-by: Krishna Srinivas <krishna@minio.io> Signed-off-by: Harshavardhana <harsha@minio.io>
		
			
				
	
	
		
			104 lines
		
	
	
		
			3.1 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			104 lines
		
	
	
		
			3.1 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
// Copyright (c) 2015-2021 MinIO, Inc.
 | 
						|
//
 | 
						|
// This file is part of MinIO Object Storage stack
 | 
						|
//
 | 
						|
// This program is free software: you can redistribute it and/or modify
 | 
						|
// it under the terms of the GNU Affero General Public License as published by
 | 
						|
// the Free Software Foundation, either version 3 of the License, or
 | 
						|
// (at your option) any later version.
 | 
						|
//
 | 
						|
// This program is distributed in the hope that it will be useful
 | 
						|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
						|
// GNU Affero General Public License for more details.
 | 
						|
//
 | 
						|
// You should have received a copy of the GNU Affero General Public License
 | 
						|
// along with this program.  If not, see <http://www.gnu.org/licenses/>.
 | 
						|
 | 
						|
package cmd
 | 
						|
 | 
						|
import (
 | 
						|
	"context"
 | 
						|
	"runtime"
 | 
						|
	"testing"
 | 
						|
	"time"
 | 
						|
)
 | 
						|
 | 
						|
// WARNING:
 | 
						|
//
 | 
						|
// Expected source line number is hard coded, 35, in the
 | 
						|
// following test. Adding new code before this test or changing its
 | 
						|
// position will cause the line number to change and the test to FAIL
 | 
						|
// Tests getSource().
 | 
						|
func TestGetSource(t *testing.T) {
 | 
						|
	currentSource := func() string { return getSource(2) }
 | 
						|
	gotSource := currentSource()
 | 
						|
	// Hard coded line number, 35, in the "expectedSource" value
 | 
						|
	expectedSource := "[namespace-lock_test.go:35:TestGetSource()]"
 | 
						|
	if gotSource != expectedSource {
 | 
						|
		t.Errorf("expected : %s, got : %s", expectedSource, gotSource)
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
// Test lock race
 | 
						|
func TestNSLockRace(t *testing.T) {
 | 
						|
	t.Skip("long test skip it")
 | 
						|
 | 
						|
	ctx := context.Background()
 | 
						|
 | 
						|
	for i := 0; i < 10000; i++ {
 | 
						|
		nsLk := newNSLock(false)
 | 
						|
 | 
						|
		// lk1; ref=1
 | 
						|
		if !nsLk.lock(ctx, "volume", "path", "source", "opsID", false, time.Second) {
 | 
						|
			t.Fatal("failed to acquire lock")
 | 
						|
		}
 | 
						|
 | 
						|
		// lk2
 | 
						|
		lk2ch := make(chan struct{})
 | 
						|
		go func() {
 | 
						|
			defer close(lk2ch)
 | 
						|
			nsLk.lock(ctx, "volume", "path", "source", "opsID", false, 1*time.Millisecond)
 | 
						|
		}()
 | 
						|
		time.Sleep(1 * time.Millisecond) // wait for goroutine to advance; ref=2
 | 
						|
 | 
						|
		// Unlock the 1st lock; ref=1 after this line
 | 
						|
		nsLk.unlock("volume", "path", false)
 | 
						|
 | 
						|
		// Taking another lockMapMutex here allows queuing up additional lockers. This should
 | 
						|
		// not be required but makes reproduction much easier.
 | 
						|
		nsLk.lockMapMutex.Lock()
 | 
						|
 | 
						|
		// lk3 blocks.
 | 
						|
		lk3ch := make(chan bool)
 | 
						|
		go func() {
 | 
						|
			lk3ch <- nsLk.lock(ctx, "volume", "path", "source", "opsID", false, 0)
 | 
						|
		}()
 | 
						|
 | 
						|
		// lk4, blocks.
 | 
						|
		lk4ch := make(chan bool)
 | 
						|
		go func() {
 | 
						|
			lk4ch <- nsLk.lock(ctx, "volume", "path", "source", "opsID", false, 0)
 | 
						|
		}()
 | 
						|
		runtime.Gosched()
 | 
						|
 | 
						|
		// unlock the manual lock
 | 
						|
		nsLk.lockMapMutex.Unlock()
 | 
						|
 | 
						|
		// To trigger the race:
 | 
						|
		// 1) lk3 or lk4 need to advance and increment the ref on the existing resource,
 | 
						|
		//    successfully acquiring the lock.
 | 
						|
		// 2) lk2 then needs to advance and remove the resource from lockMap.
 | 
						|
		// 3) lk3 or lk4 (whichever didn't execute in step 1) then executes and creates
 | 
						|
		//    a new entry in lockMap and acquires a lock for the same resource.
 | 
						|
 | 
						|
		<-lk2ch
 | 
						|
		lk3ok := <-lk3ch
 | 
						|
		lk4ok := <-lk4ch
 | 
						|
 | 
						|
		if lk3ok && lk4ok {
 | 
						|
			t.Fatalf("multiple locks acquired; iteration=%d, lk3=%t, lk4=%t", i, lk3ok, lk4ok)
 | 
						|
		}
 | 
						|
	}
 | 
						|
}
 |