mirror of
				https://github.com/minio/minio.git
				synced 2025-10-31 00:01:27 +01:00 
			
		
		
		
	With this change, MinIO's ILM supports transitioning objects to a remote tier. This change includes support for Azure Blob Storage, AWS S3 compatible object storage incl. MinIO and Google Cloud Storage as remote tier storage backends. Some new additions include: - Admin APIs remote tier configuration management - Simple journal to track remote objects to be 'collected' This is used by object API handlers which 'mutate' object versions by overwriting/replacing content (Put/CopyObject) or removing the version itself (e.g DeleteObjectVersion). - Rework of previous ILM transition to fit the new model In the new model, a storage class (a.k.a remote tier) is defined by the 'remote' object storage type (one of s3, azure, GCS), bucket name and a prefix. * Fixed bugs, review comments, and more unit-tests - Leverage inline small object feature - Migrate legacy objects to the latest object format before transitioning - Fix restore to particular version if specified - Extend SharedDataDirCount to handle transitioned and restored objects - Restore-object should accept version-id for version-suspended bucket (#12091) - Check if remote tier creds have sufficient permissions - Bonus minor fixes to existing error messages Co-authored-by: Poorna Krishnamoorthy <poorna@minio.io> Co-authored-by: Krishna Srinivas <krishna@minio.io> Signed-off-by: Harshavardhana <harsha@minio.io>
		
			
				
	
	
		
			104 lines
		
	
	
		
			3.1 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			104 lines
		
	
	
		
			3.1 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // Copyright (c) 2015-2021 MinIO, Inc.
 | |
| //
 | |
| // This file is part of MinIO Object Storage stack
 | |
| //
 | |
| // This program is free software: you can redistribute it and/or modify
 | |
| // it under the terms of the GNU Affero General Public License as published by
 | |
| // the Free Software Foundation, either version 3 of the License, or
 | |
| // (at your option) any later version.
 | |
| //
 | |
| // This program is distributed in the hope that it will be useful
 | |
| // but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
| // GNU Affero General Public License for more details.
 | |
| //
 | |
| // You should have received a copy of the GNU Affero General Public License
 | |
| // along with this program.  If not, see <http://www.gnu.org/licenses/>.
 | |
| 
 | |
| package cmd
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"runtime"
 | |
| 	"testing"
 | |
| 	"time"
 | |
| )
 | |
| 
 | |
| // WARNING:
 | |
| //
 | |
| // Expected source line number is hard coded, 35, in the
 | |
| // following test. Adding new code before this test or changing its
 | |
| // position will cause the line number to change and the test to FAIL
 | |
| // Tests getSource().
 | |
| func TestGetSource(t *testing.T) {
 | |
| 	currentSource := func() string { return getSource(2) }
 | |
| 	gotSource := currentSource()
 | |
| 	// Hard coded line number, 35, in the "expectedSource" value
 | |
| 	expectedSource := "[namespace-lock_test.go:35:TestGetSource()]"
 | |
| 	if gotSource != expectedSource {
 | |
| 		t.Errorf("expected : %s, got : %s", expectedSource, gotSource)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // Test lock race
 | |
| func TestNSLockRace(t *testing.T) {
 | |
| 	t.Skip("long test skip it")
 | |
| 
 | |
| 	ctx := context.Background()
 | |
| 
 | |
| 	for i := 0; i < 10000; i++ {
 | |
| 		nsLk := newNSLock(false)
 | |
| 
 | |
| 		// lk1; ref=1
 | |
| 		if !nsLk.lock(ctx, "volume", "path", "source", "opsID", false, time.Second) {
 | |
| 			t.Fatal("failed to acquire lock")
 | |
| 		}
 | |
| 
 | |
| 		// lk2
 | |
| 		lk2ch := make(chan struct{})
 | |
| 		go func() {
 | |
| 			defer close(lk2ch)
 | |
| 			nsLk.lock(ctx, "volume", "path", "source", "opsID", false, 1*time.Millisecond)
 | |
| 		}()
 | |
| 		time.Sleep(1 * time.Millisecond) // wait for goroutine to advance; ref=2
 | |
| 
 | |
| 		// Unlock the 1st lock; ref=1 after this line
 | |
| 		nsLk.unlock("volume", "path", false)
 | |
| 
 | |
| 		// Taking another lockMapMutex here allows queuing up additional lockers. This should
 | |
| 		// not be required but makes reproduction much easier.
 | |
| 		nsLk.lockMapMutex.Lock()
 | |
| 
 | |
| 		// lk3 blocks.
 | |
| 		lk3ch := make(chan bool)
 | |
| 		go func() {
 | |
| 			lk3ch <- nsLk.lock(ctx, "volume", "path", "source", "opsID", false, 0)
 | |
| 		}()
 | |
| 
 | |
| 		// lk4, blocks.
 | |
| 		lk4ch := make(chan bool)
 | |
| 		go func() {
 | |
| 			lk4ch <- nsLk.lock(ctx, "volume", "path", "source", "opsID", false, 0)
 | |
| 		}()
 | |
| 		runtime.Gosched()
 | |
| 
 | |
| 		// unlock the manual lock
 | |
| 		nsLk.lockMapMutex.Unlock()
 | |
| 
 | |
| 		// To trigger the race:
 | |
| 		// 1) lk3 or lk4 need to advance and increment the ref on the existing resource,
 | |
| 		//    successfully acquiring the lock.
 | |
| 		// 2) lk2 then needs to advance and remove the resource from lockMap.
 | |
| 		// 3) lk3 or lk4 (whichever didn't execute in step 1) then executes and creates
 | |
| 		//    a new entry in lockMap and acquires a lock for the same resource.
 | |
| 
 | |
| 		<-lk2ch
 | |
| 		lk3ok := <-lk3ch
 | |
| 		lk4ok := <-lk4ch
 | |
| 
 | |
| 		if lk3ok && lk4ok {
 | |
| 			t.Fatalf("multiple locks acquired; iteration=%d, lk3=%t, lk4=%t", i, lk3ok, lk4ok)
 | |
| 		}
 | |
| 	}
 | |
| }
 |