2017-12-09 22:54:26 +01:00
|
|
|
// +build !plan9,go1.7
|
2017-11-12 18:54:25 +01:00
|
|
|
|
|
|
|
package cache_test
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"io"
|
|
|
|
"io/ioutil"
|
2018-01-29 23:05:04 +01:00
|
|
|
"log"
|
2017-11-12 18:54:25 +01:00
|
|
|
"math/rand"
|
2018-01-29 23:05:04 +01:00
|
|
|
"os"
|
2017-11-12 18:54:25 +01:00
|
|
|
"path"
|
2017-11-20 15:38:28 +01:00
|
|
|
"path/filepath"
|
2017-12-18 13:55:37 +01:00
|
|
|
"runtime"
|
2017-11-12 18:54:25 +01:00
|
|
|
"strconv"
|
2018-01-29 23:05:04 +01:00
|
|
|
"strings"
|
2017-11-12 18:54:25 +01:00
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
"github.com/pkg/errors"
|
|
|
|
|
|
|
|
"encoding/base64"
|
|
|
|
goflag "flag"
|
|
|
|
"fmt"
|
|
|
|
"runtime/debug"
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-11 17:05:41 +01:00
|
|
|
"github.com/ncw/rclone/backend/cache"
|
2018-01-29 23:05:04 +01:00
|
|
|
"github.com/ncw/rclone/backend/crypt"
|
2018-01-11 17:05:41 +01:00
|
|
|
_ "github.com/ncw/rclone/backend/drive"
|
|
|
|
"github.com/ncw/rclone/backend/local"
|
2017-11-12 18:54:25 +01:00
|
|
|
"github.com/ncw/rclone/fs"
|
2018-01-29 23:05:04 +01:00
|
|
|
"github.com/ncw/rclone/fs/config"
|
|
|
|
"github.com/ncw/rclone/fs/object"
|
2017-11-12 18:54:25 +01:00
|
|
|
"github.com/ncw/rclone/fstest"
|
2018-01-29 23:05:04 +01:00
|
|
|
"github.com/ncw/rclone/vfs"
|
|
|
|
"github.com/ncw/rclone/vfs/vfsflags"
|
2017-11-12 18:54:25 +01:00
|
|
|
flag "github.com/spf13/pflag"
|
|
|
|
"github.com/stretchr/testify/require"
|
|
|
|
)
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
const (
|
|
|
|
// these 2 passwords are test random
|
|
|
|
cryptPassword1 = "3XcvMMdsV3d-HGAReTMdNH-5FcX5q32_lUeA" // oGJdUbQc7s8
|
|
|
|
cryptPassword2 = "NlgTBEIe-qibA7v-FoMfuX6Cw8KlLai_aMvV" // mv4mZW572HM
|
|
|
|
cryptedTextBase64 = "UkNMT05FAAC320i2xIee0BiNyknSPBn+Qcw3q9FhIFp3tvq6qlqvbsno3PnxmEFeJG3jDBnR/wku2gHWeQ==" // one content
|
|
|
|
cryptedText2Base64 = "UkNMT05FAAATcQkVsgjBh8KafCKcr0wdTa1fMmV0U8hsCLGFoqcvxKVmvv7wx3Hf5EXxFcki2FFV4sdpmSrb9Q==" // updated content
|
|
|
|
)
|
|
|
|
|
2017-11-12 18:54:25 +01:00
|
|
|
var (
|
2018-01-29 23:05:04 +01:00
|
|
|
remoteName string
|
|
|
|
mountDir string
|
|
|
|
uploadDir string
|
|
|
|
useMount bool
|
|
|
|
runInstance *run
|
|
|
|
errNotSupported = errors.New("not supported")
|
|
|
|
decryptedToEncryptedRemotes = map[string]string{
|
|
|
|
"one": "lm4u7jjt3c85bf56vjqgeenuno",
|
|
|
|
"second": "qvt1ochrkcfbptp5mu9ugb2l14",
|
|
|
|
"test": "jn4tegjtpqro30t3o11thb4b5s",
|
|
|
|
"test2": "qakvqnh8ttei89e0gc76crpql4",
|
|
|
|
"data.bin": "0q2847tfko6mhj3dag3r809qbc",
|
|
|
|
"ticw/data.bin": "5mv97b0ule6pht33srae5pice8/0q2847tfko6mhj3dag3r809qbc",
|
|
|
|
"tiutfo/test/one": "legd371aa8ol36tjfklt347qnc/jn4tegjtpqro30t3o11thb4b5s/lm4u7jjt3c85bf56vjqgeenuno",
|
|
|
|
"tiuufo/test/one": "vi6u1olqhirqv14cd8qlej1mgo/jn4tegjtpqro30t3o11thb4b5s/lm4u7jjt3c85bf56vjqgeenuno",
|
|
|
|
"tiutfo/second/one": "legd371aa8ol36tjfklt347qnc/qvt1ochrkcfbptp5mu9ugb2l14/lm4u7jjt3c85bf56vjqgeenuno",
|
|
|
|
"second/one": "qvt1ochrkcfbptp5mu9ugb2l14/lm4u7jjt3c85bf56vjqgeenuno",
|
|
|
|
"test/one": "jn4tegjtpqro30t3o11thb4b5s/lm4u7jjt3c85bf56vjqgeenuno",
|
|
|
|
"test/second": "jn4tegjtpqro30t3o11thb4b5s/qvt1ochrkcfbptp5mu9ugb2l14",
|
|
|
|
"test/third": "jn4tegjtpqro30t3o11thb4b5s/2nd7fjiop5h3ihfj1vl953aa5g",
|
|
|
|
"test/0.bin": "jn4tegjtpqro30t3o11thb4b5s/e6frddt058b6kvbpmlstlndmtk",
|
|
|
|
"test/1.bin": "jn4tegjtpqro30t3o11thb4b5s/kck472nt1k7qbmob0mt1p1crgc",
|
|
|
|
"test/2.bin": "jn4tegjtpqro30t3o11thb4b5s/744oe9ven2rmak4u27if51qk24",
|
|
|
|
"test/3.bin": "jn4tegjtpqro30t3o11thb4b5s/2bjd8kef0u5lmsu6qhqll34bcs",
|
|
|
|
"test/4.bin": "jn4tegjtpqro30t3o11thb4b5s/cvjs73iv0a82v0c7r67avllh7s",
|
|
|
|
"test/5.bin": "jn4tegjtpqro30t3o11thb4b5s/0plkdo790b6bnmt33qsdqmhv9c",
|
|
|
|
"test/6.bin": "jn4tegjtpqro30t3o11thb4b5s/s5r633srnjtbh83893jovjt5d0",
|
|
|
|
"test/7.bin": "jn4tegjtpqro30t3o11thb4b5s/6rq45tr9bjsammku622flmqsu4",
|
|
|
|
"test/8.bin": "jn4tegjtpqro30t3o11thb4b5s/37bc6tcl3e31qb8cadvjb749vk",
|
|
|
|
"test/9.bin": "jn4tegjtpqro30t3o11thb4b5s/t4pr35hnls32789o8fk0chk1ec",
|
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
)
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func init() {
|
|
|
|
goflag.StringVar(&remoteName, "remote-internal", "TestCache", "Remote to test with, defaults to local filesystem")
|
|
|
|
goflag.StringVar(&mountDir, "mount-dir-internal", "", "")
|
|
|
|
goflag.StringVar(&uploadDir, "upload-dir-internal", "", "")
|
|
|
|
goflag.BoolVar(&useMount, "cache-use-mount", false, "Test only with mount")
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestMain drives the tests
|
|
|
|
func TestMain(m *testing.M) {
|
|
|
|
goflag.Parse()
|
|
|
|
var rc int
|
|
|
|
|
|
|
|
runInstance = newRun()
|
|
|
|
rc = m.Run()
|
|
|
|
os.Exit(rc)
|
|
|
|
}
|
|
|
|
|
2017-11-12 18:54:25 +01:00
|
|
|
func TestInternalListRootAndInnerRemotes(t *testing.T) {
|
2018-01-29 23:05:04 +01:00
|
|
|
id := fmt.Sprintf("tilrair%v", time.Now().Unix())
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, true, true, nil, nil)
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2017-11-12 18:54:25 +01:00
|
|
|
// Instantiate inner fs
|
|
|
|
innerFolder := "inner"
|
2018-01-29 23:05:04 +01:00
|
|
|
runInstance.mkdir(t, rootFs, innerFolder)
|
|
|
|
rootFs2, boltDb2 := runInstance.newCacheFs(t, remoteName, id+"/"+innerFolder, true, true, nil, nil)
|
|
|
|
defer runInstance.cleanupFs(t, rootFs2, boltDb2)
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
runInstance.writeObjectString(t, rootFs2, "one", "content")
|
|
|
|
listRoot := runInstance.list(t, rootFs, "")
|
|
|
|
listRootInner := runInstance.list(t, rootFs, innerFolder)
|
|
|
|
listInner, err := rootFs2.List("")
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
require.Len(t, listRoot, 1)
|
|
|
|
require.Len(t, listRootInner, 1)
|
|
|
|
require.Len(t, listInner, 1)
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
func TestInternalObjWrapFsFound(t *testing.T) {
|
2018-01-29 23:05:04 +01:00
|
|
|
id := fmt.Sprintf("tiowff%v", time.Now().Unix())
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, true, true, nil, nil)
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
cfs, err := runInstance.getCacheFs(rootFs)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
wrappedFs := cfs.UnWrap()
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
var testData []byte
|
|
|
|
if runInstance.rootIsCrypt {
|
|
|
|
testData, err = base64.StdEncoding.DecodeString(cryptedTextBase64)
|
|
|
|
require.NoError(t, err)
|
|
|
|
} else {
|
|
|
|
testData = []byte("test content")
|
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
runInstance.writeObjectBytes(t, wrappedFs, runInstance.encryptRemoteIfNeeded(t, "test"), testData)
|
|
|
|
listRoot := runInstance.list(t, rootFs, "")
|
|
|
|
require.Len(t, listRoot, 1)
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
cachedData := runInstance.readDataFromRemote(t, rootFs, "test", 0, int64(len([]byte("test content"))), false)
|
|
|
|
require.Equal(t, "test content", string(cachedData))
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
err = runInstance.rm(t, rootFs, "test")
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
listRoot = runInstance.list(t, rootFs, "")
|
|
|
|
require.Len(t, listRoot, 0)
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
func TestInternalObjNotFound(t *testing.T) {
|
2018-01-29 23:05:04 +01:00
|
|
|
id := fmt.Sprintf("tionf%v", time.Now().Unix())
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, false, true, nil, nil)
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2017-11-12 18:54:25 +01:00
|
|
|
obj, err := rootFs.NewObject("404")
|
|
|
|
require.Error(t, err)
|
|
|
|
require.Nil(t, obj)
|
|
|
|
}
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func TestInternalRemoteWrittenFileFoundInMount(t *testing.T) {
|
|
|
|
if !runInstance.useMount {
|
|
|
|
t.Skip("test needs mount mode")
|
|
|
|
}
|
|
|
|
id := fmt.Sprintf("tirwffim%v", time.Now().Unix())
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, true, true, nil, nil)
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
|
|
|
|
|
|
|
cfs, err := runInstance.getCacheFs(rootFs)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
var testData []byte
|
|
|
|
if runInstance.rootIsCrypt {
|
|
|
|
testData, err = base64.StdEncoding.DecodeString(cryptedTextBase64)
|
|
|
|
require.NoError(t, err)
|
|
|
|
} else {
|
|
|
|
testData = []byte("test content")
|
|
|
|
}
|
|
|
|
|
|
|
|
runInstance.writeObjectBytes(t, cfs.UnWrap(), runInstance.encryptRemoteIfNeeded(t, "test"), testData)
|
|
|
|
data := runInstance.readDataFromRemote(t, rootFs, "test", 0, int64(len([]byte("test content"))), false)
|
|
|
|
require.Equal(t, "test content", string(data))
|
|
|
|
}
|
|
|
|
|
2017-11-12 18:54:25 +01:00
|
|
|
func TestInternalCachedWrittenContentMatches(t *testing.T) {
|
2018-01-29 23:05:04 +01:00
|
|
|
id := fmt.Sprintf("ticwcm%v", time.Now().Unix())
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, false, true, nil, nil)
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
cfs, err := runInstance.getCacheFs(rootFs)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
chunkSize := cfs.ChunkSize()
|
|
|
|
|
|
|
|
// create some rand test data
|
2018-01-29 23:05:04 +01:00
|
|
|
testData := runInstance.randomBytes(t, chunkSize*4+chunkSize/2)
|
2017-11-12 18:54:25 +01:00
|
|
|
|
|
|
|
// write the object
|
2018-01-29 23:05:04 +01:00
|
|
|
runInstance.writeRemoteBytes(t, rootFs, "data.bin", testData)
|
2017-11-12 18:54:25 +01:00
|
|
|
|
|
|
|
// check sample of data from in-file
|
|
|
|
sampleStart := chunkSize / 2
|
|
|
|
sampleEnd := chunkSize
|
|
|
|
testSample := testData[sampleStart:sampleEnd]
|
2018-01-29 23:05:04 +01:00
|
|
|
checkSample := runInstance.readDataFromRemote(t, rootFs, "data.bin", sampleStart, sampleEnd, false)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.Equal(t, int64(len(checkSample)), sampleEnd-sampleStart)
|
|
|
|
require.Equal(t, checkSample, testSample)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestInternalCachedUpdatedContentMatches(t *testing.T) {
|
2018-01-29 23:05:04 +01:00
|
|
|
id := fmt.Sprintf("ticucm%v", time.Now().Unix())
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, false, true, nil, nil)
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
|
|
|
var err error
|
2017-11-12 18:54:25 +01:00
|
|
|
|
|
|
|
// create some rand test data
|
2018-01-29 23:05:04 +01:00
|
|
|
var testData1 []byte
|
|
|
|
var testData2 []byte
|
|
|
|
if runInstance.rootIsCrypt {
|
|
|
|
testData1, err = base64.StdEncoding.DecodeString(cryptedTextBase64)
|
|
|
|
require.NoError(t, err)
|
|
|
|
testData2, err = base64.StdEncoding.DecodeString(cryptedText2Base64)
|
|
|
|
require.NoError(t, err)
|
|
|
|
} else {
|
|
|
|
testData1 = []byte(fstest.RandomString(100))
|
|
|
|
testData2 = []byte(fstest.RandomString(200))
|
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
|
|
|
|
// write the object
|
2018-01-29 23:05:04 +01:00
|
|
|
o := runInstance.updateObjectRemote(t, rootFs, "data.bin", testData1, testData2)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.Equal(t, o.Size(), int64(len(testData2)))
|
|
|
|
|
|
|
|
// check data from in-file
|
2018-01-29 23:05:04 +01:00
|
|
|
checkSample := runInstance.readDataFromRemote(t, rootFs, "data.bin", 0, int64(len(testData2)), false)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.Equal(t, checkSample, testData2)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestInternalWrappedWrittenContentMatches(t *testing.T) {
|
2018-01-29 23:05:04 +01:00
|
|
|
id := fmt.Sprintf("tiwwcm%v", time.Now().Unix())
|
|
|
|
vfsflags.Opt.DirCacheTime = time.Second
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, true, true, nil, nil)
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
|
|
|
if runInstance.rootIsCrypt {
|
|
|
|
t.Skip("test skipped with crypt remote")
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
cfs, err := runInstance.getCacheFs(rootFs)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
chunkSize := cfs.ChunkSize()
|
|
|
|
|
|
|
|
// create some rand test data
|
2018-01-29 23:05:04 +01:00
|
|
|
testSize := chunkSize*4 + chunkSize/2
|
|
|
|
testData := runInstance.randomBytes(t, testSize)
|
2017-11-12 18:54:25 +01:00
|
|
|
|
|
|
|
// write the object
|
2018-01-29 23:05:04 +01:00
|
|
|
o := runInstance.writeObjectBytes(t, cfs.UnWrap(), "data.bin", testData)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.Equal(t, o.Size(), int64(testSize))
|
2018-01-29 23:05:04 +01:00
|
|
|
time.Sleep(time.Second * 3)
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
data2 := runInstance.readDataFromRemote(t, rootFs, "data.bin", 0, int64(testSize), false)
|
|
|
|
require.Equal(t, int64(len(data2)), o.Size())
|
2017-11-12 18:54:25 +01:00
|
|
|
|
|
|
|
// check sample of data from in-file
|
|
|
|
sampleStart := chunkSize / 2
|
|
|
|
sampleEnd := chunkSize
|
|
|
|
testSample := testData[sampleStart:sampleEnd]
|
2018-01-29 23:05:04 +01:00
|
|
|
checkSample := runInstance.readDataFromRemote(t, rootFs, "data.bin", sampleStart, sampleEnd, false)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.Equal(t, len(checkSample), len(testSample))
|
|
|
|
|
|
|
|
for i := 0; i < len(checkSample); i++ {
|
|
|
|
require.Equal(t, testSample[i], checkSample[i])
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestInternalLargeWrittenContentMatches(t *testing.T) {
|
2018-01-29 23:05:04 +01:00
|
|
|
id := fmt.Sprintf("tilwcm%v", time.Now().Unix())
|
|
|
|
vfsflags.Opt.DirCacheTime = time.Second
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, true, true, nil, nil)
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
|
|
|
if runInstance.rootIsCrypt {
|
|
|
|
t.Skip("test skipped with crypt remote")
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
cfs, err := runInstance.getCacheFs(rootFs)
|
2017-12-18 13:55:37 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
chunkSize := cfs.ChunkSize()
|
|
|
|
|
|
|
|
// create some rand test data
|
2018-01-29 23:05:04 +01:00
|
|
|
testSize := chunkSize*10 + chunkSize/2
|
|
|
|
testData := runInstance.randomBytes(t, testSize)
|
2017-12-18 13:55:37 +01:00
|
|
|
|
|
|
|
// write the object
|
2018-01-29 23:05:04 +01:00
|
|
|
runInstance.writeObjectBytes(t, cfs.UnWrap(), "data.bin", testData)
|
|
|
|
time.Sleep(time.Second * 3)
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
readData := runInstance.readDataFromRemote(t, rootFs, "data.bin", 0, testSize, false)
|
|
|
|
for i := 0; i < len(readData); i++ {
|
|
|
|
require.Equalf(t, testData[i], readData[i], "at byte %v", i)
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-11-12 18:54:25 +01:00
|
|
|
func TestInternalWrappedFsChangeNotSeen(t *testing.T) {
|
2018-01-29 23:05:04 +01:00
|
|
|
id := fmt.Sprintf("tiwfcns%v", time.Now().Unix())
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, false, true, nil, nil)
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
cfs, err := runInstance.getCacheFs(rootFs)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
chunkSize := cfs.ChunkSize()
|
|
|
|
|
|
|
|
// create some rand test data
|
2018-01-29 23:05:04 +01:00
|
|
|
testData := runInstance.randomBytes(t, (chunkSize*4 + chunkSize/2))
|
|
|
|
runInstance.writeRemoteBytes(t, rootFs, "data.bin", testData)
|
2017-11-12 18:54:25 +01:00
|
|
|
|
|
|
|
// update in the wrapped fs
|
2018-01-29 23:05:04 +01:00
|
|
|
o, err := cfs.UnWrap().NewObject(runInstance.encryptRemoteIfNeeded(t, "data.bin"))
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
wrappedTime := time.Now().Add(time.Hour * -1)
|
|
|
|
err = o.SetModTime(wrappedTime)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// get a new instance from the cache
|
2018-01-29 23:05:04 +01:00
|
|
|
co, err := rootFs.NewObject("data.bin")
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
require.NotEqual(t, co.ModTime().String(), o.ModTime().String())
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
func TestInternalChangeSeenAfterDirCacheFlush(t *testing.T) {
|
2018-01-29 23:05:04 +01:00
|
|
|
id := fmt.Sprintf("ticsadcf%v", time.Now().Unix())
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, false, true, nil, nil)
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
cfs, err := runInstance.getCacheFs(rootFs)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2017-12-18 13:55:37 +01:00
|
|
|
chunkSize := cfs.ChunkSize()
|
|
|
|
|
|
|
|
// create some rand test data
|
2018-01-29 23:05:04 +01:00
|
|
|
testData := runInstance.randomBytes(t, (chunkSize*4 + chunkSize/2))
|
|
|
|
runInstance.writeRemoteBytes(t, rootFs, "data.bin", testData)
|
2017-12-18 13:55:37 +01:00
|
|
|
|
|
|
|
// update in the wrapped fs
|
2018-01-29 23:05:04 +01:00
|
|
|
o, err := cfs.UnWrap().NewObject(runInstance.encryptRemoteIfNeeded(t, "data.bin"))
|
2017-12-18 13:55:37 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
wrappedTime := time.Now().Add(-1 * time.Hour)
|
|
|
|
err = o.SetModTime(wrappedTime)
|
2017-12-18 13:55:37 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// get a new instance from the cache
|
2018-01-29 23:05:04 +01:00
|
|
|
co, err := rootFs.NewObject("data.bin")
|
2017-12-18 13:55:37 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.NotEqual(t, o.ModTime().String(), co.ModTime().String())
|
2017-11-12 18:54:25 +01:00
|
|
|
|
|
|
|
cfs.DirCacheFlush() // flush the cache
|
|
|
|
|
|
|
|
// get a new instance from the cache
|
2018-01-29 23:05:04 +01:00
|
|
|
co, err = rootFs.NewObject("data.bin")
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
require.Equal(t, wrappedTime.Unix(), co.ModTime().Unix())
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
func TestInternalCacheWrites(t *testing.T) {
|
2018-01-29 23:05:04 +01:00
|
|
|
id := "ticw"
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, false, true, nil, map[string]string{"cache-writes": "true"})
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
cfs, err := runInstance.getCacheFs(rootFs)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
chunkSize := cfs.ChunkSize()
|
|
|
|
|
|
|
|
// create some rand test data
|
2018-01-29 23:05:04 +01:00
|
|
|
earliestTime := time.Now()
|
|
|
|
testData := runInstance.randomBytes(t, (chunkSize*4 + chunkSize/2))
|
|
|
|
runInstance.writeRemoteBytes(t, rootFs, "data.bin", testData)
|
2017-12-09 22:54:26 +01:00
|
|
|
expectedTs := time.Now()
|
2018-01-29 23:05:04 +01:00
|
|
|
ts, err := boltDb.GetChunkTs(runInstance.encryptRemoteIfNeeded(t, path.Join(rootFs.Root(), "data.bin")), 0)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
require.WithinDuration(t, expectedTs, ts, expectedTs.Sub(earliestTime))
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
|
|
|
|
2017-12-09 22:54:26 +01:00
|
|
|
func TestInternalMaxChunkSizeRespected(t *testing.T) {
|
2018-01-29 23:05:04 +01:00
|
|
|
id := fmt.Sprintf("timcsr%v", time.Now().Unix())
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, false, true, nil, map[string]string{"cache-workers": "1"})
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
cfs, err := runInstance.getCacheFs(rootFs)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
chunkSize := cfs.ChunkSize()
|
|
|
|
totalChunks := 20
|
|
|
|
|
|
|
|
// create some rand test data
|
2018-01-29 23:05:04 +01:00
|
|
|
testData := runInstance.randomBytes(t, (int64(totalChunks-1)*chunkSize + chunkSize/2))
|
|
|
|
runInstance.writeRemoteBytes(t, rootFs, "data.bin", testData)
|
|
|
|
o, err := cfs.NewObject(runInstance.encryptRemoteIfNeeded(t, "data.bin"))
|
2017-12-18 13:55:37 +01:00
|
|
|
require.NoError(t, err)
|
2017-12-09 22:54:26 +01:00
|
|
|
co, ok := o.(*cache.Object)
|
|
|
|
require.True(t, ok)
|
|
|
|
|
|
|
|
for i := 0; i < 4; i++ { // read first 4
|
2018-01-29 23:05:04 +01:00
|
|
|
_ = runInstance.readDataFromObj(t, co, chunkSize*int64(i), chunkSize*int64(i+1), false)
|
2017-12-09 22:54:26 +01:00
|
|
|
}
|
|
|
|
cfs.CleanUpCache(true)
|
|
|
|
// the last 2 **must** be in the cache
|
|
|
|
require.True(t, boltDb.HasChunk(co, chunkSize*2))
|
|
|
|
require.True(t, boltDb.HasChunk(co, chunkSize*3))
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2017-12-09 22:54:26 +01:00
|
|
|
for i := 4; i < 6; i++ { // read next 2
|
2018-01-29 23:05:04 +01:00
|
|
|
_ = runInstance.readDataFromObj(t, co, chunkSize*int64(i), chunkSize*int64(i+1), false)
|
2017-12-09 22:54:26 +01:00
|
|
|
}
|
|
|
|
cfs.CleanUpCache(true)
|
|
|
|
// the last 2 **must** be in the cache
|
|
|
|
require.True(t, boltDb.HasChunk(co, chunkSize*4))
|
|
|
|
require.True(t, boltDb.HasChunk(co, chunkSize*5))
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
func TestInternalExpiredEntriesRemoved(t *testing.T) {
|
2018-01-29 23:05:04 +01:00
|
|
|
id := fmt.Sprintf("tieer%v", time.Now().Unix())
|
|
|
|
vfsflags.Opt.DirCacheTime = time.Second * 4 // needs to be lower than the defined
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, true, true, map[string]string{"info_age": "5s"}, nil)
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
|
|
|
cfs, err := runInstance.getCacheFs(rootFs)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// create some rand test data
|
|
|
|
runInstance.writeRemoteString(t, rootFs, "one", "one content")
|
|
|
|
runInstance.mkdir(t, rootFs, "test")
|
|
|
|
runInstance.writeRemoteString(t, rootFs, "test/second", "second content")
|
|
|
|
|
|
|
|
l := runInstance.list(t, rootFs, "test")
|
|
|
|
require.Len(t, l, 1)
|
|
|
|
|
|
|
|
err = cfs.UnWrap().Mkdir(runInstance.encryptRemoteIfNeeded(t, "test/third"))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
l = runInstance.list(t, rootFs, "test")
|
|
|
|
require.Len(t, l, 1)
|
|
|
|
|
|
|
|
err = runInstance.retryBlock(func() error {
|
|
|
|
l = runInstance.list(t, rootFs, "test")
|
|
|
|
if len(l) != 2 {
|
|
|
|
return errors.New("list is not 2")
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}, 10, time.Second)
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestInternalUploadTempDirCreated(t *testing.T) {
|
|
|
|
id := fmt.Sprintf("tiutdc%v", time.Now().Unix())
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, false, true,
|
|
|
|
nil,
|
|
|
|
map[string]string{"cache-tmp-upload-path": path.Join(runInstance.tmpUploadDir, id)})
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
|
|
|
|
|
|
|
_, err := os.Stat(path.Join(runInstance.tmpUploadDir, id))
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
func testInternalUploadQueueOneFile(t *testing.T, id string, rootFs fs.Fs, boltDb *cache.Persistent) {
|
|
|
|
// create some rand test data
|
|
|
|
testSize := int64(524288000)
|
|
|
|
testReader := runInstance.randomReader(t, testSize)
|
|
|
|
bu := runInstance.listenForBackgroundUpload(t, rootFs, "one")
|
|
|
|
runInstance.writeRemoteReader(t, rootFs, "one", testReader)
|
|
|
|
// validate that it exists in temp fs
|
|
|
|
ti, err := os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "one")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
if runInstance.rootIsCrypt {
|
|
|
|
require.Equal(t, int64(524416032), ti.Size())
|
|
|
|
} else {
|
|
|
|
require.Equal(t, testSize, ti.Size())
|
|
|
|
}
|
|
|
|
de1 := runInstance.list(t, rootFs, "")
|
|
|
|
require.Len(t, de1, 1)
|
|
|
|
|
|
|
|
runInstance.completeBackgroundUpload(t, "one", bu)
|
|
|
|
// check if it was removed from temp fs
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "one")))
|
|
|
|
require.True(t, os.IsNotExist(err))
|
|
|
|
|
|
|
|
// check if it can be read
|
|
|
|
data2 := runInstance.readDataFromRemote(t, rootFs, "one", 0, int64(1024), false)
|
|
|
|
require.Len(t, data2, 1024)
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func TestInternalUploadQueueOneFileNoRest(t *testing.T) {
|
|
|
|
id := fmt.Sprintf("tiuqofnr%v", time.Now().Unix())
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, true, true,
|
|
|
|
nil,
|
|
|
|
map[string]string{"cache-tmp-upload-path": path.Join(runInstance.tmpUploadDir, id), "cache-tmp-wait-time": "0s"})
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
|
|
|
|
|
|
|
testInternalUploadQueueOneFile(t, id, rootFs, boltDb)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestInternalUploadQueueOneFileWithRest(t *testing.T) {
|
|
|
|
id := fmt.Sprintf("tiuqofwr%v", time.Now().Unix())
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, true, true,
|
|
|
|
nil,
|
|
|
|
map[string]string{"cache-tmp-upload-path": path.Join(runInstance.tmpUploadDir, id), "cache-tmp-wait-time": "1m"})
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
|
|
|
|
|
|
|
testInternalUploadQueueOneFile(t, id, rootFs, boltDb)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestInternalUploadQueueMoreFiles(t *testing.T) {
|
|
|
|
id := fmt.Sprintf("tiuqmf%v", time.Now().Unix())
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, true, true,
|
|
|
|
nil,
|
|
|
|
map[string]string{"cache-tmp-upload-path": path.Join(runInstance.tmpUploadDir, id), "cache-tmp-wait-time": "1s"})
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
|
|
|
|
|
|
|
err := rootFs.Mkdir("test")
|
|
|
|
require.NoError(t, err)
|
|
|
|
minSize := 5242880
|
|
|
|
maxSize := 10485760
|
|
|
|
totalFiles := 10
|
|
|
|
rand.Seed(time.Now().Unix())
|
|
|
|
|
|
|
|
lastFile := ""
|
|
|
|
for i := 0; i < totalFiles; i++ {
|
|
|
|
size := int64(rand.Intn(maxSize-minSize) + minSize)
|
|
|
|
testReader := runInstance.randomReader(t, size)
|
|
|
|
remote := "test/" + strconv.Itoa(i) + ".bin"
|
|
|
|
runInstance.writeRemoteReader(t, rootFs, remote, testReader)
|
|
|
|
|
|
|
|
// validate that it exists in temp fs
|
|
|
|
ti, err := os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, remote)))
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, size, runInstance.cleanSize(t, ti.Size()))
|
|
|
|
|
|
|
|
if runInstance.wrappedIsExternal && i < totalFiles-1 {
|
|
|
|
time.Sleep(time.Second * 3)
|
|
|
|
}
|
|
|
|
lastFile = remote
|
|
|
|
}
|
|
|
|
|
|
|
|
// check if cache lists all files, likely temp upload didn't finish yet
|
|
|
|
de1 := runInstance.list(t, rootFs, "test")
|
|
|
|
require.Len(t, de1, totalFiles)
|
|
|
|
|
|
|
|
// wait for background uploader to do its thing
|
|
|
|
runInstance.completeAllBackgroundUploads(t, rootFs, lastFile)
|
|
|
|
|
|
|
|
// retry until we have no more temp files and fail if they don't go down to 0
|
|
|
|
tf, err := ioutil.ReadDir(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test")))
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
require.Len(t, tf, 0)
|
|
|
|
|
|
|
|
// check if cache lists all files
|
|
|
|
de1 = runInstance.list(t, rootFs, "test")
|
|
|
|
require.Len(t, de1, totalFiles)
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestInternalUploadTempFileOperations(t *testing.T) {
|
|
|
|
id := "tiutfo"
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, true, true,
|
|
|
|
nil,
|
|
|
|
map[string]string{"cache-tmp-upload-path": path.Join(runInstance.tmpUploadDir, id), "cache-tmp-wait-time": "1h"})
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
|
|
|
|
|
|
|
boltDb.PurgeTempUploads()
|
2017-11-12 18:54:25 +01:00
|
|
|
|
|
|
|
// create some rand test data
|
2018-01-29 23:05:04 +01:00
|
|
|
runInstance.mkdir(t, rootFs, "test")
|
|
|
|
runInstance.writeRemoteString(t, rootFs, "test/one", "one content")
|
|
|
|
|
|
|
|
// check if it can be read
|
|
|
|
data1 := runInstance.readDataFromRemote(t, rootFs, "test/one", 0, int64(len([]byte("one content"))), false)
|
|
|
|
require.Equal(t, []byte("one content"), data1)
|
|
|
|
// validate that it exists in temp fs
|
|
|
|
_, err := os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
// test DirMove - allowed
|
|
|
|
err = runInstance.dirMove(t, rootFs, "test", "second")
|
|
|
|
if err != errNotSupported {
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = rootFs.NewObject("test/one")
|
|
|
|
require.Error(t, err)
|
|
|
|
_, err = rootFs.NewObject("second/one")
|
|
|
|
require.NoError(t, err)
|
|
|
|
// validate that it exists in temp fs
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
|
|
|
require.Error(t, err)
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "second/one")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
started, err := boltDb.SearchPendingUpload(runInstance.encryptRemoteIfNeeded(t, path.Join(id, "test/one")))
|
|
|
|
require.Error(t, err)
|
|
|
|
started, err = boltDb.SearchPendingUpload(runInstance.encryptRemoteIfNeeded(t, path.Join(id, "second/one")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.False(t, started)
|
|
|
|
runInstance.mkdir(t, rootFs, "test")
|
|
|
|
runInstance.writeRemoteString(t, rootFs, "test/one", "one content")
|
|
|
|
}
|
|
|
|
|
|
|
|
// test Rmdir - allowed
|
|
|
|
err = runInstance.rm(t, rootFs, "test")
|
|
|
|
require.Error(t, err)
|
|
|
|
require.Contains(t, err.Error(), "directory not empty")
|
|
|
|
_, err = rootFs.NewObject("test/one")
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
// validate that it exists in temp fs
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
started, err := boltDb.SearchPendingUpload(runInstance.encryptRemoteIfNeeded(t, path.Join(id, "test/one")))
|
|
|
|
require.False(t, started)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// test Move/Rename -- allowed
|
|
|
|
err = runInstance.move(t, rootFs, path.Join("test", "one"), path.Join("test", "second"))
|
|
|
|
if err != errNotSupported {
|
|
|
|
require.NoError(t, err)
|
|
|
|
// try to read from it
|
|
|
|
_, err = rootFs.NewObject("test/one")
|
|
|
|
require.Error(t, err)
|
|
|
|
_, err = rootFs.NewObject("test/second")
|
|
|
|
require.NoError(t, err)
|
|
|
|
data2 := runInstance.readDataFromRemote(t, rootFs, "test/second", 0, int64(len([]byte("one content"))), false)
|
|
|
|
require.Equal(t, []byte("one content"), data2)
|
|
|
|
// validate that it exists in temp fs
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
|
|
|
require.Error(t, err)
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/second")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
runInstance.writeRemoteString(t, rootFs, "test/one", "one content")
|
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
// test Copy -- allowed
|
|
|
|
err = runInstance.copy(t, rootFs, path.Join("test", "one"), path.Join("test", "third"))
|
|
|
|
if err != errNotSupported {
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = rootFs.NewObject("test/one")
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = rootFs.NewObject("test/third")
|
|
|
|
require.NoError(t, err)
|
|
|
|
data2 := runInstance.readDataFromRemote(t, rootFs, "test/third", 0, int64(len([]byte("one content"))), false)
|
|
|
|
require.Equal(t, []byte("one content"), data2)
|
|
|
|
// validate that it exists in temp fs
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/third")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// test Remove -- allowed
|
|
|
|
err = runInstance.rm(t, rootFs, "test/one")
|
2017-12-18 13:55:37 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
_, err = rootFs.NewObject("test/one")
|
|
|
|
require.Error(t, err)
|
|
|
|
// validate that it doesn't exist in temp fs
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
|
|
|
require.Error(t, err)
|
|
|
|
runInstance.writeRemoteString(t, rootFs, "test/one", "one content")
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
// test Update -- allowed
|
|
|
|
firstModTime, err := runInstance.modTime(t, rootFs, "test/one")
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
err = runInstance.updateData(t, rootFs, "test/one", "one content", " updated")
|
|
|
|
require.NoError(t, err)
|
|
|
|
obj2, err := rootFs.NewObject("test/one")
|
|
|
|
require.NoError(t, err)
|
|
|
|
data2 := runInstance.readDataFromObj(t, obj2, 0, int64(len("one content updated")), false)
|
|
|
|
require.Equal(t, "one content updated", string(data2))
|
|
|
|
tmpInfo, err := os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
if runInstance.rootIsCrypt {
|
|
|
|
require.Equal(t, int64(67), tmpInfo.Size())
|
|
|
|
} else {
|
|
|
|
require.Equal(t, int64(len(data2)), tmpInfo.Size())
|
|
|
|
}
|
|
|
|
|
|
|
|
// test SetModTime -- allowed
|
|
|
|
secondModTime, err := runInstance.modTime(t, rootFs, "test/one")
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.NotEqual(t, secondModTime, firstModTime)
|
|
|
|
require.NotEqual(t, time.Time{}, firstModTime)
|
|
|
|
require.NotEqual(t, time.Time{}, secondModTime)
|
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func TestInternalUploadUploadingFileOperations(t *testing.T) {
|
|
|
|
id := "tiuufo"
|
|
|
|
rootFs, boltDb := runInstance.newCacheFs(t, remoteName, id, true, true,
|
|
|
|
nil,
|
|
|
|
map[string]string{"cache-tmp-upload-path": path.Join(runInstance.tmpUploadDir, id), "cache-tmp-wait-time": "1h"})
|
|
|
|
defer runInstance.cleanupFs(t, rootFs, boltDb)
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
boltDb.PurgeTempUploads()
|
|
|
|
|
|
|
|
// create some rand test data
|
|
|
|
runInstance.mkdir(t, rootFs, "test")
|
|
|
|
runInstance.writeRemoteString(t, rootFs, "test/one", "one content")
|
|
|
|
|
|
|
|
// check if it can be read
|
|
|
|
data1 := runInstance.readDataFromRemote(t, rootFs, "test/one", 0, int64(len([]byte("one content"))), false)
|
|
|
|
require.Equal(t, []byte("one content"), data1)
|
|
|
|
// validate that it exists in temp fs
|
|
|
|
_, err := os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
err = boltDb.SetPendingUploadToStarted(runInstance.encryptRemoteIfNeeded(t, path.Join(rootFs.Root(), "test/one")))
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
|
|
|
|
// test DirMove
|
|
|
|
err = runInstance.dirMove(t, rootFs, "test", "second")
|
|
|
|
if err != errNotSupported {
|
|
|
|
require.Error(t, err)
|
|
|
|
_, err = rootFs.NewObject("test/one")
|
|
|
|
require.NoError(t, err)
|
|
|
|
// validate that it exists in temp fs
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "second/one")))
|
|
|
|
require.Error(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// test Rmdir
|
|
|
|
err = runInstance.rm(t, rootFs, "test")
|
|
|
|
require.Error(t, err)
|
|
|
|
_, err = rootFs.NewObject("test/one")
|
|
|
|
require.NoError(t, err)
|
|
|
|
// validate that it doesn't exist in temp fs
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// test Move/Rename
|
|
|
|
err = runInstance.move(t, rootFs, path.Join("test", "one"), path.Join("test", "second"))
|
|
|
|
if err != errNotSupported {
|
|
|
|
require.Error(t, err)
|
|
|
|
// try to read from it
|
|
|
|
_, err = rootFs.NewObject("test/one")
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = rootFs.NewObject("test/second")
|
|
|
|
require.Error(t, err)
|
|
|
|
// validate that it exists in temp fs
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/second")))
|
|
|
|
require.Error(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// test Copy -- allowed
|
|
|
|
err = runInstance.copy(t, rootFs, path.Join("test", "one"), path.Join("test", "third"))
|
|
|
|
if err != errNotSupported {
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = rootFs.NewObject("test/one")
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = rootFs.NewObject("test/third")
|
|
|
|
require.NoError(t, err)
|
|
|
|
data2 := runInstance.readDataFromRemote(t, rootFs, "test/third", 0, int64(len([]byte("one content"))), false)
|
|
|
|
require.Equal(t, []byte("one content"), data2)
|
|
|
|
// validate that it exists in temp fs
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/third")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// test Remove
|
|
|
|
err = runInstance.rm(t, rootFs, "test/one")
|
|
|
|
require.Error(t, err)
|
|
|
|
_, err = rootFs.NewObject("test/one")
|
|
|
|
require.NoError(t, err)
|
|
|
|
// validate that it doesn't exist in temp fs
|
|
|
|
_, err = os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
|
|
|
require.NoError(t, err)
|
|
|
|
runInstance.writeRemoteString(t, rootFs, "test/one", "one content")
|
|
|
|
|
|
|
|
// test Update - this seems to work. Why? FIXME
|
|
|
|
//firstModTime, err := runInstance.modTime(t, rootFs, "test/one")
|
|
|
|
//require.NoError(t, err)
|
|
|
|
//err = runInstance.updateData(t, rootFs, "test/one", "one content", " updated", func() {
|
|
|
|
// data2 := runInstance.readDataFromRemote(t, rootFs, "test/one", 0, int64(len("one content updated")), true)
|
|
|
|
// require.Equal(t, "one content", string(data2))
|
|
|
|
//
|
|
|
|
// tmpInfo, err := os.Stat(path.Join(runInstance.tmpUploadDir, id, runInstance.encryptRemoteIfNeeded(t, "test/one")))
|
|
|
|
// require.NoError(t, err)
|
|
|
|
// if runInstance.rootIsCrypt {
|
|
|
|
// require.Equal(t, int64(67), tmpInfo.Size())
|
|
|
|
// } else {
|
|
|
|
// require.Equal(t, int64(len(data2)), tmpInfo.Size())
|
|
|
|
// }
|
|
|
|
//})
|
|
|
|
//require.Error(t, err)
|
|
|
|
|
|
|
|
// test SetModTime -- seems to work cause of previous
|
|
|
|
//secondModTime, err := runInstance.modTime(t, rootFs, "test/one")
|
|
|
|
//require.NoError(t, err)
|
|
|
|
//require.Equal(t, secondModTime, firstModTime)
|
|
|
|
//require.NotEqual(t, time.Time{}, firstModTime)
|
|
|
|
//require.NotEqual(t, time.Time{}, secondModTime)
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
|
|
|
|
2017-12-18 13:55:37 +01:00
|
|
|
// FIXME, enable this when mount is sorted out
|
|
|
|
//func TestInternalFilesMissingInMount1904(t *testing.T) {
|
|
|
|
// t.Skip("Not yet")
|
|
|
|
// if runtime.GOOS == "windows" {
|
|
|
|
// t.Skip("Not yet")
|
|
|
|
// }
|
|
|
|
// id := "tifm1904"
|
2018-01-29 23:05:04 +01:00
|
|
|
// rootFs, _ := newCacheFs(t, RemoteName, id, false,
|
2017-12-18 13:55:37 +01:00
|
|
|
// map[string]string{"chunk_size": "5M", "info_age": "1m", "chunk_total_size": "500M", "cache-writes": "true"})
|
|
|
|
// mntPoint := path.Join("/tmp", "tifm1904-mnt")
|
|
|
|
// testPoint := path.Join(mntPoint, id)
|
|
|
|
// checkOutput := "1 10 100 11 12 13 14 15 16 17 18 19 2 20 21 22 23 24 25 26 27 28 29 3 30 31 32 33 34 35 36 37 38 39 4 40 41 42 43 44 45 46 47 48 49 5 50 51 52 53 54 55 56 57 58 59 6 60 61 62 63 64 65 66 67 68 69 7 70 71 72 73 74 75 76 77 78 79 8 80 81 82 83 84 85 86 87 88 89 9 90 91 92 93 94 95 96 97 98 99 "
|
|
|
|
//
|
|
|
|
// _ = os.MkdirAll(mntPoint, os.ModePerm)
|
|
|
|
//
|
|
|
|
// list, err := rootFs.List("")
|
|
|
|
// require.NoError(t, err)
|
|
|
|
// found := false
|
|
|
|
// list.ForDir(func(d fs.Directory) {
|
|
|
|
// if strings.Contains(d.Remote(), id) {
|
|
|
|
// found = true
|
|
|
|
// }
|
|
|
|
// })
|
|
|
|
//
|
|
|
|
// if !found {
|
|
|
|
// t.Skip("Test folder '%v' doesn't exist", id)
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
// mountFs(t, rootFs, mntPoint)
|
|
|
|
// defer unmountFs(t, mntPoint)
|
|
|
|
//
|
|
|
|
// for i := 1; i <= 2; i++ {
|
|
|
|
// out, err := exec.Command("ls", testPoint).Output()
|
|
|
|
// require.NoError(t, err)
|
|
|
|
// require.Equal(t, checkOutput, strings.Replace(string(out), "\n", " ", -1))
|
|
|
|
// t.Logf("root path has all files")
|
|
|
|
// _ = writeObjectString(t, rootFs, path.Join(id, strconv.Itoa(i), strconv.Itoa(i), "one_file"), "one content")
|
|
|
|
//
|
|
|
|
// for j := 1; j <= 100; j++ {
|
|
|
|
// out, err := exec.Command("ls", path.Join(testPoint, strconv.Itoa(j))).Output()
|
|
|
|
// require.NoError(t, err)
|
|
|
|
// require.Equal(t, checkOutput, strings.Replace(string(out), "\n", " ", -1), "'%v' doesn't match", j)
|
|
|
|
// }
|
|
|
|
// obj, err := rootFs.NewObject(path.Join(id, strconv.Itoa(i), strconv.Itoa(i), "one_file"))
|
|
|
|
// require.NoError(t, err)
|
|
|
|
// err = obj.Remove()
|
|
|
|
// require.NoError(t, err)
|
|
|
|
// t.Logf("folders contain all the files")
|
|
|
|
//
|
|
|
|
// out, err = exec.Command("date").Output()
|
|
|
|
// require.NoError(t, err)
|
|
|
|
// t.Logf("check #%v date: '%v'", i, strings.Replace(string(out), "\n", " ", -1))
|
|
|
|
//
|
|
|
|
// if i < 2 {
|
|
|
|
// time.Sleep(time.Second * 60)
|
|
|
|
// }
|
|
|
|
// }
|
|
|
|
//}
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
// run holds the remotes for a test run
|
|
|
|
type run struct {
|
|
|
|
okDiff time.Duration
|
|
|
|
allCfgMap map[string]string
|
|
|
|
allFlagMap map[string]string
|
|
|
|
runDefaultCfgMap map[string]string
|
|
|
|
runDefaultFlagMap map[string]string
|
|
|
|
mntDir string
|
|
|
|
tmpUploadDir string
|
|
|
|
useMount bool
|
|
|
|
isMounted bool
|
|
|
|
rootIsCrypt bool
|
|
|
|
wrappedIsExternal bool
|
|
|
|
unmountFn func() error
|
|
|
|
unmountRes chan error
|
|
|
|
vfs *vfs.VFS
|
|
|
|
tempFiles []*os.File
|
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func newRun() *run {
|
|
|
|
var err error
|
|
|
|
r := &run{
|
|
|
|
okDiff: time.Second * 9, // really big diff here but the build machines seem to be slow. need a different way for this
|
|
|
|
useMount: useMount,
|
|
|
|
isMounted: false,
|
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
r.allCfgMap = map[string]string{
|
|
|
|
"plex_url": "",
|
|
|
|
"plex_username": "",
|
|
|
|
"plex_password": "",
|
|
|
|
"chunk_size": cache.DefCacheChunkSize,
|
|
|
|
"info_age": cache.DefCacheInfoAge,
|
|
|
|
"chunk_total_size": cache.DefCacheTotalChunkSize,
|
|
|
|
}
|
|
|
|
r.allFlagMap = map[string]string{
|
|
|
|
"cache-db-path": filepath.Join(config.CacheDir, "cache-backend"),
|
|
|
|
"cache-chunk-path": filepath.Join(config.CacheDir, "cache-backend"),
|
|
|
|
"cache-db-purge": "true",
|
|
|
|
"cache-chunk-size": cache.DefCacheChunkSize,
|
|
|
|
"cache-total-chunk-size": cache.DefCacheTotalChunkSize,
|
|
|
|
"cache-chunk-clean-interval": cache.DefCacheChunkCleanInterval,
|
|
|
|
"cache-info-age": cache.DefCacheInfoAge,
|
|
|
|
"cache-read-retries": strconv.Itoa(cache.DefCacheReadRetries),
|
|
|
|
"cache-workers": strconv.Itoa(cache.DefCacheTotalWorkers),
|
|
|
|
"cache-chunk-no-memory": "false",
|
|
|
|
"cache-rps": strconv.Itoa(cache.DefCacheRps),
|
|
|
|
"cache-writes": "false",
|
|
|
|
"cache-tmp-upload-path": "",
|
|
|
|
"cache-tmp-wait-time": cache.DefCacheTmpWaitTime,
|
|
|
|
}
|
|
|
|
r.runDefaultCfgMap = make(map[string]string)
|
|
|
|
for key, value := range r.allCfgMap {
|
|
|
|
r.runDefaultCfgMap[key] = value
|
|
|
|
}
|
|
|
|
r.runDefaultFlagMap = make(map[string]string)
|
|
|
|
for key, value := range r.allFlagMap {
|
|
|
|
r.runDefaultFlagMap[key] = value
|
|
|
|
}
|
|
|
|
if mountDir == "" {
|
|
|
|
if runtime.GOOS != "windows" {
|
|
|
|
r.mntDir, err = ioutil.TempDir("", "rclonecache-mount")
|
|
|
|
if err != nil {
|
|
|
|
log.Fatalf("Failed to create mount dir: %v", err)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// Find a free drive letter
|
|
|
|
drive := ""
|
|
|
|
for letter := 'E'; letter <= 'Z'; letter++ {
|
|
|
|
drive = string(letter) + ":"
|
|
|
|
_, err := os.Stat(drive + "\\")
|
|
|
|
if os.IsNotExist(err) {
|
|
|
|
goto found
|
|
|
|
}
|
|
|
|
}
|
|
|
|
log.Print("Couldn't find free drive letter for test")
|
|
|
|
found:
|
|
|
|
r.mntDir = drive
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
r.mntDir = mountDir
|
|
|
|
}
|
|
|
|
log.Printf("Mount Dir: %v", r.mntDir)
|
|
|
|
|
|
|
|
if uploadDir == "" {
|
|
|
|
r.tmpUploadDir, err = ioutil.TempDir("", "rclonecache-tmp")
|
|
|
|
if err != nil {
|
|
|
|
log.Fatalf("Failed to create temp dir: %v", err)
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
r.tmpUploadDir = uploadDir
|
|
|
|
}
|
|
|
|
log.Printf("Temp Upload Dir: %v", r.tmpUploadDir)
|
|
|
|
|
|
|
|
return r
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) encryptRemoteIfNeeded(t *testing.T, remote string) string {
|
|
|
|
if !runInstance.rootIsCrypt || len(decryptedToEncryptedRemotes) == 0 {
|
|
|
|
return remote
|
|
|
|
}
|
|
|
|
|
|
|
|
enc, ok := decryptedToEncryptedRemotes[remote]
|
|
|
|
if !ok {
|
|
|
|
t.Fatalf("Failed to find decrypted -> encrypted mapping for '%v'", remote)
|
|
|
|
return remote
|
|
|
|
}
|
|
|
|
return enc
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) newCacheFs(t *testing.T, remote, id string, needRemote, purge bool, cfg map[string]string, flags map[string]string) (fs.Fs, *cache.Persistent) {
|
|
|
|
fstest.Initialise()
|
|
|
|
remoteExists := false
|
|
|
|
for _, s := range config.FileSections() {
|
|
|
|
if s == remote {
|
|
|
|
remoteExists = true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if !remoteExists && needRemote {
|
|
|
|
t.Skipf("Need remote (%v) to exist", remote)
|
|
|
|
return nil, nil
|
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
// if the remote doesn't exist, create a new one with a local one for it
|
|
|
|
// identify which is the cache remote (it can be wrapped by a crypt too)
|
|
|
|
rootIsCrypt := false
|
|
|
|
cacheRemote := remote
|
|
|
|
if !remoteExists {
|
|
|
|
localRemote := remote + "-local"
|
|
|
|
config.FileSet(localRemote, "type", "local")
|
|
|
|
config.FileSet(localRemote, "nounc", "true")
|
|
|
|
config.FileSet(remote, "type", "cache")
|
|
|
|
config.FileSet(remote, "remote", localRemote+":/var/tmp/"+localRemote)
|
|
|
|
} else {
|
|
|
|
remoteType := fs.ConfigFileGet(remote, "type", "")
|
|
|
|
if remoteType == "" {
|
|
|
|
t.Skipf("skipped due to invalid remote type for %v", remote)
|
|
|
|
return nil, nil
|
|
|
|
}
|
|
|
|
if remoteType != "cache" {
|
|
|
|
if remoteType == "crypt" {
|
|
|
|
rootIsCrypt = true
|
|
|
|
config.FileSet(remote, "password", cryptPassword1)
|
|
|
|
config.FileSet(remote, "password2", cryptPassword2)
|
|
|
|
}
|
|
|
|
remoteRemote := fs.ConfigFileGet(remote, "remote", "")
|
|
|
|
if remoteRemote == "" {
|
|
|
|
t.Skipf("skipped due to invalid remote wrapper for %v", remote)
|
|
|
|
return nil, nil
|
|
|
|
}
|
|
|
|
remoteRemoteParts := strings.Split(remoteRemote, ":")
|
|
|
|
remoteWrapping := remoteRemoteParts[0]
|
|
|
|
remoteType := fs.ConfigFileGet(remoteWrapping, "type", "")
|
|
|
|
if remoteType != "cache" {
|
|
|
|
t.Skipf("skipped due to invalid remote type for %v: '%v'", remoteWrapping, remoteType)
|
|
|
|
return nil, nil
|
|
|
|
}
|
|
|
|
cacheRemote = remoteWrapping
|
|
|
|
}
|
|
|
|
}
|
|
|
|
runInstance.rootIsCrypt = rootIsCrypt
|
|
|
|
dbPath := filepath.Join(config.CacheDir, "cache-backend", cacheRemote+".db")
|
|
|
|
chunkPath := filepath.Join(config.CacheDir, "cache-backend", cacheRemote)
|
|
|
|
boltDb, err := cache.GetPersistent(dbPath, chunkPath, &cache.Features{PurgeDb: true})
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
for k, v := range r.runDefaultCfgMap {
|
|
|
|
if c, ok := cfg[k]; ok {
|
|
|
|
config.FileSet(cacheRemote, k, c)
|
|
|
|
} else {
|
|
|
|
config.FileSet(cacheRemote, k, v)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for k, v := range r.runDefaultFlagMap {
|
|
|
|
if c, ok := flags[k]; ok {
|
|
|
|
_ = flag.Set(k, c)
|
|
|
|
} else {
|
|
|
|
_ = flag.Set(k, v)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
fs.Config.LowLevelRetries = 1
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
// Instantiate root
|
|
|
|
if purge {
|
|
|
|
boltDb.PurgeTempUploads()
|
|
|
|
_ = os.RemoveAll(path.Join(runInstance.tmpUploadDir, id))
|
|
|
|
}
|
|
|
|
f, err := fs.NewFs(remote + ":" + id)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
cfs, err := r.getCacheFs(f)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
_, isCache := cfs.Features().UnWrap().(*cache.Fs)
|
|
|
|
_, isCrypt := cfs.Features().UnWrap().(*crypt.Fs)
|
|
|
|
_, isLocal := cfs.Features().UnWrap().(*local.Fs)
|
|
|
|
if isCache || isCrypt || isLocal {
|
|
|
|
r.wrappedIsExternal = true
|
|
|
|
} else {
|
|
|
|
r.wrappedIsExternal = true
|
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
if purge {
|
|
|
|
_ = f.Features().Purge()
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
err = f.Mkdir("")
|
|
|
|
require.NoError(t, err)
|
|
|
|
if r.useMount && !r.isMounted {
|
|
|
|
r.mountFs(t, f)
|
|
|
|
}
|
|
|
|
|
|
|
|
return f, boltDb
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) cleanupFs(t *testing.T, f fs.Fs, b *cache.Persistent) {
|
|
|
|
if r.useMount && r.isMounted {
|
|
|
|
r.unmountFs(t, f)
|
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
err := f.Features().Purge()
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
cfs, err := r.getCacheFs(f)
|
|
|
|
require.NoError(t, err)
|
|
|
|
cfs.StopBackgroundRunners()
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
if r.useMount && runtime.GOOS != "windows" {
|
|
|
|
err = os.RemoveAll(r.mntDir)
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
err = os.RemoveAll(r.tmpUploadDir)
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
for _, f := range r.tempFiles {
|
|
|
|
_ = f.Close()
|
|
|
|
_ = os.Remove(f.Name())
|
|
|
|
}
|
|
|
|
r.tempFiles = nil
|
|
|
|
debug.FreeOSMemory()
|
|
|
|
for k, v := range r.runDefaultFlagMap {
|
|
|
|
_ = flag.Set(k, v)
|
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) randomBytes(t *testing.T, size int64) []byte {
|
|
|
|
testData := make([]byte, size)
|
|
|
|
testSize, err := rand.Read(testData)
|
|
|
|
require.Equal(t, size, int64(len(testData)))
|
|
|
|
require.Equal(t, size, int64(testSize))
|
2017-12-18 13:55:37 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
return testData
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) randomReader(t *testing.T, size int64) io.ReadCloser {
|
|
|
|
chunk := int64(1024)
|
|
|
|
cnt := size / chunk
|
|
|
|
left := size % chunk
|
|
|
|
f, err := ioutil.TempFile("", "rclonecache-tempfile")
|
2017-12-18 13:55:37 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
for i := 0; i < int(cnt); i++ {
|
|
|
|
data := r.randomBytes(t, chunk)
|
|
|
|
_, _ = f.Write(data)
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
data := r.randomBytes(t, int64(left))
|
|
|
|
_, _ = f.Write(data)
|
|
|
|
_, _ = f.Seek(int64(0), 0)
|
|
|
|
r.tempFiles = append(r.tempFiles, f)
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
return f
|
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) writeRemoteRandomBytes(t *testing.T, f fs.Fs, p string, size int64) string {
|
|
|
|
remote := path.Join(p, strconv.Itoa(rand.Int())+".bin")
|
|
|
|
// create some rand test data
|
|
|
|
testData := r.randomBytes(t, size)
|
|
|
|
|
|
|
|
r.writeRemoteBytes(t, f, remote, testData)
|
|
|
|
return remote
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) writeObjectRandomBytes(t *testing.T, f fs.Fs, p string, size int64) fs.Object {
|
|
|
|
remote := path.Join(p, strconv.Itoa(rand.Int())+".bin")
|
|
|
|
// create some rand test data
|
|
|
|
testData := r.randomBytes(t, size)
|
|
|
|
|
|
|
|
return r.writeObjectBytes(t, f, remote, testData)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) writeRemoteString(t *testing.T, f fs.Fs, remote, content string) {
|
|
|
|
r.writeRemoteBytes(t, f, remote, []byte(content))
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) writeObjectString(t *testing.T, f fs.Fs, remote, content string) fs.Object {
|
|
|
|
return r.writeObjectBytes(t, f, remote, []byte(content))
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) writeRemoteBytes(t *testing.T, f fs.Fs, remote string, data []byte) {
|
|
|
|
var err error
|
|
|
|
|
|
|
|
if r.useMount {
|
|
|
|
err = r.retryBlock(func() error {
|
|
|
|
return ioutil.WriteFile(path.Join(r.mntDir, remote), data, 0600)
|
|
|
|
}, 3, time.Second*3)
|
|
|
|
require.NoError(t, err)
|
|
|
|
r.vfs.WaitForWriters(10 * time.Second)
|
2017-12-18 13:55:37 +01:00
|
|
|
} else {
|
2018-01-29 23:05:04 +01:00
|
|
|
r.writeObjectBytes(t, f, remote, data)
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) writeRemoteReader(t *testing.T, f fs.Fs, remote string, in io.ReadCloser) {
|
|
|
|
defer func() {
|
|
|
|
_ = in.Close()
|
|
|
|
}()
|
|
|
|
|
|
|
|
if r.useMount {
|
|
|
|
out, err := os.Create(path.Join(r.mntDir, remote))
|
|
|
|
require.NoError(t, err)
|
|
|
|
defer func() {
|
|
|
|
_ = out.Close()
|
|
|
|
}()
|
|
|
|
|
|
|
|
_, err = io.Copy(out, in)
|
|
|
|
require.NoError(t, err)
|
|
|
|
r.vfs.WaitForWriters(10 * time.Second)
|
2017-12-18 13:55:37 +01:00
|
|
|
} else {
|
2018-01-29 23:05:04 +01:00
|
|
|
r.writeObjectReader(t, f, remote, in)
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) writeObjectBytes(t *testing.T, f fs.Fs, remote string, data []byte) fs.Object {
|
|
|
|
in := bytes.NewReader(data)
|
|
|
|
_ = r.writeObjectReader(t, f, remote, in)
|
|
|
|
o, err := f.NewObject(remote)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, int64(len(data)), o.Size())
|
|
|
|
return o
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) writeObjectReader(t *testing.T, f fs.Fs, remote string, in io.Reader) fs.Object {
|
|
|
|
modTime := time.Now()
|
|
|
|
objInfo := object.NewStaticObjectInfo(remote, modTime, -1, true, nil, f)
|
|
|
|
obj, err := f.Put(in, objInfo)
|
|
|
|
require.NoError(t, err)
|
|
|
|
if r.useMount {
|
|
|
|
r.vfs.WaitForWriters(10 * time.Second)
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
|
|
|
|
return obj
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) updateObjectRemote(t *testing.T, f fs.Fs, remote string, data1 []byte, data2 []byte) fs.Object {
|
|
|
|
var err error
|
|
|
|
var obj fs.Object
|
|
|
|
|
|
|
|
if r.useMount {
|
|
|
|
err = ioutil.WriteFile(path.Join(r.mntDir, remote), data1, 0600)
|
|
|
|
require.NoError(t, err)
|
|
|
|
r.vfs.WaitForWriters(10 * time.Second)
|
|
|
|
err = ioutil.WriteFile(path.Join(r.mntDir, remote), data2, 0600)
|
|
|
|
require.NoError(t, err)
|
|
|
|
r.vfs.WaitForWriters(10 * time.Second)
|
|
|
|
obj, err = f.NewObject(remote)
|
2017-12-18 13:55:37 +01:00
|
|
|
} else {
|
2018-01-29 23:05:04 +01:00
|
|
|
in1 := bytes.NewReader(data1)
|
|
|
|
in2 := bytes.NewReader(data2)
|
|
|
|
objInfo1 := object.NewStaticObjectInfo(remote, time.Now(), int64(len(data1)), true, nil, f)
|
|
|
|
objInfo2 := object.NewStaticObjectInfo(remote, time.Now(), int64(len(data2)), true, nil, f)
|
|
|
|
|
|
|
|
obj, err = f.Put(in1, objInfo1)
|
|
|
|
require.NoError(t, err)
|
|
|
|
obj, err = f.NewObject(remote)
|
|
|
|
require.NoError(t, err)
|
|
|
|
err = obj.Update(in2, objInfo2)
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
return obj
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) readDataFromRemote(t *testing.T, f fs.Fs, remote string, offset, end int64, noLengthCheck bool) []byte {
|
|
|
|
size := end - offset
|
|
|
|
checkSample := make([]byte, size)
|
|
|
|
|
|
|
|
if r.useMount {
|
|
|
|
f, err := os.Open(path.Join(r.mntDir, remote))
|
|
|
|
defer func() {
|
|
|
|
_ = f.Close()
|
|
|
|
}()
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, _ = f.Seek(offset, 0)
|
|
|
|
totalRead, err := io.ReadFull(f, checkSample)
|
|
|
|
checkSample = checkSample[:totalRead]
|
|
|
|
if err == io.EOF || err == io.ErrUnexpectedEOF {
|
|
|
|
err = nil
|
|
|
|
}
|
|
|
|
require.NoError(t, err)
|
|
|
|
if !noLengthCheck {
|
|
|
|
require.Equal(t, size, int64(totalRead))
|
|
|
|
}
|
|
|
|
require.NoError(t, err)
|
2017-12-18 13:55:37 +01:00
|
|
|
} else {
|
2018-01-29 23:05:04 +01:00
|
|
|
co, err := f.NewObject(remote)
|
|
|
|
require.NoError(t, err)
|
|
|
|
checkSample = r.readDataFromObj(t, co, offset, end, noLengthCheck)
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
if !noLengthCheck {
|
|
|
|
require.Equal(t, size, int64(len(checkSample)), "wrong data read size from file")
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
return checkSample
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) readDataFromObj(t *testing.T, o fs.Object, offset, end int64, noLengthCheck bool) []byte {
|
|
|
|
size := end - offset
|
|
|
|
checkSample := make([]byte, size)
|
|
|
|
reader, err := o.Open(&fs.SeekOption{Offset: offset})
|
2017-12-18 13:55:37 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
totalRead, err := io.ReadFull(reader, checkSample)
|
|
|
|
if (err == io.EOF || err == io.ErrUnexpectedEOF) && noLengthCheck {
|
|
|
|
err = nil
|
|
|
|
checkSample = checkSample[:totalRead]
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
_ = reader.Close()
|
|
|
|
return checkSample
|
|
|
|
}
|
2017-11-12 18:54:25 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) mkdir(t *testing.T, f fs.Fs, remote string) {
|
|
|
|
var err error
|
|
|
|
if r.useMount {
|
|
|
|
err = os.Mkdir(path.Join(r.mntDir, remote), 0700)
|
|
|
|
} else {
|
|
|
|
err = f.Mkdir(remote)
|
|
|
|
}
|
|
|
|
require.NoError(t, err)
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) rm(t *testing.T, f fs.Fs, remote string) error {
|
|
|
|
var err error
|
|
|
|
|
|
|
|
if r.useMount {
|
|
|
|
err = os.Remove(path.Join(r.mntDir, remote))
|
|
|
|
} else {
|
|
|
|
var obj fs.Object
|
|
|
|
obj, err = f.NewObject(remote)
|
|
|
|
if err != nil {
|
|
|
|
err = f.Rmdir(remote)
|
|
|
|
} else {
|
|
|
|
err = obj.Remove()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) list(t *testing.T, f fs.Fs, remote string) []interface{} {
|
|
|
|
var err error
|
|
|
|
var l []interface{}
|
|
|
|
if r.useMount {
|
|
|
|
var list []os.FileInfo
|
|
|
|
list, err = ioutil.ReadDir(path.Join(r.mntDir, remote))
|
|
|
|
for _, ll := range list {
|
|
|
|
l = append(l, ll)
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
var list fs.DirEntries
|
|
|
|
list, err = f.List(remote)
|
|
|
|
for _, ll := range list {
|
|
|
|
l = append(l, ll)
|
|
|
|
}
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
return l
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) listPath(t *testing.T, f fs.Fs, remote string) []string {
|
|
|
|
var err error
|
|
|
|
var l []string
|
|
|
|
if r.useMount {
|
|
|
|
var list []os.FileInfo
|
|
|
|
list, err = ioutil.ReadDir(path.Join(r.mntDir, remote))
|
|
|
|
for _, ll := range list {
|
|
|
|
l = append(l, ll.Name())
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
} else {
|
|
|
|
var list fs.DirEntries
|
|
|
|
list, err = f.List(remote)
|
|
|
|
for _, ll := range list {
|
|
|
|
l = append(l, ll.Remote())
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
require.NoError(t, err)
|
|
|
|
return l
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) copyFile(t *testing.T, f fs.Fs, src, dst string) error {
|
|
|
|
in, err := os.Open(src)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
defer func() {
|
|
|
|
_ = in.Close()
|
|
|
|
}()
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
out, err := os.Create(dst)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
defer func() {
|
|
|
|
_ = out.Close()
|
|
|
|
}()
|
|
|
|
|
|
|
|
_, err = io.Copy(out, in)
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) dirMove(t *testing.T, rootFs fs.Fs, src, dst string) error {
|
|
|
|
var err error
|
|
|
|
|
|
|
|
if runInstance.useMount {
|
|
|
|
err = os.Rename(path.Join(runInstance.mntDir, src), path.Join(runInstance.mntDir, dst))
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
r.vfs.WaitForWriters(10 * time.Second)
|
|
|
|
} else if rootFs.Features().DirMove != nil {
|
|
|
|
err = rootFs.Features().DirMove(rootFs, src, dst)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
} else {
|
2018-01-29 23:05:04 +01:00
|
|
|
t.Logf("DirMove not supported by %v", rootFs)
|
|
|
|
return errNotSupported
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) move(t *testing.T, rootFs fs.Fs, src, dst string) error {
|
|
|
|
var err error
|
|
|
|
|
|
|
|
if runInstance.useMount {
|
|
|
|
err = os.Rename(path.Join(runInstance.mntDir, src), path.Join(runInstance.mntDir, dst))
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
r.vfs.WaitForWriters(10 * time.Second)
|
|
|
|
} else if rootFs.Features().Move != nil {
|
|
|
|
obj1, err := rootFs.NewObject(src)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
_, err = rootFs.Features().Move(obj1, dst)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
} else {
|
2018-01-29 23:05:04 +01:00
|
|
|
t.Logf("Move not supported by %v", rootFs)
|
|
|
|
return errNotSupported
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) copy(t *testing.T, rootFs fs.Fs, src, dst string) error {
|
|
|
|
var err error
|
|
|
|
|
|
|
|
if r.useMount {
|
|
|
|
err = r.copyFile(t, rootFs, path.Join(r.mntDir, src), path.Join(r.mntDir, dst))
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
r.vfs.WaitForWriters(10 * time.Second)
|
|
|
|
} else if rootFs.Features().Copy != nil {
|
|
|
|
obj, err := rootFs.NewObject(src)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
_, err = rootFs.Features().Copy(obj, dst)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
} else {
|
2018-01-29 23:05:04 +01:00
|
|
|
t.Logf("Copy not supported by %v", rootFs)
|
|
|
|
return errNotSupported
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) modTime(t *testing.T, rootFs fs.Fs, src string) (time.Time, error) {
|
|
|
|
var err error
|
|
|
|
|
|
|
|
if r.useMount {
|
|
|
|
fi, err := os.Stat(path.Join(runInstance.mntDir, src))
|
|
|
|
if err != nil {
|
|
|
|
return time.Time{}, err
|
|
|
|
}
|
|
|
|
return fi.ModTime(), nil
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
obj1, err := rootFs.NewObject(src)
|
|
|
|
if err != nil {
|
|
|
|
return time.Time{}, err
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
return obj1.ModTime(), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) updateData(t *testing.T, rootFs fs.Fs, src, data, append string) error {
|
|
|
|
var err error
|
|
|
|
|
|
|
|
if r.useMount {
|
|
|
|
f, err := os.OpenFile(path.Join(runInstance.mntDir, src), os.O_TRUNC|os.O_CREATE|os.O_WRONLY, 0644)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
_, err = f.WriteString(data + append)
|
|
|
|
if err != nil {
|
|
|
|
_ = f.Close()
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
err = f.Close()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
r.vfs.WaitForWriters(10 * time.Second)
|
2017-12-18 13:55:37 +01:00
|
|
|
} else {
|
2018-01-29 23:05:04 +01:00
|
|
|
obj1, err := rootFs.NewObject(src)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
data1 := []byte(data + append)
|
|
|
|
r := bytes.NewReader(data1)
|
|
|
|
objInfo1 := object.NewStaticObjectInfo(src, time.Now(), int64(len(data1)), true, nil, rootFs)
|
|
|
|
err = obj1.Update(r, objInfo1)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) cleanSize(t *testing.T, size int64) int64 {
|
|
|
|
if r.rootIsCrypt {
|
|
|
|
denominator := int64(65536 + 16)
|
|
|
|
size = size - 32
|
|
|
|
quotient := size / denominator
|
|
|
|
remainder := size % denominator
|
|
|
|
return (quotient*65536 + remainder - 16)
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
return size
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *run) listenForBackgroundUpload(t *testing.T, f fs.Fs, remote string) chan error {
|
|
|
|
cfs, err := r.getCacheFs(f)
|
2017-12-18 13:55:37 +01:00
|
|
|
require.NoError(t, err)
|
2018-01-29 23:05:04 +01:00
|
|
|
buCh := cfs.GetBackgroundUploadChannel()
|
|
|
|
require.NotNil(t, buCh)
|
|
|
|
maxDuration := time.Minute * 3
|
|
|
|
if r.wrappedIsExternal {
|
|
|
|
maxDuration = time.Minute * 10
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
waitCh := make(chan error)
|
|
|
|
go func() {
|
|
|
|
var err error
|
|
|
|
var state cache.BackgroundUploadState
|
|
|
|
|
|
|
|
for i := 0; i < 2; i++ {
|
|
|
|
select {
|
|
|
|
case state = <-buCh:
|
|
|
|
// continue
|
|
|
|
case <-time.After(maxDuration):
|
|
|
|
waitCh <- errors.Errorf("Timed out waiting for background upload: %v", remote)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
checkRemote := state.Remote
|
|
|
|
if r.rootIsCrypt {
|
|
|
|
cryptFs := f.(*crypt.Fs)
|
|
|
|
checkRemote, err = cryptFs.DecryptFileName(checkRemote)
|
|
|
|
if err != nil {
|
|
|
|
waitCh <- err
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if checkRemote == remote && cache.BackgroundUploadStarted != state.Status {
|
|
|
|
waitCh <- state.Error
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
waitCh <- errors.Errorf("Too many attempts to wait for the background upload: %v", remote)
|
|
|
|
}()
|
|
|
|
return waitCh
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) completeBackgroundUpload(t *testing.T, remote string, waitCh chan error) {
|
|
|
|
var err error
|
|
|
|
maxDuration := time.Minute * 3
|
|
|
|
if r.wrappedIsExternal {
|
|
|
|
maxDuration = time.Minute * 10
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case err = <-waitCh:
|
|
|
|
// continue
|
|
|
|
case <-time.After(maxDuration):
|
|
|
|
t.Fatalf("Timed out waiting to complete the background upload %v", remote)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) completeAllBackgroundUploads(t *testing.T, f fs.Fs, lastRemote string) {
|
|
|
|
var state cache.BackgroundUploadState
|
2017-12-18 13:55:37 +01:00
|
|
|
var err error
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
maxDuration := time.Minute * 5
|
|
|
|
if r.wrappedIsExternal {
|
|
|
|
maxDuration = time.Minute * 15
|
|
|
|
}
|
|
|
|
cfs, err := r.getCacheFs(f)
|
|
|
|
require.NoError(t, err)
|
|
|
|
buCh := cfs.GetBackgroundUploadChannel()
|
|
|
|
require.NotNil(t, buCh)
|
|
|
|
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case state = <-buCh:
|
|
|
|
checkRemote := state.Remote
|
|
|
|
if r.rootIsCrypt {
|
|
|
|
cryptFs := f.(*crypt.Fs)
|
|
|
|
checkRemote, err = cryptFs.DecryptFileName(checkRemote)
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
if checkRemote == lastRemote && cache.BackgroundUploadCompleted == state.Status {
|
|
|
|
require.NoError(t, state.Error)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
case <-time.After(maxDuration):
|
|
|
|
t.Fatalf("Timed out waiting to complete the background upload %v", lastRemote)
|
|
|
|
return
|
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
}
|
2018-01-29 23:05:04 +01:00
|
|
|
}
|
2017-12-18 13:55:37 +01:00
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) retryBlock(block func() error, maxRetries int, rate time.Duration) error {
|
|
|
|
var err error
|
|
|
|
for i := 0; i < maxRetries; i++ {
|
|
|
|
err = block()
|
|
|
|
if err == nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
time.Sleep(rate)
|
|
|
|
}
|
|
|
|
return err
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
func (r *run) getCacheFs(f fs.Fs) (*cache.Fs, error) {
|
2017-11-12 18:54:25 +01:00
|
|
|
cfs, ok := f.(*cache.Fs)
|
|
|
|
if ok {
|
|
|
|
return cfs, nil
|
|
|
|
} else {
|
|
|
|
if f.Features().UnWrap != nil {
|
|
|
|
cfs, ok := f.Features().UnWrap().(*cache.Fs)
|
|
|
|
if ok {
|
|
|
|
return cfs, nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-01-29 23:05:04 +01:00
|
|
|
return nil, errors.New("didn't found a cache fs")
|
2017-11-12 18:54:25 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
var (
|
|
|
|
_ fs.Fs = (*cache.Fs)(nil)
|
|
|
|
_ fs.Fs = (*local.Fs)(nil)
|
|
|
|
)
|