mirror of
https://github.com/rclone/rclone.git
synced 2025-01-29 21:04:23 +02:00
b4216648e4
Before this change, bisync could only detect changes based on modtime, and would refuse to run if either path lacked modtime support. This made bisync unavailable for many of rclone's backends. Additionally, bisync did not account for the Fs's precision when comparing modtimes, meaning that they could only be reliably compared within the same side -- not against the opposite side. Size and checksum (even when available) were ignored completely for deltas. After this change, bisync now fully supports comparing based on any combination of size, modtime, and checksum, lifting the prior restriction on backends without modtime support. The comparison logic considers the backend's precision, hash types, and other features as appropriate. The comparison features optionally use a new --compare flag (which takes any combination of size,modtime,checksum) and even supports some combinations not otherwise supported in `sync` (like comparing all three at the same time.) By default (without the --compare flag), bisync inherits the same comparison options as `sync` (that is: size and modtime by default, unless modified with flags such as --checksum or --size-only.) If the --compare flag is set, it will override these defaults. If --compare includes checksum and both remotes support checksums but have no hash types in common with each other, checksums will be considered only for comparisons within the same side (to determine what has changed since the prior sync), but not for comparisons against the opposite side. If one side supports checksums and the other does not, checksums will only be considered on the side that supports them. When comparing with checksum and/or size without modtime, bisync cannot determine whether a file is newer or older -- only whether it is changed or unchanged. (If it is changed on both sides, bisync still does the standard equality-check to avoid declaring a sync conflict unless it absolutely has to.) Also included are some new flags to customize the checksum comparison behavior on backends where hashes are slow or unavailable. --no-slow-hash and --slow-hash-sync-only allow selectively ignoring checksums on backends such as local where they are slow. --download-hash allows computing them by downloading when (and only when) they're otherwise not available. Of course, this option probably won't be practical with large files, but may be a good option for syncing small-but-important files with maximum accuracy (for example, a source code repo on a crypt remote.) An additional advantage over methods like cryptcheck is that the original file is not required for comparison (for example, --download-hash can be used to bisync two different crypt remotes with different passwords.) Additionally, all of the above are now considered during the final --check-sync for much-improved accuracy (before this change, it only compared filenames!) Many other details are explained in the included docs.
238 lines
5.8 KiB
Go
238 lines
5.8 KiB
Go
package bisync
|
|
|
|
import (
|
|
"context"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/rclone/rclone/fs"
|
|
"github.com/rclone/rclone/fs/accounting"
|
|
"github.com/rclone/rclone/fs/filter"
|
|
"github.com/rclone/rclone/fs/hash"
|
|
"github.com/rclone/rclone/fs/march"
|
|
)
|
|
|
|
var ls1 = newFileList()
|
|
var ls2 = newFileList()
|
|
var err error
|
|
var firstErr error
|
|
var marchAliasLock sync.Mutex
|
|
var marchLsLock sync.Mutex
|
|
var marchErrLock sync.Mutex
|
|
var marchCtx context.Context
|
|
|
|
func (b *bisyncRun) makeMarchListing(ctx context.Context) (*fileList, *fileList, error) {
|
|
ci := fs.GetConfig(ctx)
|
|
marchCtx = ctx
|
|
b.setupListing()
|
|
fs.Debugf(b, "starting to march!")
|
|
|
|
// set up a march over fdst (Path2) and fsrc (Path1)
|
|
m := &march.March{
|
|
Ctx: ctx,
|
|
Fdst: b.fs2,
|
|
Fsrc: b.fs1,
|
|
Dir: "",
|
|
NoTraverse: false,
|
|
Callback: b,
|
|
DstIncludeAll: false,
|
|
NoCheckDest: false,
|
|
NoUnicodeNormalization: ci.NoUnicodeNormalization,
|
|
}
|
|
err = m.Run(ctx)
|
|
|
|
fs.Debugf(b, "march completed. err: %v", err)
|
|
if err == nil {
|
|
err = firstErr
|
|
}
|
|
if err != nil {
|
|
b.handleErr("march", "error during march", err, true, true)
|
|
b.abort = true
|
|
}
|
|
|
|
// save files
|
|
if b.opt.Compare.DownloadHash && ls1.hash == hash.None {
|
|
ls1.hash = hash.MD5
|
|
}
|
|
if b.opt.Compare.DownloadHash && ls2.hash == hash.None {
|
|
ls2.hash = hash.MD5
|
|
}
|
|
err = ls1.save(ctx, b.newListing1)
|
|
b.handleErr(ls1, "error saving ls1 from march", err, true, true)
|
|
err = ls2.save(ctx, b.newListing2)
|
|
b.handleErr(ls2, "error saving ls2 from march", err, true, true)
|
|
|
|
return ls1, ls2, err
|
|
}
|
|
|
|
// SrcOnly have an object which is on path1 only
|
|
func (b *bisyncRun) SrcOnly(o fs.DirEntry) (recurse bool) {
|
|
fs.Debugf(o, "path1 only")
|
|
b.parse(o, true)
|
|
return isDir(o)
|
|
}
|
|
|
|
// DstOnly have an object which is on path2 only
|
|
func (b *bisyncRun) DstOnly(o fs.DirEntry) (recurse bool) {
|
|
fs.Debugf(o, "path2 only")
|
|
b.parse(o, false)
|
|
return isDir(o)
|
|
}
|
|
|
|
// Match is called when object exists on both path1 and path2 (whether equal or not)
|
|
func (b *bisyncRun) Match(ctx context.Context, o2, o1 fs.DirEntry) (recurse bool) {
|
|
fs.Debugf(o1, "both path1 and path2")
|
|
marchAliasLock.Lock()
|
|
b.aliases.Add(o1.Remote(), o2.Remote())
|
|
marchAliasLock.Unlock()
|
|
b.parse(o1, true)
|
|
b.parse(o2, false)
|
|
return isDir(o1)
|
|
}
|
|
|
|
func isDir(e fs.DirEntry) bool {
|
|
switch x := e.(type) {
|
|
case fs.Object:
|
|
fs.Debugf(x, "is Object")
|
|
return false
|
|
case fs.Directory:
|
|
fs.Debugf(x, "is Dir")
|
|
return true
|
|
default:
|
|
fs.Debugf(e, "is unknown")
|
|
}
|
|
return false
|
|
}
|
|
|
|
func (b *bisyncRun) parse(e fs.DirEntry, isPath1 bool) {
|
|
switch x := e.(type) {
|
|
case fs.Object:
|
|
b.ForObject(x, isPath1)
|
|
case fs.Directory:
|
|
if b.opt.CreateEmptySrcDirs {
|
|
b.ForDir(x, isPath1)
|
|
}
|
|
default:
|
|
fs.Debugf(e, "is unknown")
|
|
}
|
|
}
|
|
|
|
func (b *bisyncRun) setupListing() {
|
|
ls1 = newFileList()
|
|
ls2 = newFileList()
|
|
|
|
// note that --ignore-listing-checksum is different from --ignore-checksum
|
|
// and we already checked it when we set b.opt.Compare.HashType1 and 2
|
|
ls1.hash = b.opt.Compare.HashType1
|
|
ls2.hash = b.opt.Compare.HashType2
|
|
}
|
|
|
|
func (b *bisyncRun) ForObject(o fs.Object, isPath1 bool) {
|
|
tr := accounting.Stats(marchCtx).NewCheckingTransfer(o, "listing file - "+whichPath(isPath1))
|
|
defer func() {
|
|
tr.Done(marchCtx, nil)
|
|
}()
|
|
var (
|
|
hashVal string
|
|
hashErr error
|
|
)
|
|
ls := whichLs(isPath1)
|
|
hashType := ls.hash
|
|
if hashType != hash.None {
|
|
hashVal, hashErr = o.Hash(marchCtx, hashType)
|
|
marchErrLock.Lock()
|
|
if firstErr == nil {
|
|
firstErr = hashErr
|
|
}
|
|
marchErrLock.Unlock()
|
|
}
|
|
hashVal, hashErr = tryDownloadHash(marchCtx, o, hashVal)
|
|
marchErrLock.Lock()
|
|
if firstErr == nil {
|
|
firstErr = hashErr
|
|
}
|
|
if firstErr != nil {
|
|
b.handleErr(hashType, "error hashing during march", firstErr, false, true)
|
|
}
|
|
marchErrLock.Unlock()
|
|
|
|
var modtime time.Time
|
|
if b.opt.Compare.Modtime {
|
|
modtime = o.ModTime(marchCtx).In(TZ)
|
|
}
|
|
id := "" // TODO
|
|
flags := "-" // "-" for a file and "d" for a directory
|
|
marchLsLock.Lock()
|
|
ls.put(o.Remote(), o.Size(), modtime, hashVal, id, flags)
|
|
marchLsLock.Unlock()
|
|
}
|
|
|
|
func (b *bisyncRun) ForDir(o fs.Directory, isPath1 bool) {
|
|
tr := accounting.Stats(marchCtx).NewCheckingTransfer(o, "listing dir - "+whichPath(isPath1))
|
|
defer func() {
|
|
tr.Done(marchCtx, nil)
|
|
}()
|
|
ls := whichLs(isPath1)
|
|
var modtime time.Time
|
|
if b.opt.Compare.Modtime {
|
|
modtime = o.ModTime(marchCtx).In(TZ)
|
|
}
|
|
id := "" // TODO
|
|
flags := "d" // "-" for a file and "d" for a directory
|
|
marchLsLock.Lock()
|
|
ls.put(o.Remote(), -1, modtime, "", id, flags)
|
|
marchLsLock.Unlock()
|
|
}
|
|
|
|
func whichLs(isPath1 bool) *fileList {
|
|
ls := ls1
|
|
if !isPath1 {
|
|
ls = ls2
|
|
}
|
|
return ls
|
|
}
|
|
|
|
func whichPath(isPath1 bool) string {
|
|
s := "Path1"
|
|
if !isPath1 {
|
|
s = "Path2"
|
|
}
|
|
return s
|
|
}
|
|
|
|
func (b *bisyncRun) findCheckFiles(ctx context.Context) (*fileList, *fileList, error) {
|
|
ctxCheckFile, filterCheckFile := filter.AddConfig(ctx)
|
|
b.handleErr(b.opt.CheckFilename, "error adding CheckFilename to filter", filterCheckFile.Add(true, b.opt.CheckFilename), true, true)
|
|
b.handleErr(b.opt.CheckFilename, "error adding ** exclusion to filter", filterCheckFile.Add(false, "**"), true, true)
|
|
ci := fs.GetConfig(ctxCheckFile)
|
|
marchCtx = ctxCheckFile
|
|
|
|
b.setupListing()
|
|
fs.Debugf(b, "starting to march!")
|
|
|
|
// set up a march over fdst (Path2) and fsrc (Path1)
|
|
m := &march.March{
|
|
Ctx: ctxCheckFile,
|
|
Fdst: b.fs2,
|
|
Fsrc: b.fs1,
|
|
Dir: "",
|
|
NoTraverse: false,
|
|
Callback: b,
|
|
DstIncludeAll: false,
|
|
NoCheckDest: false,
|
|
NoUnicodeNormalization: ci.NoUnicodeNormalization,
|
|
}
|
|
err = m.Run(ctxCheckFile)
|
|
|
|
fs.Debugf(b, "march completed. err: %v", err)
|
|
if err == nil {
|
|
err = firstErr
|
|
}
|
|
if err != nil {
|
|
b.handleErr("march", "error during findCheckFiles", err, true, true)
|
|
b.abort = true
|
|
}
|
|
|
|
return ls1, ls2, err
|
|
}
|