mirror of
https://github.com/seaweedfs/seaweedfs.git
synced 2026-05-14 05:41:29 +00:00
* refactor(command): expand "~" in all path-style CLI flags Many of weed's path-bearing flags (-s3.config, -s3.iam.config, -admin.dataDir, -webdav.cacheDir, -volume.dir.idx, TLS cert/key files, profile output paths, mount cache dirs, sftp key files, ...) were never run through util.ResolvePath, so a value like "~/iam.json" was used literally. Tilde only worked when the shell expanded it, which silently fails for the common -flag=~/path form (bash leaves the tilde literal in --opt=~/path). - Extend util.ResolvePath to also handle "~user" / "~user/rest", matching shell tilde expansion. Add unit tests. - Apply util.ResolvePath at the top of each shared start* function (s3, webdav, sftp) so mini/server/filer/standalone callers all inherit it; resolve at the few one-off use sites (mount cache dirs, volume idx folder, mini admin.dataDir, profile paths). - Drop the duplicate expandHomeDir helper from admin.go in favor of the now-equivalent util.ResolvePath. * fixup: handle comma-separated -dir flags for tilde expansion `weed mini -dir`, `weed server -dir`, and `weed volume -dir` accept comma-separated paths (`dir[,dir]...`). Calling util.ResolvePath on the whole string mishandled multi-folder values with tilde, e.g. "~/d1,~/d2" would resolve as if "d1,~/d2" were a single subpath. - Add util.ResolveCommaSeparatedPaths: split on ",", run each entry through ResolvePath, rejoin. Short-circuits when no "~" present. - Use it for *miniDataFolders (mini.go), *volumeDataFolders (server.go), and resolve each entry of v.folders in-place (volume.go) so all downstream consumers see resolved paths. - Add 7-case TestResolveCommaSeparatedPaths covering empty, single, multiple, and mixed inputs. * address PR review: metaFolder + Windows backslash - master.go: resolve *m.metaFolder at the top of runMaster so util.FullPath(*m.metaFolder) on the next line sees an expanded path. Drop the now-redundant ResolvePath in TestFolderWritable. - server.go: same treatment for *masterOptions.metaFolder, paired with the existing cpu/mem profile resolves. Drop the redundant inner ResolvePath at TestFolderWritable. - file_util.go: ResolvePath now accepts filepath.Separator as a separator after the tilde, so "~\\data" works on Windows. Other platforms keep current behaviour (backslash stays literal because it is a valid filename character in usernames and paths). - file_util_test.go: add two cases using filepath.Separator that exercise the new code path on Windows and remain a no-op on Unix. * address PR review: resolve "~" in remaining command path flags Comprehensive sweep of path-bearing flags across every weed subcommand, applying util.ResolvePath in-place at the top of each run* function so all downstream consumers see expanded paths. - webdav.go: resolve *wo.cacheDir at the top of startWebDav so mini/server/filer/standalone callers all inherit it. - mount_std.go: cpu/mem profile paths. - filer_sync.go: cpu/mem profile paths. - mq_broker.go: cpu/mem profile paths. - benchmark.go: cpuprofile output path. - backup.go: -dir resolved once at runBackup; drop the duplicated inline ResolvePath in NewVolume calls. - compact.go: -dir resolved at runCompact; drop inline ResolvePath. - export.go: -dir and -o resolved at runExport; drop inline ResolvePath in LoadFromIdx and ScanVolumeFile. - download.go: -dir resolved at runDownload; drop inline. - update.go: -dir resolved at runUpdate so filepath.Join uses the expanded path; drop inline ResolvePath in TestFolderWritable. - scaffold.go: -output expanded before filepath.Join. - worker.go: -workingDir expanded before being passed to runtime. * address PR review: resolve option-struct paths at run* entry points server.go:381 propagates s3Options.config to filerOptions.s3ConfigFile *before* startS3Server runs, which meant the filer-side code saw the unresolved tilde-prefixed pointer. Same pattern for webdavOptions and sftpOptions (and equivalent in mini.go / filer.go). The fix: hoist resolution from the shared start* functions up to the run* entry points, where every shared pointer is set up before any propagation happens. - s3.go, webdav.go, sftp.go: extract a resolvePaths() method on each Options struct that runs every path field through util.ResolvePath in-place. Idempotent. - runS3, runWebDav, runSftp: call the standalone struct's resolvePaths before starting metrics / loading security config. - runServer, runMini, runFiler: call resolvePaths on every embedded options struct, plus resolve loose flags (serverIamConfig, miniS3Config, miniIamConfig, miniMasterOptions.metaFolder, and filer's defaultLevelDbDirectory) so they're expanded before any pointer copy or use. - Drop the now-redundant inline ResolvePath at filer's defaultLevelDbDirectory composition. * address PR review: re-resolve mini -dir post-config, cover misc paths - mini.go: applyConfigFileOptions can overwrite -dir with a literal ~/data from mini.options. Re-resolve *miniDataFolders after the config-file apply, alongside the other path resolves, so the mini filer no longer ends up with a literal ~/data/filerldb2. - benchmark.go: resolve *b.idListFile (-list). - filer_sync.go: resolve *syncOptions.aSecurity / .bSecurity (-a.security / -b.security) before LoadClientTLSFromFile. - filer_cat.go: resolve *filerCat.output (-o) before os.OpenFile. - admin.go: drop trailing blank line at EOF (git diff --check). * address PR review: resolve -a.security/-b.security/-config before use Three follow-up fixes: - filer_sync.go: the -a.security / -b.security resolves were placed *after* LoadClientTLSFromFile / LoadHTTPClientFromFile were called, so weed filer.sync -a.security=~/a.toml still passed the literal tilde path. Hoist the resolves above the security-loading block so TLS clients see expanded paths. - filer_sync_verify.go: same flag pair was never resolved at all in the verify command; resolve at the top of runFilerSyncVerify. - filer_meta_backup.go: -config (the backup_filer.toml path) was passed directly to viper. Resolve at the top of runFilerMetaBackup. - mini.go: master.dir defaulted to the entire comma-joined miniDataFolders. With weed mini -dir=~/d1,~/d2 (or any multi-dir setup), TestFolderWritable then stat'd the joined string instead of a single directory. Default to the first entry via StringSplit to mirror the disk-space calculation a few lines below, and drop the now-redundant ResolvePath in TestFolderWritable.
271 lines
7.7 KiB
Go
271 lines
7.7 KiB
Go
package command
|
|
|
|
import (
|
|
"archive/tar"
|
|
"bytes"
|
|
"fmt"
|
|
"io"
|
|
"os"
|
|
"path"
|
|
"path/filepath"
|
|
"strconv"
|
|
"strings"
|
|
"text/template"
|
|
"time"
|
|
|
|
"github.com/seaweedfs/seaweedfs/weed/glog"
|
|
"github.com/seaweedfs/seaweedfs/weed/storage"
|
|
"github.com/seaweedfs/seaweedfs/weed/storage/needle"
|
|
"github.com/seaweedfs/seaweedfs/weed/storage/needle_map"
|
|
"github.com/seaweedfs/seaweedfs/weed/storage/super_block"
|
|
"github.com/seaweedfs/seaweedfs/weed/storage/types"
|
|
"github.com/seaweedfs/seaweedfs/weed/util"
|
|
)
|
|
|
|
const (
|
|
defaultFnFormat = `{{.Id}}_{{.Name}}{{.Ext}}`
|
|
timeFormat = "2006-01-02T15:04:05"
|
|
)
|
|
|
|
var (
|
|
export ExportOptions
|
|
)
|
|
|
|
type ExportOptions struct {
|
|
dir *string
|
|
collection *string
|
|
volumeId *int
|
|
}
|
|
|
|
var cmdExport = &Command{
|
|
UsageLine: "export -dir=/tmp -volumeId=234 -o=/dir/name.tar -fileNameFormat={{.Name}} -newer='" + timeFormat + "'",
|
|
Short: "list or export files from one volume data file",
|
|
Long: `List all files in a volume, or Export all files in a volume to a tar file if the output is specified.
|
|
|
|
The format of file name in the tar file can be customized. Default is {{.Mime}}/{{.Id}}:{{.Name}}. Also available is {{.Key}}.
|
|
|
|
`,
|
|
}
|
|
|
|
func init() {
|
|
cmdExport.Run = runExport // break init cycle
|
|
export.dir = cmdExport.Flag.String("dir", ".", "input data directory to store volume data files")
|
|
export.collection = cmdExport.Flag.String("collection", "", "the volume collection name")
|
|
export.volumeId = cmdExport.Flag.Int("volumeId", -1, "a volume id. The volume .dat and .idx files should already exist in the dir.")
|
|
}
|
|
|
|
var (
|
|
output = cmdExport.Flag.String("o", "", "output tar file name, must ends with .tar, or just a \"-\" for stdout")
|
|
format = cmdExport.Flag.String("fileNameFormat", defaultFnFormat, "filename formatted with {{.Id}} {{.Name}} {{.Ext}}")
|
|
newer = cmdExport.Flag.String("newer", "", "export only files newer than this time, default is all files. Must be specified in RFC3339 without timezone, e.g. 2006-01-02T15:04:05")
|
|
showDeleted = cmdExport.Flag.Bool("deleted", false, "export deleted files. only applies if -o is not specified")
|
|
limit = cmdExport.Flag.Int("limit", 0, "only show first n entries if specified")
|
|
|
|
tarOutputFile *tar.Writer
|
|
tarHeader tar.Header
|
|
fileNameTemplate *template.Template
|
|
fileNameTemplateBuffer = bytes.NewBuffer(nil)
|
|
newerThan time.Time
|
|
newerThanUnix int64 = -1
|
|
localLocation, _ = time.LoadLocation("Local")
|
|
)
|
|
|
|
func printNeedle(vid needle.VolumeId, n *needle.Needle, version needle.Version, deleted bool, offset int64, onDiskSize int64) {
|
|
key := needle.NewFileIdFromNeedle(vid, n).String()
|
|
size := int32(n.DataSize)
|
|
if version == needle.Version1 {
|
|
size = int32(n.Size)
|
|
}
|
|
fmt.Printf("%s\t%s\t%d\t%t\t%s\t%s\t%s\t%t\t%d\t%d\n",
|
|
key,
|
|
n.Name,
|
|
size,
|
|
n.IsCompressed(),
|
|
n.Mime,
|
|
n.LastModifiedString(),
|
|
n.Ttl.String(),
|
|
deleted,
|
|
offset,
|
|
offset+onDiskSize,
|
|
)
|
|
}
|
|
|
|
type VolumeFileScanner4Export struct {
|
|
version needle.Version
|
|
counter int
|
|
needleMap *needle_map.MemDb
|
|
vid needle.VolumeId
|
|
}
|
|
|
|
func (scanner *VolumeFileScanner4Export) VisitSuperBlock(superBlock super_block.SuperBlock) error {
|
|
scanner.version = superBlock.Version
|
|
return nil
|
|
|
|
}
|
|
func (scanner *VolumeFileScanner4Export) ReadNeedleBody() bool {
|
|
return true
|
|
}
|
|
|
|
func (scanner *VolumeFileScanner4Export) VisitNeedle(n *needle.Needle, offset int64, needleHeader, needleBody []byte) error {
|
|
needleMap := scanner.needleMap
|
|
vid := scanner.vid
|
|
|
|
nv, ok := needleMap.Get(n.Id)
|
|
glog.V(3).Infof("key %d offset %d size %d disk_size %d compressed %v ok %v nv %+v",
|
|
n.Id, offset, n.Size, n.DiskSize(scanner.version), n.IsCompressed(), ok, nv)
|
|
if *showDeleted && n.Size > 0 || ok && nv.Size.IsValid() && nv.Offset.ToActualOffset() == offset {
|
|
if newerThanUnix >= 0 && n.HasLastModifiedDate() && n.LastModified < uint64(newerThanUnix) {
|
|
glog.V(3).Infof("Skipping this file, as it's old enough: LastModified %d vs %d",
|
|
n.LastModified, newerThanUnix)
|
|
return nil
|
|
}
|
|
scanner.counter++
|
|
if *limit > 0 && scanner.counter > *limit {
|
|
return io.EOF
|
|
}
|
|
if tarOutputFile != nil {
|
|
return writeFile(vid, n)
|
|
} else {
|
|
printNeedle(vid, n, scanner.version, false, offset, n.DiskSize(scanner.version))
|
|
return nil
|
|
}
|
|
}
|
|
if !ok {
|
|
if *showDeleted && tarOutputFile == nil {
|
|
if n.DataSize > 0 {
|
|
printNeedle(vid, n, scanner.version, true, offset, n.DiskSize(scanner.version))
|
|
} else {
|
|
n.Name = []byte("*tombstone")
|
|
printNeedle(vid, n, scanner.version, true, offset, n.DiskSize(scanner.version))
|
|
}
|
|
}
|
|
glog.V(2).Infof("This seems deleted %d size %d", n.Id, n.Size)
|
|
} else {
|
|
glog.V(2).Infof("Skipping later-updated Id %d size %d", n.Id, n.Size)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func runExport(cmd *Command, args []string) bool {
|
|
|
|
*export.dir = util.ResolvePath(*export.dir)
|
|
*output = util.ResolvePath(*output)
|
|
|
|
var err error
|
|
|
|
if *newer != "" {
|
|
if newerThan, err = time.ParseInLocation(timeFormat, *newer, localLocation); err != nil {
|
|
fmt.Println("cannot parse 'newer' argument: " + err.Error())
|
|
return false
|
|
}
|
|
newerThanUnix = newerThan.Unix()
|
|
}
|
|
|
|
if *export.volumeId == -1 {
|
|
return false
|
|
}
|
|
|
|
if *output != "" {
|
|
if *output != "-" && !strings.HasSuffix(*output, ".tar") {
|
|
fmt.Println("the output file", *output, "should be '-' or end with .tar")
|
|
return false
|
|
}
|
|
|
|
if fileNameTemplate, err = template.New("name").Parse(*format); err != nil {
|
|
fmt.Println("cannot parse format " + *format + ": " + err.Error())
|
|
return false
|
|
}
|
|
|
|
var outputFile *os.File
|
|
if *output == "-" {
|
|
outputFile = os.Stdout
|
|
} else {
|
|
if outputFile, err = os.Create(*output); err != nil {
|
|
glog.Fatalf("cannot open output tar %s: %s", *output, err)
|
|
}
|
|
}
|
|
defer outputFile.Close()
|
|
tarOutputFile = tar.NewWriter(outputFile)
|
|
defer tarOutputFile.Close()
|
|
t := time.Now()
|
|
tarHeader = tar.Header{Mode: 0644,
|
|
ModTime: t, Uid: os.Getuid(), Gid: os.Getgid(),
|
|
Typeflag: tar.TypeReg,
|
|
AccessTime: t, ChangeTime: t}
|
|
}
|
|
|
|
fileName := strconv.Itoa(*export.volumeId)
|
|
if *export.collection != "" {
|
|
fileName = *export.collection + "_" + fileName
|
|
}
|
|
vid := needle.VolumeId(*export.volumeId)
|
|
|
|
needleMap := needle_map.NewMemDb()
|
|
defer needleMap.Close()
|
|
|
|
if err := needleMap.LoadFromIdx(path.Join(*export.dir, fileName+".idx")); err != nil {
|
|
glog.Fatalf("cannot load needle map from %s.idx: %s", fileName, err)
|
|
}
|
|
|
|
volumeFileScanner := &VolumeFileScanner4Export{
|
|
needleMap: needleMap,
|
|
vid: vid,
|
|
}
|
|
|
|
if tarOutputFile == nil {
|
|
fmt.Printf("key\tname\tsize\tgzip\tmime\tmodified\tttl\tdeleted\tstart\tstop\n")
|
|
}
|
|
|
|
err = storage.ScanVolumeFile(*export.dir, *export.collection, vid, storage.NeedleMapInMemory, volumeFileScanner)
|
|
if err != nil && err != io.EOF {
|
|
glog.Errorf("Export Volume File [ERROR] %s\n", err)
|
|
}
|
|
return true
|
|
}
|
|
|
|
type nameParams struct {
|
|
Name string
|
|
Id types.NeedleId
|
|
Mime string
|
|
Key string
|
|
Ext string
|
|
}
|
|
|
|
func writeFile(vid needle.VolumeId, n *needle.Needle) (err error) {
|
|
key := needle.NewFileIdFromNeedle(vid, n).String()
|
|
fileNameTemplateBuffer.Reset()
|
|
if err = fileNameTemplate.Execute(fileNameTemplateBuffer,
|
|
nameParams{
|
|
Name: string(n.Name),
|
|
Id: n.Id,
|
|
Mime: string(n.Mime),
|
|
Key: key,
|
|
Ext: filepath.Ext(string(n.Name)),
|
|
},
|
|
); err != nil {
|
|
return err
|
|
}
|
|
|
|
fileName := fileNameTemplateBuffer.String()
|
|
|
|
if n.IsCompressed() {
|
|
if util.IsGzippedContent(n.Data) && path.Ext(fileName) != ".gz" {
|
|
fileName = fileName + ".gz"
|
|
}
|
|
// TODO other compression method
|
|
}
|
|
|
|
tarHeader.Name, tarHeader.Size = fileName, int64(len(n.Data))
|
|
if n.HasLastModifiedDate() {
|
|
tarHeader.ModTime = time.Unix(int64(n.LastModified), 0)
|
|
} else {
|
|
tarHeader.ModTime = time.Unix(0, 0)
|
|
}
|
|
tarHeader.ChangeTime = tarHeader.ModTime
|
|
if err = tarOutputFile.WriteHeader(&tarHeader); err != nil {
|
|
return err
|
|
}
|
|
_, err = tarOutputFile.Write(n.Data)
|
|
return
|
|
}
|