// _ _ // __ _____ __ ___ ___ __ _| |_ ___ // \ \ /\ / / _ \/ _` \ \ / / |/ _` | __/ _ \ // \ V V / __/ (_| |\ V /| | (_| | || __/ // \_/\_/ \___|\__,_| \_/ |_|\__,_|\__\___| // // Copyright © 2016 - 2025 Weaviate B.V. All rights reserved. // // CONTACT: hello@weaviate.io // package modsloads3 import ( "context" "fmt" "os" "path/filepath" "strconv" "strings" "sync/atomic" "time" "github.com/prometheus/client_golang/prometheus" "github.com/sirupsen/logrus" "github.com/urfave/cli/v2" "github.com/weaviate/s5cmd/v2/command" "github.com/weaviate/s5cmd/v2/log" "github.com/weaviate/s5cmd/v2/log/stat" "github.com/weaviate/s5cmd/v2/parallel" entcfg "github.com/weaviate/weaviate/entities/config" "github.com/weaviate/weaviate/entities/modulecapabilities" "github.com/weaviate/weaviate/entities/moduletools" "github.com/weaviate/weaviate/usecases/config" "github.com/weaviate/weaviate/usecases/monitoring" ) const ( Name = "offload-s3" s3Endpoint = "OFFLOAD_S3_ENDPOINT" s3BucketAutoCreate = "OFFLOAD_S3_BUCKET_AUTO_CREATE" s3Bucket = "OFFLOAD_S3_BUCKET" concurrency = "OFFLOAD_S3_CONCURRENCY" timeout = "OFFLOAD_TIMEOUT" ) // verify we implement the modules.Module interface var ( _ = modulecapabilities.Module(&Module{}) ) type Module struct { Endpoint string Bucket string BucketExists atomic.Bool Concurrency int DataPath string logger logrus.FieldLogger timeout time.Duration app *cli.App metrics *monitoring.TenantOffloadMetrics } func New() *Module { return &Module{ Endpoint: "", Bucket: "weaviate-offload", Concurrency: 25, DataPath: config.DefaultPersistenceDataPath, timeout: 120 * time.Second, // we use custom cli app to avoid some bugs in underlying dependencies // specially with .After implementation. metrics: monitoring.NewTenantOffloadMetrics(monitoring.Config{ MetricsNamespace: "weaviate", }, prometheus.DefaultRegisterer), app: &cli.App{ Name: "weaviate-s5cmd", Usage: "weaviate fast S3 and local filesystem execution tool", EnableBashCompletion: true, Commands: command.Commands(), Flags: []cli.Flag{ &cli.IntFlag{ Name: "numworkers", Value: 256, Usage: "number of workers execute operation on each object", }, &cli.IntFlag{ Name: "retry-count", Aliases: []string{"r"}, Value: 10, Usage: "number of times that a request will be retried for failures", }, &cli.StringFlag{ Name: "endpoint-url", Usage: "override default S3 host for custom services", EnvVars: []string{"OFFLOAD_S3_ENDPOINT"}, }, &cli.BoolFlag{ Name: "no-verify-ssl", Usage: "disable SSL certificate verification", }, }, Before: func(c *cli.Context) error { retryCount := c.Int("retry-count") workerCount := c.Int("numworkers") isStat := c.Bool("stat") endpointURL := c.String("endpoint-url") log.Init("error", false) // print level error only parallel.Init(workerCount) if retryCount < 0 { err := fmt.Errorf("retry count cannot be a negative value") return err } if c.Bool("no-sign-request") && c.String("profile") != "" { err := fmt.Errorf(`"no-sign-request" and "profile" flags cannot be used together`) return err } if c.Bool("no-sign-request") && c.String("credentials-file") != "" { err := fmt.Errorf(`"no-sign-request" and "credentials-file" flags cannot be used together`) return err } if isStat { stat.InitStat() } if endpointURL != "" { if !strings.HasPrefix(endpointURL, "http") { err := fmt.Errorf(`bad value for --endpoint-url %v: scheme is missing. Must be of the form http:/// or https:///`, endpointURL) return err } } return nil }, Action: func(c *cli.Context) error { if c.Bool("install-completion") { return nil } args := c.Args() if args.Present() { cli.ShowCommandHelp(c, args.First()) return cli.Exit("", 1) } return cli.ShowAppHelp(c) }, After: func(c *cli.Context) error { if c.Bool("stat") && len(stat.Statistics()) > 0 { log.Stat(stat.Statistics()) } return nil }, }, } } func (m *Module) Name() string { return Name } func (m *Module) Type() modulecapabilities.ModuleType { return modulecapabilities.Offload } func (m *Module) Init(ctx context.Context, params moduletools.ModuleInitParams, ) error { m.logger = params.GetLogger() if path := os.Getenv("PERSISTENCE_DATA_PATH"); path != "" { m.DataPath = path } if bucket := os.Getenv(s3Bucket); bucket != "" { m.Bucket = bucket } if endpoint := os.Getenv(s3Endpoint); endpoint != "" { m.Endpoint = endpoint } if eTimeout := os.Getenv(timeout); eTimeout != "" { timeoutN, err := time.ParseDuration(fmt.Sprintf("%ss", eTimeout)) if err != nil { return err } m.timeout = time.Duration(timeoutN.Seconds()) * time.Second } if concc := os.Getenv(concurrency); concc != "" { conccN, err := strconv.Atoi(concc) if err != nil { return err } m.Concurrency = conccN } if entcfg.Enabled(os.Getenv(s3BucketAutoCreate)) { if err := m.create(ctx); err != nil && !strings.Contains(err.Error(), "BucketAlreadyOwnedByYou") { return fmt.Errorf("can't create offload bucket: %s at endpoint %s %w", m.Bucket, m.Endpoint, err) } } m.logger.WithFields(logrus.Fields{ concurrency: m.Concurrency, timeout: m.timeout, s3Endpoint: m.Endpoint, s3Bucket: m.Bucket, "PERSISTENCE_DATA_PATH": m.DataPath, }).Info("offload module loaded") return nil } func (m *Module) VerifyBucket(ctx context.Context) error { if m.BucketExists.Load() { return nil } ctx, cancel := context.WithTimeout(ctx, m.timeout) defer cancel() cmd := []string{ fmt.Sprintf("--endpoint-url=%s", m.Endpoint), "ls", fmt.Sprintf("s3://%s", m.Bucket), } if err := m.app.RunContext(ctx, cmd); err != nil { return err } m.BucketExists.Store(true) return nil } func (m *Module) create(ctx context.Context) error { ctx, cancel := context.WithTimeout(ctx, m.timeout) defer cancel() cmd := []string{ fmt.Sprintf("--endpoint-url=%s", m.Endpoint), "mb", fmt.Sprintf("s3://%s", m.Bucket), } return m.app.RunContext(ctx, cmd) } // Upload uploads the content of a shard assigned to specific node to // cloud provider (S3, Azure Blob storage, Google cloud storage) // {cloud_provider}://{configured_bucket}/{className}/{shardName}/{nodeName}/{shard content} func (m *Module) Upload(ctx context.Context, className, shardName, nodeName string) error { start := time.Now() if err := validate(className, shardName, nodeName); err != nil { return err } ctx, cancel := context.WithTimeout(ctx, m.timeout) defer cancel() localPath := fmt.Sprintf("%s/%s/%s", m.DataPath, strings.ToLower(className), shardName) cmd := []string{ fmt.Sprintf("--endpoint-url=%s", m.Endpoint), "cp", fmt.Sprintf("--concurrency=%s", fmt.Sprintf("%d", m.Concurrency)), fmt.Sprintf("%s/*", localPath), fmt.Sprintf("s3://%s/%s/%s/%s/", m.Bucket, strings.ToLower(className), shardName, nodeName), } var err error defer func() { // Update few useful metrics size, _ := dirSize(localPath) m.metrics.FetchedBytes.Add(float64(size)) status := "success" if err != nil { status = "failed" } m.metrics.OpsDuration.WithLabelValues("upload", status).Observe(time.Since(start).Seconds()) }() err = m.app.RunContext(ctx, cmd) return err } // Download downloads the content of a shard to desired node from // cloud provider (S3, Azure Blob storage, Google cloud storage) // {dataPath}/{className}/{shardName}/{content} func (m *Module) Download(ctx context.Context, className, shardName, nodeName string) error { localPath := fmt.Sprintf("%s/%s/%s", m.DataPath, strings.ToLower(className), shardName) return m.DownloadToPath(ctx, className, shardName, nodeName, localPath) } func (m *Module) DownloadToPath(ctx context.Context, className, shardName, nodeName, localPath string) error { start := time.Now() if err := validate(className, shardName, nodeName); err != nil { return err } ctx, cancel := context.WithTimeout(ctx, m.timeout) defer cancel() cmd := []string{ fmt.Sprintf("--endpoint-url=%s", m.Endpoint), "cp", fmt.Sprintf("--concurrency=%s", fmt.Sprintf("%d", m.Concurrency)), fmt.Sprintf("s3://%s/%s/%s/%s/*", m.Bucket, strings.ToLower(className), shardName, nodeName), fmt.Sprintf("%s/", localPath), } var err error defer func() { // Update few useful metrics size, _ := dirSize(localPath) m.metrics.FetchedBytes.Add(float64(size)) status := "success" if err != nil { status = "failed" } m.metrics.OpsDuration.WithLabelValues("download", status).Observe(time.Since(start).Seconds()) }() err = m.app.RunContext(ctx, cmd) return err } // Delete deletes content of a shard assigned to specific node in // cloud provider (S3, Azure Blob storage, Google cloud storage) // Careful: if shardName and nodeName is passed empty it will delete all class frozen shards in cloud storage // {cloud_provider}://{configured_bucket}/{className}/{shardName}/{nodeName}/{shard content} func (m *Module) Delete(ctx context.Context, className, shardName, nodeName string) error { start := time.Now() if className == "" { return fmt.Errorf("can't pass empty class name") } if shardName == "" && nodeName != "" { return fmt.Errorf("can't pass empty shard name") } if nodeName == "" && shardName != "" { return fmt.Errorf("can't pass empty node name") } ctx, cancel := context.WithTimeout(ctx, m.timeout) defer cancel() cloudPath := fmt.Sprintf("s3://%s/%s/%s/%s/*", m.Bucket, strings.ToLower(className), shardName, nodeName) // update cloud path on deleting a class if shardName == "" && nodeName == "" { cloudPath = fmt.Sprintf("s3://%s/%s/*", m.Bucket, strings.ToLower(className)) } cmd := []string{ fmt.Sprintf("--endpoint-url=%s", m.Endpoint), "rm", cloudPath, } var err error defer func() { // Update few useful metrics status := "success" if err != nil { status = "failed" } m.metrics.OpsDuration.WithLabelValues("delete", status).Observe(time.Since(start).Seconds()) }() err = m.app.RunContext(ctx, cmd) if err != nil && !strings.Contains(err.Error(), "no object found") { return err } return nil } func dirSize(path string) (int64, error) { var size int64 err := filepath.Walk(path, func(_ string, info os.FileInfo, err error) error { if err != nil { return err } if !info.IsDir() { size += info.Size() } return err }) return size, err } func validate(className, shardName, nodeName string) error { if className == "" { return fmt.Errorf("can't pass empty class name") } if shardName == "" { return fmt.Errorf("can't pass empty tenant name") } if nodeName == "" { return fmt.Errorf("can't pass empty node name") } return nil }