ADAPT-Chase's picture
Add files using upload-large-folder tool
59bb539 verified
// _ _
// __ _____ __ ___ ___ __ _| |_ ___
// \ \ /\ / / _ \/ _` \ \ / / |/ _` | __/ _ \
// \ V V / __/ (_| |\ V /| | (_| | || __/
// \_/\_/ \___|\__,_| \_/ |_|\__,_|\__\___|
//
// Copyright © 2016 - 2025 Weaviate B.V. All rights reserved.
//
// CONTACT: hello@weaviate.io
//
package modsloads3
import (
"context"
"fmt"
"os"
"path/filepath"
"strconv"
"strings"
"sync/atomic"
"time"
"github.com/prometheus/client_golang/prometheus"
"github.com/sirupsen/logrus"
"github.com/urfave/cli/v2"
"github.com/weaviate/s5cmd/v2/command"
"github.com/weaviate/s5cmd/v2/log"
"github.com/weaviate/s5cmd/v2/log/stat"
"github.com/weaviate/s5cmd/v2/parallel"
entcfg "github.com/weaviate/weaviate/entities/config"
"github.com/weaviate/weaviate/entities/modulecapabilities"
"github.com/weaviate/weaviate/entities/moduletools"
"github.com/weaviate/weaviate/usecases/config"
"github.com/weaviate/weaviate/usecases/monitoring"
)
const (
Name = "offload-s3"
s3Endpoint = "OFFLOAD_S3_ENDPOINT"
s3BucketAutoCreate = "OFFLOAD_S3_BUCKET_AUTO_CREATE"
s3Bucket = "OFFLOAD_S3_BUCKET"
concurrency = "OFFLOAD_S3_CONCURRENCY"
timeout = "OFFLOAD_TIMEOUT"
)
// verify we implement the modules.Module interface
var (
_ = modulecapabilities.Module(&Module{})
)
type Module struct {
Endpoint string
Bucket string
BucketExists atomic.Bool
Concurrency int
DataPath string
logger logrus.FieldLogger
timeout time.Duration
app *cli.App
metrics *monitoring.TenantOffloadMetrics
}
func New() *Module {
return &Module{
Endpoint: "",
Bucket: "weaviate-offload",
Concurrency: 25,
DataPath: config.DefaultPersistenceDataPath,
timeout: 120 * time.Second,
// we use custom cli app to avoid some bugs in underlying dependencies
// specially with .After implementation.
metrics: monitoring.NewTenantOffloadMetrics(monitoring.Config{
MetricsNamespace: "weaviate",
}, prometheus.DefaultRegisterer),
app: &cli.App{
Name: "weaviate-s5cmd",
Usage: "weaviate fast S3 and local filesystem execution tool",
EnableBashCompletion: true,
Commands: command.Commands(),
Flags: []cli.Flag{
&cli.IntFlag{
Name: "numworkers",
Value: 256,
Usage: "number of workers execute operation on each object",
},
&cli.IntFlag{
Name: "retry-count",
Aliases: []string{"r"},
Value: 10,
Usage: "number of times that a request will be retried for failures",
},
&cli.StringFlag{
Name: "endpoint-url",
Usage: "override default S3 host for custom services",
EnvVars: []string{"OFFLOAD_S3_ENDPOINT"},
},
&cli.BoolFlag{
Name: "no-verify-ssl",
Usage: "disable SSL certificate verification",
},
},
Before: func(c *cli.Context) error {
retryCount := c.Int("retry-count")
workerCount := c.Int("numworkers")
isStat := c.Bool("stat")
endpointURL := c.String("endpoint-url")
log.Init("error", false) // print level error only
parallel.Init(workerCount)
if retryCount < 0 {
err := fmt.Errorf("retry count cannot be a negative value")
return err
}
if c.Bool("no-sign-request") && c.String("profile") != "" {
err := fmt.Errorf(`"no-sign-request" and "profile" flags cannot be used together`)
return err
}
if c.Bool("no-sign-request") && c.String("credentials-file") != "" {
err := fmt.Errorf(`"no-sign-request" and "credentials-file" flags cannot be used together`)
return err
}
if isStat {
stat.InitStat()
}
if endpointURL != "" {
if !strings.HasPrefix(endpointURL, "http") {
err := fmt.Errorf(`bad value for --endpoint-url %v: scheme is missing. Must be of the form http://<hostname>/ or https://<hostname>/`, endpointURL)
return err
}
}
return nil
},
Action: func(c *cli.Context) error {
if c.Bool("install-completion") {
return nil
}
args := c.Args()
if args.Present() {
cli.ShowCommandHelp(c, args.First())
return cli.Exit("", 1)
}
return cli.ShowAppHelp(c)
},
After: func(c *cli.Context) error {
if c.Bool("stat") && len(stat.Statistics()) > 0 {
log.Stat(stat.Statistics())
}
return nil
},
},
}
}
func (m *Module) Name() string {
return Name
}
func (m *Module) Type() modulecapabilities.ModuleType {
return modulecapabilities.Offload
}
func (m *Module) Init(ctx context.Context,
params moduletools.ModuleInitParams,
) error {
m.logger = params.GetLogger()
if path := os.Getenv("PERSISTENCE_DATA_PATH"); path != "" {
m.DataPath = path
}
if bucket := os.Getenv(s3Bucket); bucket != "" {
m.Bucket = bucket
}
if endpoint := os.Getenv(s3Endpoint); endpoint != "" {
m.Endpoint = endpoint
}
if eTimeout := os.Getenv(timeout); eTimeout != "" {
timeoutN, err := time.ParseDuration(fmt.Sprintf("%ss", eTimeout))
if err != nil {
return err
}
m.timeout = time.Duration(timeoutN.Seconds()) * time.Second
}
if concc := os.Getenv(concurrency); concc != "" {
conccN, err := strconv.Atoi(concc)
if err != nil {
return err
}
m.Concurrency = conccN
}
if entcfg.Enabled(os.Getenv(s3BucketAutoCreate)) {
if err := m.create(ctx); err != nil && !strings.Contains(err.Error(), "BucketAlreadyOwnedByYou") {
return fmt.Errorf("can't create offload bucket: %s at endpoint %s %w", m.Bucket, m.Endpoint, err)
}
}
m.logger.WithFields(logrus.Fields{
concurrency: m.Concurrency,
timeout: m.timeout,
s3Endpoint: m.Endpoint,
s3Bucket: m.Bucket,
"PERSISTENCE_DATA_PATH": m.DataPath,
}).Info("offload module loaded")
return nil
}
func (m *Module) VerifyBucket(ctx context.Context) error {
if m.BucketExists.Load() {
return nil
}
ctx, cancel := context.WithTimeout(ctx, m.timeout)
defer cancel()
cmd := []string{
fmt.Sprintf("--endpoint-url=%s", m.Endpoint),
"ls",
fmt.Sprintf("s3://%s", m.Bucket),
}
if err := m.app.RunContext(ctx, cmd); err != nil {
return err
}
m.BucketExists.Store(true)
return nil
}
func (m *Module) create(ctx context.Context) error {
ctx, cancel := context.WithTimeout(ctx, m.timeout)
defer cancel()
cmd := []string{
fmt.Sprintf("--endpoint-url=%s", m.Endpoint),
"mb",
fmt.Sprintf("s3://%s", m.Bucket),
}
return m.app.RunContext(ctx, cmd)
}
// Upload uploads the content of a shard assigned to specific node to
// cloud provider (S3, Azure Blob storage, Google cloud storage)
// {cloud_provider}://{configured_bucket}/{className}/{shardName}/{nodeName}/{shard content}
func (m *Module) Upload(ctx context.Context, className, shardName, nodeName string) error {
start := time.Now()
if err := validate(className, shardName, nodeName); err != nil {
return err
}
ctx, cancel := context.WithTimeout(ctx, m.timeout)
defer cancel()
localPath := fmt.Sprintf("%s/%s/%s", m.DataPath, strings.ToLower(className), shardName)
cmd := []string{
fmt.Sprintf("--endpoint-url=%s", m.Endpoint),
"cp",
fmt.Sprintf("--concurrency=%s", fmt.Sprintf("%d", m.Concurrency)),
fmt.Sprintf("%s/*", localPath),
fmt.Sprintf("s3://%s/%s/%s/%s/", m.Bucket, strings.ToLower(className), shardName, nodeName),
}
var err error
defer func() {
// Update few useful metrics
size, _ := dirSize(localPath)
m.metrics.FetchedBytes.Add(float64(size))
status := "success"
if err != nil {
status = "failed"
}
m.metrics.OpsDuration.WithLabelValues("upload", status).Observe(time.Since(start).Seconds())
}()
err = m.app.RunContext(ctx, cmd)
return err
}
// Download downloads the content of a shard to desired node from
// cloud provider (S3, Azure Blob storage, Google cloud storage)
// {dataPath}/{className}/{shardName}/{content}
func (m *Module) Download(ctx context.Context, className, shardName, nodeName string) error {
localPath := fmt.Sprintf("%s/%s/%s", m.DataPath, strings.ToLower(className), shardName)
return m.DownloadToPath(ctx, className, shardName, nodeName, localPath)
}
func (m *Module) DownloadToPath(ctx context.Context, className, shardName, nodeName, localPath string) error {
start := time.Now()
if err := validate(className, shardName, nodeName); err != nil {
return err
}
ctx, cancel := context.WithTimeout(ctx, m.timeout)
defer cancel()
cmd := []string{
fmt.Sprintf("--endpoint-url=%s", m.Endpoint),
"cp",
fmt.Sprintf("--concurrency=%s", fmt.Sprintf("%d", m.Concurrency)),
fmt.Sprintf("s3://%s/%s/%s/%s/*", m.Bucket, strings.ToLower(className), shardName, nodeName),
fmt.Sprintf("%s/", localPath),
}
var err error
defer func() {
// Update few useful metrics
size, _ := dirSize(localPath)
m.metrics.FetchedBytes.Add(float64(size))
status := "success"
if err != nil {
status = "failed"
}
m.metrics.OpsDuration.WithLabelValues("download", status).Observe(time.Since(start).Seconds())
}()
err = m.app.RunContext(ctx, cmd)
return err
}
// Delete deletes content of a shard assigned to specific node in
// cloud provider (S3, Azure Blob storage, Google cloud storage)
// Careful: if shardName and nodeName is passed empty it will delete all class frozen shards in cloud storage
// {cloud_provider}://{configured_bucket}/{className}/{shardName}/{nodeName}/{shard content}
func (m *Module) Delete(ctx context.Context, className, shardName, nodeName string) error {
start := time.Now()
if className == "" {
return fmt.Errorf("can't pass empty class name")
}
if shardName == "" && nodeName != "" {
return fmt.Errorf("can't pass empty shard name")
}
if nodeName == "" && shardName != "" {
return fmt.Errorf("can't pass empty node name")
}
ctx, cancel := context.WithTimeout(ctx, m.timeout)
defer cancel()
cloudPath := fmt.Sprintf("s3://%s/%s/%s/%s/*", m.Bucket, strings.ToLower(className), shardName, nodeName)
// update cloud path on deleting a class
if shardName == "" && nodeName == "" {
cloudPath = fmt.Sprintf("s3://%s/%s/*", m.Bucket, strings.ToLower(className))
}
cmd := []string{
fmt.Sprintf("--endpoint-url=%s", m.Endpoint),
"rm",
cloudPath,
}
var err error
defer func() {
// Update few useful metrics
status := "success"
if err != nil {
status = "failed"
}
m.metrics.OpsDuration.WithLabelValues("delete", status).Observe(time.Since(start).Seconds())
}()
err = m.app.RunContext(ctx, cmd)
if err != nil && !strings.Contains(err.Error(), "no object found") {
return err
}
return nil
}
func dirSize(path string) (int64, error) {
var size int64
err := filepath.Walk(path, func(_ string, info os.FileInfo, err error) error {
if err != nil {
return err
}
if !info.IsDir() {
size += info.Size()
}
return err
})
return size, err
}
func validate(className, shardName, nodeName string) error {
if className == "" {
return fmt.Errorf("can't pass empty class name")
}
if shardName == "" {
return fmt.Errorf("can't pass empty tenant name")
}
if nodeName == "" {
return fmt.Errorf("can't pass empty node name")
}
return nil
}