mirror of
https://github.com/opencloud-eu/opencloud.git
synced 2026-03-07 16:56:58 -05:00
272 lines
7.0 KiB
Go
272 lines
7.0 KiB
Go
package command
|
|
|
|
import (
|
|
"context"
|
|
"flag"
|
|
"os"
|
|
"os/signal"
|
|
"path"
|
|
"time"
|
|
|
|
"github.com/gofrs/uuid"
|
|
"github.com/owncloud/ocis/storage/pkg/service/external"
|
|
|
|
"github.com/cs3org/reva/cmd/revad/runtime"
|
|
"github.com/micro/cli/v2"
|
|
"github.com/oklog/run"
|
|
"github.com/owncloud/ocis/storage/pkg/config"
|
|
"github.com/owncloud/ocis/storage/pkg/flagset"
|
|
"github.com/owncloud/ocis/storage/pkg/server/debug"
|
|
)
|
|
|
|
// StorageMetadata the entrypoint for the storage-storage-metadata command.
|
|
//
|
|
// It provides a ocis-specific storage store metadata (shares,account,settings...)
|
|
func StorageMetadata(cfg *config.Config) *cli.Command {
|
|
return &cli.Command{
|
|
Name: "storage-metadata",
|
|
Usage: "Start storage-metadata service",
|
|
// TODO(refs) at this point it might make sense delegate log flags to each individual storage command.
|
|
Flags: append(flagset.StorageMetadata(cfg), flagset.RootWithConfig(cfg)...),
|
|
Category: "Extensions",
|
|
Before: func(c *cli.Context) error {
|
|
storageRoot := c.String("storage-root")
|
|
|
|
cfg.Reva.Storages.OwnCloud.Root = storageRoot
|
|
cfg.Reva.Storages.EOS.Root = storageRoot
|
|
cfg.Reva.Storages.Local.Root = storageRoot
|
|
cfg.Reva.Storages.S3.Root = storageRoot
|
|
cfg.Reva.Storages.Home.Root = storageRoot
|
|
|
|
return nil
|
|
},
|
|
Action: func(c *cli.Context) error {
|
|
logger := NewLogger(cfg)
|
|
if cfg.Tracing.Enabled {
|
|
switch t := cfg.Tracing.Type; t {
|
|
case "agent":
|
|
logger.Error().
|
|
Str("type", t).
|
|
Msg("Reva only supports the jaeger tracing backend")
|
|
|
|
case "jaeger":
|
|
logger.Info().
|
|
Str("type", t).
|
|
Msg("configuring storage to use the jaeger tracing backend")
|
|
|
|
case "zipkin":
|
|
logger.Error().
|
|
Str("type", t).
|
|
Msg("Reva only supports the jaeger tracing backend")
|
|
|
|
default:
|
|
logger.Warn().
|
|
Str("type", t).
|
|
Msg("Unknown tracing backend")
|
|
}
|
|
|
|
} else {
|
|
logger.Debug().
|
|
Msg("Tracing is not enabled")
|
|
}
|
|
|
|
var (
|
|
gr = run.Group{}
|
|
ctx, cancel = context.WithCancel(context.Background())
|
|
//metrics = metrics.New()
|
|
)
|
|
|
|
defer cancel()
|
|
|
|
{
|
|
uuid := uuid.Must(uuid.NewV4())
|
|
pidFile := path.Join(os.TempDir(), "revad-"+c.Command.Name+"-"+uuid.String()+".pid")
|
|
|
|
// Disable home because the metadata is stored independently
|
|
// of the user. This also means that a valid-token without any user-id
|
|
// is allowed to write to the metadata-storage.
|
|
cfg.Reva.Storages.Common.EnableHome = false
|
|
cfg.Reva.Storages.EOS.EnableHome = false
|
|
cfg.Reva.Storages.Local.EnableHome = false
|
|
cfg.Reva.Storages.OwnCloud.EnableHome = false
|
|
cfg.Reva.Storages.S3.EnableHome = false
|
|
|
|
rcfg := map[string]interface{}{
|
|
"core": map[string]interface{}{
|
|
"max_cpus": cfg.Reva.StorageMetadata.MaxCPUs,
|
|
"tracing_enabled": cfg.Tracing.Enabled,
|
|
"tracing_endpoint": cfg.Tracing.Endpoint,
|
|
"tracing_collector": cfg.Tracing.Collector,
|
|
"tracing_service_name": c.Command.Name,
|
|
},
|
|
"shared": map[string]interface{}{
|
|
"jwt_secret": cfg.Reva.JWTSecret,
|
|
"gatewaysvc": cfg.Reva.Gateway.Endpoint,
|
|
},
|
|
"grpc": map[string]interface{}{
|
|
"network": cfg.Reva.StorageMetadata.GRPCNetwork,
|
|
"address": cfg.Reva.StorageMetadata.GRPCAddr,
|
|
"interceptors": map[string]interface{}{
|
|
"log": map[string]interface{}{},
|
|
},
|
|
"services": map[string]interface{}{
|
|
"storageprovider": map[string]interface{}{
|
|
"mount_path": "/meta",
|
|
"driver": cfg.Reva.StorageMetadata.Driver,
|
|
"drivers": drivers(cfg),
|
|
"data_server_url": cfg.Reva.StorageMetadata.DataServerURL,
|
|
"tmp_folder": cfg.Reva.StorageMetadata.TempFolder,
|
|
},
|
|
},
|
|
},
|
|
"http": map[string]interface{}{
|
|
"network": cfg.Reva.StorageMetadata.HTTPNetwork,
|
|
"address": cfg.Reva.StorageMetadata.HTTPAddr,
|
|
// TODO build services dynamically
|
|
"services": map[string]interface{}{
|
|
"dataprovider": map[string]interface{}{
|
|
"prefix": "data",
|
|
"driver": cfg.Reva.StorageMetadata.Driver,
|
|
"drivers": drivers(cfg),
|
|
"timeout": 86400,
|
|
"insecure": true,
|
|
"disable_tus": true,
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
gr.Add(func() error {
|
|
runtime.RunWithOptions(
|
|
rcfg,
|
|
pidFile,
|
|
runtime.WithLogger(&logger.Logger),
|
|
)
|
|
return nil
|
|
}, func(_ error) {
|
|
logger.Info().
|
|
Str("server", c.Command.Name).
|
|
Msg("Shutting down server")
|
|
|
|
cancel()
|
|
})
|
|
}
|
|
|
|
{
|
|
server, err := debug.Server(
|
|
debug.Name(c.Command.Name+"-debug"),
|
|
debug.Addr(cfg.Reva.StorageMetadata.DebugAddr),
|
|
debug.Logger(logger),
|
|
debug.Context(ctx),
|
|
debug.Config(cfg),
|
|
)
|
|
|
|
if err != nil {
|
|
logger.Info().
|
|
Err(err).
|
|
Str("server", c.Command.Name+"-debug").
|
|
Msg("Failed to initialize server")
|
|
|
|
return err
|
|
}
|
|
|
|
gr.Add(func() error {
|
|
return server.ListenAndServe()
|
|
}, func(_ error) {
|
|
ctx, timeout := context.WithTimeout(ctx, 5*time.Second)
|
|
|
|
defer timeout()
|
|
defer cancel()
|
|
|
|
if err := server.Shutdown(ctx); err != nil {
|
|
logger.Info().
|
|
Err(err).
|
|
Str("server", c.Command.Name+"-debug").
|
|
Msg("Failed to shutdown server")
|
|
} else {
|
|
logger.Info().
|
|
Str("server", c.Command.Name+"-debug").
|
|
Msg("Shutting down server")
|
|
}
|
|
})
|
|
}
|
|
|
|
{
|
|
stop := make(chan os.Signal, 1)
|
|
|
|
gr.Add(func() error {
|
|
signal.Notify(stop, os.Interrupt)
|
|
<-stop
|
|
|
|
return nil
|
|
}, func(err error) {
|
|
close(stop)
|
|
cancel()
|
|
})
|
|
}
|
|
|
|
if err := external.RegisterGRPCEndpoint(
|
|
ctx,
|
|
"com.owncloud.storage.metadata",
|
|
uuid.Must(uuid.NewV4()).String(),
|
|
cfg.Reva.StorageMetadata.GRPCAddr,
|
|
logger,
|
|
); err != nil {
|
|
logger.Fatal().Err(err).Msg("failed to register the grpc endpoint")
|
|
}
|
|
|
|
return gr.Run()
|
|
},
|
|
}
|
|
}
|
|
|
|
// SutureService allows for the storage-metadata command to be embedded and supervised by a suture supervisor tree.
|
|
type SutureService struct {
|
|
ctx context.Context
|
|
cancel context.CancelFunc // used to cancel the context go-micro services used to shutdown a service.
|
|
cfg *config.Config
|
|
}
|
|
|
|
// NewSutureService creates a new storagemetadata.SutureService
|
|
func NewStorageMetadata(ctx context.Context, o ...Option) SutureService {
|
|
sctx, cancel := context.WithCancel(ctx)
|
|
cfg := config.New()
|
|
|
|
opts := newOptions(o...)
|
|
|
|
// merge config and options
|
|
cfg.Context = sctx
|
|
|
|
cfg.Log.Level = opts.LogLevel
|
|
cfg.Log.Pretty = opts.LogPretty
|
|
cfg.Log.Color = opts.LogColor
|
|
|
|
return SutureService{
|
|
ctx: sctx,
|
|
cancel: cancel,
|
|
cfg: cfg,
|
|
}
|
|
}
|
|
|
|
func (s SutureService) Serve() {
|
|
f := &flag.FlagSet{}
|
|
for k := range StorageMetadata(s.cfg).Flags {
|
|
if err := StorageMetadata(s.cfg).Flags[k].Apply(f); err != nil {
|
|
return
|
|
}
|
|
}
|
|
ctx := cli.NewContext(nil, f, nil)
|
|
if StorageMetadata(s.cfg).Before != nil {
|
|
if err := StorageMetadata(s.cfg).Before(ctx); err != nil {
|
|
return
|
|
}
|
|
}
|
|
if err := StorageMetadata(s.cfg).Action(ctx); err != nil {
|
|
return
|
|
}
|
|
}
|
|
|
|
func (s SutureService) Stop() {
|
|
s.cancel()
|
|
}
|