mirror of
https://github.com/syncthing/syncthing.git
synced 2026-04-26 17:28:25 -04:00
This updates our logging framework from legacy freetext strings using the `log` package to structured log entries using `log/slog`. I have updated all INFO or higher level entries, but not yet DEBUG (😓)... So, at a high level: There is a slight change in log levels, effectively adding a new warning level: - DEBUG is still debug (ideally not for users but developers, though this is something we need to work on) - INFO is still info, though I've added more data here, effectively making Syncthing more verbose by default (more on this below) - WARNING is a new log level that is different from the _old_ WARNING (more below) - ERROR is what was WARNING before -- problems that must be dealt with, and also bubbled as a popup in the GUI. A new feature is that the logging level can be set per package to something other than just debug or info, and hence I feel that we can add a bit more things into INFO while moving some (in fact, most) current INFO level warnings into WARNING. For example, I think it's justified to get a log of synced files in INFO and sync failures in WARNING. These are things that have historically been tricky to debug properly, and having more information by default will be useful to many, while still making it possible get close to told level of inscrutability by setting the log level to WARNING. I'd like to get to a stage where DEBUG is never necessary to just figure out what's going on, as opposed to trying to narrow down a likely bug. Code wise: - Our logging object, generally known as `l` in each package, is now a new adapter object that provides the old API on top of the newer one. (This should go away once all old log entries are migrated.) This is only for `l.Debugln` and `l.Debugf`. - There is a new level tracker that keeps the log level for each package. - There is a nested setup of handlers, since the structure mandated by `log/slog` is slightly convoluted (imho). We do this because we need to do formatting at a "medium" level internally so we can buffer log lines in text format but with separate timestamp and log level for the API/GUI to consume. - The `debug` API call becomes a `loglevels` API call, which can set the log level to `DEBUG`, `INFO`, `WARNING` or `ERROR` per package. The GUI is updated to handle this. - Our custom `sync` package provided some debugging of mutexes quite strongly integrated into the old logging framework, only turned on when `STTRACE` was set to certain values at startup, etc. It's been a long time since this has been useful; I removed it. - The `STTRACE` env var remains and can be used the same way as before, while additionally permitting specific log levels to be specified, `STTRACE=model:WARN,scanner:DEBUG`. - There is a new command line option `--log-level=INFO` to set the default log level. - The command line options `--log-flags` and `--verbose` go away, but are currently retained as hidden & ignored options since we set them by default in some of our startup examples and Syncthing would otherwise fail to start. Sample format messages: ``` 2009-02-13 23:31:30 INF A basic info line (attr1="val with spaces" attr2=2 attr3="val\"quote" a=a log.pkg=slogutil) 2009-02-13 23:31:30 INF An info line with grouped values (attr1=val1 foo.attr2=2 foo.bar.attr3=3 a=a log.pkg=slogutil) 2009-02-13 23:31:30 INF An info line with grouped values via logger (foo.attr1=val1 foo.attr2=2 a=a log.pkg=slogutil) 2009-02-13 23:31:30 INF An info line with nested grouped values via logger (bar.foo.attr1=val1 bar.foo.attr2=2 a=a log.pkg=slogutil) 2009-02-13 23:31:30 WRN A warning entry (a=a log.pkg=slogutil) 2009-02-13 23:31:30 ERR An error (a=a log.pkg=slogutil) ``` --------- Co-authored-by: Ross Smith II <ross@smithii.com>
416 lines
14 KiB
Go
416 lines
14 KiB
Go
// Copyright (C) 2015 The Syncthing Authors.
|
|
//
|
|
// This Source Code Form is subject to the terms of the Mozilla Public
|
|
// License, v. 2.0. If a copy of the MPL was not distributed with this file,
|
|
// You can obtain one at https://mozilla.org/MPL/2.0/.
|
|
|
|
//go:generate -command counterfeiter go run github.com/maxbrunsfeld/counterfeiter/v6
|
|
//go:generate counterfeiter -o mocks/folderSummaryService.go --fake-name FolderSummaryService . FolderSummaryService
|
|
|
|
package model
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/thejerf/suture/v4"
|
|
|
|
"github.com/syncthing/syncthing/internal/db"
|
|
"github.com/syncthing/syncthing/lib/config"
|
|
"github.com/syncthing/syncthing/lib/events"
|
|
"github.com/syncthing/syncthing/lib/protocol"
|
|
"github.com/syncthing/syncthing/lib/svcutil"
|
|
)
|
|
|
|
type FolderSummaryService interface {
|
|
suture.Service
|
|
Summary(folder string) (*FolderSummary, error)
|
|
}
|
|
|
|
// The folderSummaryService adds summary information events (FolderSummary and
|
|
// FolderCompletion) into the event stream at certain intervals.
|
|
type folderSummaryService struct {
|
|
*suture.Supervisor
|
|
|
|
cfg config.Wrapper
|
|
model Model
|
|
id protocol.DeviceID
|
|
evLogger events.Logger
|
|
immediate chan string
|
|
|
|
// For keeping track of folders to recalculate for
|
|
foldersMut sync.Mutex
|
|
folders map[string]struct{}
|
|
}
|
|
|
|
func NewFolderSummaryService(cfg config.Wrapper, m Model, id protocol.DeviceID, evLogger events.Logger) FolderSummaryService {
|
|
service := &folderSummaryService{
|
|
Supervisor: suture.New("folderSummaryService", svcutil.SpecWithDebugLogger()),
|
|
cfg: cfg,
|
|
model: m,
|
|
id: id,
|
|
evLogger: evLogger,
|
|
immediate: make(chan string),
|
|
folders: make(map[string]struct{}),
|
|
}
|
|
|
|
service.Add(svcutil.AsService(service.listenForUpdates, fmt.Sprintf("%s/listenForUpdates", service)))
|
|
service.Add(svcutil.AsService(service.calculateSummaries, fmt.Sprintf("%s/calculateSummaries", service)))
|
|
|
|
return service
|
|
}
|
|
|
|
func (c *folderSummaryService) String() string {
|
|
return fmt.Sprintf("FolderSummaryService@%p", c)
|
|
}
|
|
|
|
// FolderSummary replaces the previously used map[string]interface{}, and needs
|
|
// to keep the structure/naming for api backwards compatibility
|
|
type FolderSummary struct {
|
|
Errors int `json:"errors"`
|
|
PullErrors int `json:"pullErrors"` // deprecated
|
|
|
|
Invalid string `json:"invalid"` // deprecated
|
|
|
|
GlobalFiles int `json:"globalFiles"`
|
|
GlobalDirectories int `json:"globalDirectories"`
|
|
GlobalSymlinks int `json:"globalSymlinks"`
|
|
GlobalDeleted int `json:"globalDeleted"`
|
|
GlobalBytes int64 `json:"globalBytes"`
|
|
GlobalTotalItems int `json:"globalTotalItems"`
|
|
|
|
LocalFiles int `json:"localFiles"`
|
|
LocalDirectories int `json:"localDirectories"`
|
|
LocalSymlinks int `json:"localSymlinks"`
|
|
LocalDeleted int `json:"localDeleted"`
|
|
LocalBytes int64 `json:"localBytes"`
|
|
LocalTotalItems int `json:"localTotalItems"`
|
|
|
|
NeedFiles int `json:"needFiles"`
|
|
NeedDirectories int `json:"needDirectories"`
|
|
NeedSymlinks int `json:"needSymlinks"`
|
|
NeedDeletes int `json:"needDeletes"`
|
|
NeedBytes int64 `json:"needBytes"`
|
|
NeedTotalItems int `json:"needTotalItems"`
|
|
|
|
ReceiveOnlyChangedFiles int `json:"receiveOnlyChangedFiles"`
|
|
ReceiveOnlyChangedDirectories int `json:"receiveOnlyChangedDirectories"`
|
|
ReceiveOnlyChangedSymlinks int `json:"receiveOnlyChangedSymlinks"`
|
|
ReceiveOnlyChangedDeletes int `json:"receiveOnlyChangedDeletes"`
|
|
ReceiveOnlyChangedBytes int64 `json:"receiveOnlyChangedBytes"`
|
|
ReceiveOnlyTotalItems int `json:"receiveOnlyTotalItems"`
|
|
|
|
InSyncFiles int `json:"inSyncFiles"`
|
|
InSyncBytes int64 `json:"inSyncBytes"`
|
|
|
|
State string `json:"state"`
|
|
StateChanged time.Time `json:"stateChanged"`
|
|
Error string `json:"error"`
|
|
|
|
Version int64 `json:"version"` // deprecated
|
|
Sequence int64 `json:"sequence"`
|
|
RemoteSequence map[protocol.DeviceID]int64 `json:"remoteSequence"`
|
|
|
|
IgnorePatterns bool `json:"ignorePatterns"`
|
|
WatchError string `json:"watchError"`
|
|
}
|
|
|
|
func (c *folderSummaryService) Summary(folder string) (*FolderSummary, error) {
|
|
res := new(FolderSummary)
|
|
|
|
var local, global, need, ro db.Counts
|
|
var ourSeq int64
|
|
var remoteSeq map[protocol.DeviceID]int64
|
|
errs, err := c.model.FolderErrors(folder)
|
|
if err == nil {
|
|
global, _ = c.model.GlobalSize(folder)
|
|
local, _ = c.model.LocalSize(folder, protocol.LocalDeviceID)
|
|
need, _ = c.model.NeedSize(folder, protocol.LocalDeviceID)
|
|
ro, _ = c.model.ReceiveOnlySize(folder)
|
|
ourSeq, _ = c.model.Sequence(folder, protocol.LocalDeviceID)
|
|
remoteSeq, _ = c.model.RemoteSequences(folder)
|
|
}
|
|
// For API backwards compatibility (SyncTrayzor needs it) an empty folder
|
|
// summary is returned for not running folders, an error might actually be
|
|
// more appropriate
|
|
if err != nil && !errors.Is(err, ErrFolderPaused) && !errors.Is(err, ErrFolderNotRunning) {
|
|
return nil, err
|
|
}
|
|
|
|
res.Errors = len(errs)
|
|
res.PullErrors = len(errs) // deprecated
|
|
|
|
res.Invalid = "" // Deprecated, retains external API for now
|
|
|
|
res.GlobalFiles, res.GlobalDirectories, res.GlobalSymlinks, res.GlobalDeleted, res.GlobalBytes, res.GlobalTotalItems = global.Files, global.Directories, global.Symlinks, global.Deleted, global.Bytes, global.TotalItems()
|
|
|
|
res.LocalFiles, res.LocalDirectories, res.LocalSymlinks, res.LocalDeleted, res.LocalBytes, res.LocalTotalItems = local.Files, local.Directories, local.Symlinks, local.Deleted, local.Bytes, local.TotalItems()
|
|
|
|
fcfg, haveFcfg := c.cfg.Folder(folder)
|
|
|
|
if haveFcfg && fcfg.IgnoreDelete {
|
|
need.Deleted = 0
|
|
}
|
|
|
|
need.Bytes -= c.model.FolderProgressBytesCompleted(folder)
|
|
// This may happen if we are in progress of pulling files that were
|
|
// deleted globally after the pull started.
|
|
if need.Bytes < 0 {
|
|
need.Bytes = 0
|
|
}
|
|
res.NeedFiles, res.NeedDirectories, res.NeedSymlinks, res.NeedDeletes, res.NeedBytes, res.NeedTotalItems = need.Files, need.Directories, need.Symlinks, need.Deleted, need.Bytes, need.TotalItems()
|
|
|
|
if haveFcfg && (fcfg.Type == config.FolderTypeReceiveOnly || fcfg.Type == config.FolderTypeReceiveEncrypted) {
|
|
// Add statistics for things that have changed locally in a receive
|
|
// only or receive encrypted folder.
|
|
res.ReceiveOnlyChangedFiles = ro.Files
|
|
res.ReceiveOnlyChangedDirectories = ro.Directories
|
|
res.ReceiveOnlyChangedSymlinks = ro.Symlinks
|
|
res.ReceiveOnlyChangedDeletes = ro.Deleted
|
|
res.ReceiveOnlyChangedBytes = ro.Bytes
|
|
res.ReceiveOnlyTotalItems = ro.TotalItems()
|
|
}
|
|
|
|
res.InSyncFiles, res.InSyncBytes = global.Files-need.Files, global.Bytes-need.Bytes
|
|
|
|
res.State, res.StateChanged, err = c.model.State(folder)
|
|
if err != nil {
|
|
res.Error = err.Error()
|
|
}
|
|
|
|
res.Version = ourSeq // legacy
|
|
res.Sequence = ourSeq
|
|
res.RemoteSequence = remoteSeq
|
|
|
|
ignorePatterns, _, _ := c.model.CurrentIgnores(folder)
|
|
res.IgnorePatterns = false
|
|
for _, line := range ignorePatterns {
|
|
if len(line) > 0 && !strings.HasPrefix(line, "//") {
|
|
res.IgnorePatterns = true
|
|
break
|
|
}
|
|
}
|
|
|
|
err = c.model.WatchError(folder)
|
|
if err != nil {
|
|
res.WatchError = err.Error()
|
|
}
|
|
|
|
return res, nil
|
|
}
|
|
|
|
// listenForUpdates subscribes to the event bus and makes note of folders that
|
|
// need their data recalculated.
|
|
func (c *folderSummaryService) listenForUpdates(ctx context.Context) error {
|
|
sub := c.evLogger.Subscribe(events.LocalIndexUpdated | events.RemoteIndexUpdated | events.StateChanged | events.RemoteDownloadProgress | events.DeviceConnected | events.ClusterConfigReceived | events.FolderWatchStateChanged | events.DownloadProgress)
|
|
defer sub.Unsubscribe()
|
|
|
|
for {
|
|
// This loop needs to be fast so we don't miss too many events.
|
|
|
|
select {
|
|
case ev, ok := <-sub.C():
|
|
if !ok {
|
|
<-ctx.Done()
|
|
return ctx.Err()
|
|
}
|
|
c.processUpdate(ev)
|
|
case <-ctx.Done():
|
|
return ctx.Err()
|
|
}
|
|
}
|
|
}
|
|
|
|
func (c *folderSummaryService) processUpdate(ev events.Event) {
|
|
var folder string
|
|
|
|
switch ev.Type {
|
|
case events.DeviceConnected, events.ClusterConfigReceived:
|
|
// When a device connects we schedule a refresh of all
|
|
// folders shared with that device.
|
|
|
|
var deviceID protocol.DeviceID
|
|
if ev.Type == events.DeviceConnected {
|
|
data := ev.Data.(map[string]string)
|
|
deviceID, _ = protocol.DeviceIDFromString(data["id"])
|
|
} else {
|
|
data := ev.Data.(ClusterConfigReceivedEventData)
|
|
deviceID = data.Device
|
|
}
|
|
|
|
c.foldersMut.Lock()
|
|
nextFolder:
|
|
for _, folder := range c.cfg.Folders() {
|
|
for _, dev := range folder.Devices {
|
|
if dev.DeviceID == deviceID {
|
|
c.folders[folder.ID] = struct{}{}
|
|
continue nextFolder
|
|
}
|
|
}
|
|
}
|
|
c.foldersMut.Unlock()
|
|
|
|
return
|
|
|
|
case events.DownloadProgress:
|
|
data := ev.Data.(map[string]map[string]*PullerProgress)
|
|
c.foldersMut.Lock()
|
|
for folder := range data {
|
|
c.folders[folder] = struct{}{}
|
|
}
|
|
c.foldersMut.Unlock()
|
|
return
|
|
|
|
case events.StateChanged:
|
|
data := ev.Data.(map[string]interface{})
|
|
if data["to"].(string) != "idle" {
|
|
return
|
|
}
|
|
if from := data["from"].(string); from != "syncing" && from != "sync-preparing" {
|
|
return
|
|
}
|
|
|
|
// The folder changed to idle from syncing. We should do an
|
|
// immediate refresh to update the GUI. The send to
|
|
// c.immediate must be nonblocking so that we can continue
|
|
// handling events.
|
|
|
|
folder = data["folder"].(string)
|
|
select {
|
|
case c.immediate <- folder:
|
|
c.foldersMut.Lock()
|
|
delete(c.folders, folder)
|
|
c.foldersMut.Unlock()
|
|
return
|
|
default:
|
|
// Refresh whenever we do the next summary.
|
|
}
|
|
|
|
default:
|
|
// The other events all have a "folder" attribute that they
|
|
// affect. Whenever the local or remote index is updated for a
|
|
// given folder we make a note of it.
|
|
// This folder needs to be refreshed whenever we do the next
|
|
// refresh.
|
|
|
|
folder = ev.Data.(map[string]interface{})["folder"].(string)
|
|
}
|
|
|
|
c.foldersMut.Lock()
|
|
c.folders[folder] = struct{}{}
|
|
c.foldersMut.Unlock()
|
|
}
|
|
|
|
// calculateSummaries periodically recalculates folder summaries and
|
|
// completion percentage, and sends the results on the event bus.
|
|
func (c *folderSummaryService) calculateSummaries(ctx context.Context) error {
|
|
const pumpInterval = 2 * time.Second
|
|
pump := time.NewTimer(pumpInterval)
|
|
|
|
for {
|
|
select {
|
|
case <-pump.C:
|
|
t0 := time.Now()
|
|
for _, folder := range c.foldersToHandle() {
|
|
select {
|
|
case <-ctx.Done():
|
|
return ctx.Err()
|
|
default:
|
|
}
|
|
c.sendSummary(ctx, folder)
|
|
}
|
|
|
|
// We don't want to spend all our time calculating summaries. Lets
|
|
// set an arbitrary limit at not spending more than about 30% of
|
|
// our time here...
|
|
wait := 2*time.Since(t0) + pumpInterval
|
|
pump.Reset(wait)
|
|
|
|
case folder := <-c.immediate:
|
|
c.sendSummary(ctx, folder)
|
|
|
|
case <-ctx.Done():
|
|
return ctx.Err()
|
|
}
|
|
}
|
|
}
|
|
|
|
// foldersToHandle returns the list of folders needing a summary update, and
|
|
// clears the list.
|
|
func (c *folderSummaryService) foldersToHandle() []string {
|
|
c.foldersMut.Lock()
|
|
res := make([]string, 0, len(c.folders))
|
|
for folder := range c.folders {
|
|
res = append(res, folder)
|
|
delete(c.folders, folder)
|
|
}
|
|
c.foldersMut.Unlock()
|
|
return res
|
|
}
|
|
|
|
type FolderSummaryEventData struct {
|
|
Folder string `json:"folder"`
|
|
Summary *FolderSummary `json:"summary"`
|
|
}
|
|
|
|
// sendSummary send the summary events for a single folder
|
|
func (c *folderSummaryService) sendSummary(ctx context.Context, folder string) {
|
|
// The folder summary contains how many bytes, files etc
|
|
// are in the folder and how in sync we are.
|
|
data, err := c.Summary(folder)
|
|
if err != nil {
|
|
return
|
|
}
|
|
c.evLogger.Log(events.FolderSummary, FolderSummaryEventData{
|
|
Folder: folder,
|
|
Summary: data,
|
|
})
|
|
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeGlobal, metricTypeFiles).Set(float64(data.GlobalFiles))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeGlobal, metricTypeDirectories).Set(float64(data.GlobalDirectories))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeGlobal, metricTypeSymlinks).Set(float64(data.GlobalSymlinks))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeGlobal, metricTypeDeleted).Set(float64(data.GlobalDeleted))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeGlobal, metricTypeBytes).Set(float64(data.GlobalBytes))
|
|
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeLocal, metricTypeFiles).Set(float64(data.LocalFiles))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeLocal, metricTypeDirectories).Set(float64(data.LocalDirectories))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeLocal, metricTypeSymlinks).Set(float64(data.LocalSymlinks))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeLocal, metricTypeDeleted).Set(float64(data.LocalDeleted))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeLocal, metricTypeBytes).Set(float64(data.LocalBytes))
|
|
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeNeed, metricTypeFiles).Set(float64(data.NeedFiles))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeNeed, metricTypeDirectories).Set(float64(data.NeedDirectories))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeNeed, metricTypeSymlinks).Set(float64(data.NeedSymlinks))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeNeed, metricTypeDeleted).Set(float64(data.NeedDeletes))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeNeed, metricTypeBytes).Set(float64(data.NeedBytes))
|
|
|
|
for _, devCfg := range c.cfg.Folders()[folder].Devices {
|
|
select {
|
|
case <-ctx.Done():
|
|
return
|
|
default:
|
|
}
|
|
|
|
if devCfg.DeviceID.Equals(c.id) {
|
|
// We already know about ourselves.
|
|
continue
|
|
}
|
|
|
|
// Get completion percentage of this folder for the
|
|
// remote device.
|
|
comp, err := c.model.Completion(devCfg.DeviceID, folder)
|
|
if err != nil {
|
|
l.Debugf("Error getting completion for folder %v, device %v: %v", folder, devCfg.DeviceID, err)
|
|
continue
|
|
}
|
|
ev := comp.Map()
|
|
ev["folder"] = folder
|
|
ev["device"] = devCfg.DeviceID.String()
|
|
c.evLogger.Log(events.FolderCompletion, ev)
|
|
}
|
|
}
|