Files
vikunja/pkg/modules/dump/restore.go
kolaente 2053426062 chore(lint): suppress known gosec false positives
Add config-level exclusions for G117 (secret-named struct fields),
G101 in test files, G702/G704 in magefile, and goheader in plugins.
Add inline #nosec comments for specific G703/G704 false positives
in export, dump/restore, migration, and avatar code.
2026-03-23 16:23:15 +01:00

530 lines
15 KiB
Go

// Vikunja is a to-do list application to facilitate your life.
// Copyright 2018-present Vikunja and contributors. All rights reserved.
//
// This program is free software: you can redistribute it and/or modify
// it under the terms of the GNU Affero General Public License as published by
// the Free Software Foundation, either version 3 of the License, or
// (at your option) any later version.
//
// This program is distributed in the hope that it will be useful,
// but WITHOUT ANY WARRANTY; without even the implied warranty of
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
// GNU Affero General Public License for more details.
//
// You should have received a copy of the GNU Affero General Public License
// along with this program. If not, see <https://www.gnu.org/licenses/>.
package dump
import (
"archive/zip"
"bufio"
"bytes"
"encoding/base64"
"encoding/json"
"errors"
"fmt"
"io"
"os"
"path/filepath"
"sort"
"strconv"
"strings"
"github.com/hashicorp/go-version"
"code.vikunja.io/api/pkg/config"
"code.vikunja.io/api/pkg/db"
"code.vikunja.io/api/pkg/files"
"code.vikunja.io/api/pkg/initialize"
"code.vikunja.io/api/pkg/log"
"code.vikunja.io/api/pkg/migration"
"code.vikunja.io/api/pkg/utils"
vversion "code.vikunja.io/api/pkg/version"
"src.techknowlogick.com/xormigrate"
)
const maxConfigSize = 5 * 1024 * 1024 // 5 MB, should be largely enough
const maxDumpEntrySize = 500 * 1024 * 1024 // 500 MB
// parseDbFileName validates and extracts the table name from a database dump filename.
// Returns the table name and true if valid, or empty string and false if invalid.
func parseDbFileName(fname string) (string, bool) {
if !strings.HasSuffix(fname, ".json") {
return "", false
}
tableName := strings.TrimSuffix(fname, ".json")
if tableName == "" || strings.ContainsAny(tableName, "/\\") {
return "", false
}
return tableName, true
}
// Restore takes a zip file name and restores it
func Restore(filename string, overrideConfig bool) error {
r, err := zip.OpenReader(filename)
if err != nil {
return fmt.Errorf("could not open zip file: %w", err)
}
log.Warning("Restoring a dump will wipe your current installation!")
log.Warning("To confirm, please type 'Yes, I understand' and confirm with enter:")
cr := bufio.NewReader(os.Stdin)
text, err := cr.ReadString('\n')
if err != nil {
return fmt.Errorf("could not read confirmation message: %w", err)
}
if text != "Yes, I understand\n" {
return fmt.Errorf("invalid confirmation message")
}
// Find the configFile, database and files files
var configFile *zip.File
var dotEnvFile *zip.File
var versionFile *zip.File
dbfiles := make(map[string]*zip.File)
filesFiles := make(map[string]*zip.File)
for _, file := range r.File {
if utils.ContainsPathTraversal(file.Name) {
return fmt.Errorf("unsafe path in zip archive: %q", file.Name)
}
if strings.HasPrefix(file.Name, "config") {
configFile = file
continue
}
if strings.HasPrefix(file.Name, "database/") {
fname := strings.TrimPrefix(file.Name, "database/")
tableName, valid := parseDbFileName(fname)
if !valid {
return fmt.Errorf("invalid database file name in zip archive: %q", file.Name)
}
dbfiles[tableName] = file
continue
}
if file.Name == ".env" {
dotEnvFile = file
continue
}
if strings.HasPrefix(file.Name, "files/") {
filesFiles[strings.TrimPrefix(file.Name, "files/")] = file
continue
}
if file.Name == "VERSION" {
versionFile = file
}
}
///////
// Check if we're restoring to the same version as the dump
err = checkVikunjaVersion(versionFile)
if err != nil {
return err
}
///////
// Restore the config file
if overrideConfig {
err = restoreConfig(configFile, dotEnvFile)
if err != nil {
return err
}
} else {
log.Warning("Preserving existing configuration (--preserve-config flag used)")
log.Warning("Configuration preserved - ensure your current config is compatible with the restored data")
}
log.Info("Restoring...")
// Init the configFile again since the restored configuration is most likely different from the one before
initialize.LightInit()
initialize.InitEngines()
err = files.InitFileHandler()
if err != nil {
return fmt.Errorf("could not init file handler: %w", err)
}
///////
// Restore the db
// Validate archive contents before wiping to avoid leaving the database
// in a destroyed state when the archive is malformed.
migrations := dbfiles["migration"]
if migrations == nil {
return fmt.Errorf("dump does not contain database migration information")
}
rc, err := migrations.Open()
if err != nil {
return fmt.Errorf("could not open migrations: %w", err)
}
defer rc.Close()
var buf bytes.Buffer
if _, err := buf.ReadFrom(io.LimitReader(rc, maxDumpEntrySize)); err != nil {
return fmt.Errorf("could not read migrations: %w", err)
}
ms := []*xormigrate.Migration{}
if err := json.Unmarshal(buf.Bytes(), &ms); err != nil {
return fmt.Errorf("could not read migrations: %w", err)
}
sort.Slice(ms, func(i, j int) bool {
return ms[i].ID < ms[j].ID
})
if len(ms) < 2 {
return fmt.Errorf("dump does not contain enough migration information")
}
lastMigration := ms[len(ms)-2]
if err := preValidateTableData(dbfiles); err != nil {
return err
}
// Start by wiping everything - only after we've validated the archive
if err := db.WipeEverything(); err != nil {
return fmt.Errorf("could not wipe database: %w", err)
}
log.Info("Wiped database.")
log.Debugf("Last migration: %s", lastMigration.ID)
if err := migration.MigrateTo(lastMigration.ID, nil); err != nil {
return fmt.Errorf("could not create db structure: %w", err)
}
delete(dbfiles, "migration")
err = restoreTableData(dbfiles)
if err != nil {
return err
}
// Run migrations again to migrate a potentially outdated dump
migration.Migrate(nil)
///////
// Restore Files
for i, file := range filesFiles {
id, err := strconv.ParseInt(i, 10, 64)
if err != nil {
return fmt.Errorf("could not parse file id %s: %w", i, err)
}
if err := restoreFile(id, file); err != nil {
return fmt.Errorf("could not restore file %s: %w", i, err)
}
log.Infof("Restored file %s", i)
}
log.Infof("Restored %d files.", len(filesFiles))
///////
// Done
log.Infof("Done restoring dump.")
if overrideConfig {
log.Infof("Restart Vikunja to make sure the new configuration file is applied.")
}
return nil
}
func restoreFile(id int64, zipFile *zip.File) error {
f := &files.File{ID: id}
fc, err := zipFile.Open()
if err != nil {
return fmt.Errorf("could not open zip entry: %w", err)
}
defer fc.Close()
// Create a temporary file to make the content seekable without loading
// it all into memory. zip.File.Open() returns io.ReadCloser which is not
// seekable, but f.Save requires io.ReadSeeker.
tmpFile, err := os.CreateTemp("", "vikunja-restore-*")
if err != nil {
return fmt.Errorf("could not create temp file: %w", err)
}
defer func() {
_ = tmpFile.Close()
_ = os.Remove(tmpFile.Name()) // #nosec G703 -- path from os.CreateTemp, not user input
}()
// Limit copy size to prevent decompression bombs
maxSize := config.GetMaxFileSizeInMBytes() * 1024 * 1024
written, err := io.CopyN(tmpFile, fc, int64(maxSize)+1) // #nosec G115 -- maxSize is configured, not user input
if err != nil && !errors.Is(err, io.EOF) {
return fmt.Errorf("could not copy to temp file: %w", err)
}
if uint64(written) > maxSize {
return files.ErrFileIsTooLarge{Size: uint64(written)}
}
if _, err := tmpFile.Seek(0, io.SeekStart); err != nil {
return fmt.Errorf("could not seek temp file: %w", err)
}
return f.Save(tmpFile)
}
func convertFieldValue(fieldName string, value interface{}, isFloat bool) (interface{}, error) {
// Check if this is a float field and the value is already a number
if isFloat {
switch v := value.(type) {
case float64:
// Already a float64, no need to process
return v, nil
case int:
// Convert int to float64
return float64(v), nil
case string:
// Try to decode from base64 string and convert to float
decoded, err := base64.StdEncoding.DecodeString(v)
if err != nil {
var corruptErr base64.CorruptInputError
if !errors.As(err, &corruptErr) {
return nil, fmt.Errorf("could not decode field '%s' %s: %w", fieldName, value, err)
}
// If it's a CorruptInputError, treat the string as raw data
decoded = []byte(v)
}
val, err := strconv.ParseFloat(string(decoded), 64)
if err != nil {
return nil, fmt.Errorf("could not parse double value for field '%s': %w", fieldName, err)
}
return val, nil
default:
return nil, fmt.Errorf("unexpected type for float field '%s': %T", fieldName, v)
}
}
// Handle JSON fields (non-float)
switch v := value.(type) {
case string:
// Check if the string is "null" (case insensitive) and return nil for SQL NULL
if strings.ToLower(v) == "null" {
return nil, nil
}
decoded, err := base64.StdEncoding.DecodeString(v)
if err != nil {
var corruptErr base64.CorruptInputError
if !errors.As(err, &corruptErr) {
return nil, fmt.Errorf("could not decode field '%s' %s: %w", fieldName, value, err)
}
// If it's a CorruptInputError, treat the string as raw data
decoded = []byte(v)
}
return string(decoded), nil
default:
return nil, fmt.Errorf("expected string for JSON field '%s', got %T", fieldName, v)
}
}
func restoreTableData(tables map[string]*zip.File) error {
jsonFields := map[string][]string{
"api_tokens": {"permissions"},
"notifications": {"notification"},
"project_views": {"filter", "bucket_configuration"},
"saved_filters": {"filters"},
"users": {"frontend_settings"},
}
floatFields := map[string][]string{
"buckets": {"position"},
"project_views": {"position"},
"projects": {"position"},
"task_positions": {"position"},
}
// Restore all db data
for table, d := range tables {
content, err := unmarshalFileToJSON(d)
if err != nil {
return fmt.Errorf("could not read table %s: %w", table, err)
}
processFields := func(fields []string, isFloat bool) error {
for i := range content {
for _, f := range fields {
if _, hasField := content[i][f]; !hasField {
continue
}
convertedValue, err := convertFieldValue(f, content[i][f], isFloat)
if err != nil {
return err
}
content[i][f] = convertedValue
}
}
return nil
}
// Process JSON fields
if fields, hasJSONFields := jsonFields[table]; hasJSONFields {
if err := processFields(fields, false); err != nil {
return err
}
}
// Process double fields
if fields, hasDoubleFields := floatFields[table]; hasDoubleFields {
if err := processFields(fields, true); err != nil {
return err
}
}
if err := db.Restore(table, content); err != nil {
return fmt.Errorf("could not restore table data for table %s: %w", table, err)
}
log.Infof("Restored table %s", table)
}
log.Infof("Restored %d tables", len(tables))
return nil
}
// preValidateTableData checks that all table data JSON files in the archive
// are parseable before wiping the database, to avoid leaving the database
// in a destroyed state when the archive contains corrupted data.
func preValidateTableData(dbfiles map[string]*zip.File) error {
for table, d := range dbfiles {
if table == "migration" {
continue
}
rc, err := d.Open()
if err != nil {
return fmt.Errorf("could not open table data for %s: %w", table, err)
}
var buf bytes.Buffer
if _, err := buf.ReadFrom(io.LimitReader(rc, maxDumpEntrySize)); err != nil {
rc.Close()
return fmt.Errorf("could not read table data for %s: %w", table, err)
}
rc.Close()
var test []map[string]interface{}
if err := json.Unmarshal(buf.Bytes(), &test); err != nil {
return fmt.Errorf("invalid JSON in table data for %s: %w", table, err)
}
}
return nil
}
func unmarshalFileToJSON(file *zip.File) (contents []map[string]interface{}, err error) {
rc, err := file.Open()
if err != nil {
return
}
defer rc.Close()
var buf bytes.Buffer
if _, err := buf.ReadFrom(io.LimitReader(rc, maxDumpEntrySize)); err != nil {
return nil, err
}
contents = []map[string]interface{}{}
if err := json.Unmarshal(buf.Bytes(), &contents); err != nil {
return nil, err
}
return
}
func restoreConfig(configFile, dotEnvFile *zip.File) error {
if configFile != nil {
if configFile.UncompressedSize64 > maxConfigSize {
return fmt.Errorf("config file too large, is %d, max size is %d", configFile.UncompressedSize64, maxConfigSize)
}
// Use only the base name to prevent writing outside the working directory
sanitizedName := filepath.Base(configFile.Name)
outFile, err := os.OpenFile(sanitizedName, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, configFile.Mode())
if err != nil {
return fmt.Errorf("could not open config file for writing: %w", err)
}
cfgr, err := configFile.Open()
if err != nil {
return err
}
// #nosec - We eliminated the potential decompression bomb by erroring out above if the file is larger than a threshold.
_, err = io.Copy(outFile, cfgr)
if err != nil {
return fmt.Errorf("could not create config file: %w", err)
}
_ = cfgr.Close()
_ = outFile.Close()
log.Infof("The config file has been restored to '%s'.", sanitizedName)
log.Infof("You can now make changes to it, hit enter when you're done.")
if _, err := bufio.NewReader(os.Stdin).ReadString('\n'); err != nil {
return fmt.Errorf("could not read from stdin: %w", err)
}
return nil
}
log.Warning("No config file found, not restoring one.")
log.Warning("You'll likely have had Vikunja configured through environment variables.")
if dotEnvFile != nil {
dotenv, err := dotEnvFile.Open()
if err != nil {
return err
}
buf := bytes.Buffer{}
_, err = buf.ReadFrom(io.LimitReader(dotenv, maxDumpEntrySize))
if err != nil {
return err
}
log.Warningf("Please make sure the following settings are properly configured in your instance:\n%s", buf.String())
log.Warning("Make sure your current config matches the following env variables, confirm by pressing enter when done.")
log.Warning("If your config does not match, you'll have to make the changes and restart the restoring process afterwards.")
if _, err := bufio.NewReader(os.Stdin).ReadString('\n'); err != nil {
return fmt.Errorf("could not read from stdin: %w", err)
}
}
return nil
}
func checkVikunjaVersion(versionFile *zip.File) error {
if versionFile == nil {
return fmt.Errorf("dump does not contain VERSION file, refusing to continue")
}
vf, err := versionFile.Open()
if err != nil {
return fmt.Errorf("could not open version file: %w", err)
}
var bufVersion bytes.Buffer
if _, err := bufVersion.ReadFrom(io.LimitReader(vf, maxDumpEntrySize)); err != nil {
return fmt.Errorf("could not read version file: %w", err)
}
versionString := bufVersion.String()
if versionString == "dev" && vversion.Version == "dev" {
log.Debugf("Importing from dev version")
} else {
dumpedVersion, err := version.NewVersion(bufVersion.String())
if err != nil {
return err
}
currentVersion, err := version.NewVersion(vversion.Version)
if err != nil {
return err
}
if !dumpedVersion.Equal(currentVersion) {
return fmt.Errorf("export was created with version %s but this is %s - please make sure you are running the same Vikunja version before restoring", dumpedVersion, currentVersion)
}
}
return nil
}