mirror of
https://github.com/marcogll/TaxHacker_s23.git
synced 2026-01-13 13:25:18 +00:00
feat: backup + restore
This commit is contained in:
3
.gitignore
vendored
3
.gitignore
vendored
@@ -45,7 +45,8 @@ yarn-error.log*
|
||||
*.tsbuildinfo
|
||||
next-env.d.ts
|
||||
|
||||
# sqlite
|
||||
# databases
|
||||
pgdata
|
||||
*.db
|
||||
*.sqlite
|
||||
*.sqlite3
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
import { ExportFields, ExportFilters, exportImportFieldsMapping } from "@/models/export_and_import"
|
||||
import { EXPORT_AND_IMPORT_FIELD_MAP, ExportFields, ExportFilters } from "@/models/export_and_import"
|
||||
import { getFields } from "@/models/fields"
|
||||
import { getFilesByTransactionId } from "@/models/files"
|
||||
import { getTransactions } from "@/models/transactions"
|
||||
@@ -38,7 +38,7 @@ export async function GET(request: Request) {
|
||||
const row: Record<string, any> = {}
|
||||
for (const key of fieldKeys) {
|
||||
const value = transaction[key as keyof typeof transaction] ?? ""
|
||||
const exportFieldSettings = exportImportFieldsMapping[key]
|
||||
const exportFieldSettings = EXPORT_AND_IMPORT_FIELD_MAP[key]
|
||||
if (exportFieldSettings && exportFieldSettings.export) {
|
||||
row[key] = await exportFieldSettings.export(value)
|
||||
} else {
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
"use server"
|
||||
|
||||
import { exportImportFieldsMapping } from "@/models/export_and_import"
|
||||
import { EXPORT_AND_IMPORT_FIELD_MAP } from "@/models/export_and_import"
|
||||
import { createTransaction } from "@/models/transactions"
|
||||
import { parse } from "@fast-csv/parse"
|
||||
import { revalidatePath } from "next/cache"
|
||||
@@ -44,7 +44,7 @@ export async function saveTransactionsAction(prevState: any, formData: FormData)
|
||||
for (const row of rows) {
|
||||
const transactionData: Record<string, unknown> = {}
|
||||
for (const [fieldCode, value] of Object.entries(row)) {
|
||||
const fieldDef = exportImportFieldsMapping[fieldCode]
|
||||
const fieldDef = EXPORT_AND_IMPORT_FIELD_MAP[fieldCode]
|
||||
if (fieldDef?.import) {
|
||||
transactionData[fieldCode] = await fieldDef.import(value as string)
|
||||
} else {
|
||||
|
||||
@@ -1,21 +1,214 @@
|
||||
"use server"
|
||||
|
||||
import { DATABASE_FILE } from "@/lib/db"
|
||||
import { prisma } from "@/lib/db"
|
||||
import { FILE_UPLOAD_PATH } from "@/lib/files"
|
||||
import { MODEL_BACKUP } from "@/models/backups"
|
||||
import fs from "fs"
|
||||
import { mkdir } from "fs/promises"
|
||||
import JSZip from "jszip"
|
||||
import path from "path"
|
||||
|
||||
const SUPPORTED_BACKUP_VERSIONS = ["1.0"]
|
||||
|
||||
export async function restoreBackupAction(prevState: any, formData: FormData) {
|
||||
const file = formData.get("file") as File
|
||||
const removeExistingData = formData.get("removeExistingData") === "true"
|
||||
|
||||
if (!file) {
|
||||
return { success: false, error: "No file provided" }
|
||||
}
|
||||
|
||||
// Restore tables
|
||||
try {
|
||||
const fileBuffer = await file.arrayBuffer()
|
||||
const fileData = Buffer.from(fileBuffer)
|
||||
fs.writeFileSync(DATABASE_FILE, fileData)
|
||||
const zip = await JSZip.loadAsync(fileData)
|
||||
|
||||
// Check backup version
|
||||
const metadataFile = zip.file("data/metadata.json")
|
||||
if (metadataFile) {
|
||||
const metadataContent = await metadataFile.async("string")
|
||||
try {
|
||||
const metadata = JSON.parse(metadataContent)
|
||||
if (!metadata.version || !SUPPORTED_BACKUP_VERSIONS.includes(metadata.version)) {
|
||||
return {
|
||||
success: false,
|
||||
error: `Incompatible backup version: ${
|
||||
metadata.version || "unknown"
|
||||
}. Supported versions: ${SUPPORTED_BACKUP_VERSIONS.join(", ")}`,
|
||||
}
|
||||
}
|
||||
console.log(`Restoring backup version ${metadata.version} created at ${metadata.timestamp}`)
|
||||
} catch (error) {
|
||||
console.warn("Could not parse backup metadata:", error)
|
||||
}
|
||||
} else {
|
||||
console.warn("No metadata found in backup, assuming legacy format")
|
||||
}
|
||||
|
||||
if (removeExistingData) {
|
||||
await clearAllTables()
|
||||
}
|
||||
|
||||
for (const { filename, model, idField } of MODEL_BACKUP) {
|
||||
try {
|
||||
const jsonFile = zip.file(`data/${filename}`)
|
||||
if (jsonFile) {
|
||||
const jsonContent = await jsonFile.async("string")
|
||||
const restoredCount = await restoreModelFromJSON(model, jsonContent, idField)
|
||||
console.log(`Restored ${restoredCount} records from ${filename}`)
|
||||
}
|
||||
} catch (error) {
|
||||
console.error(`Error restoring model from ${filename}:`, error)
|
||||
}
|
||||
}
|
||||
|
||||
// Restore files
|
||||
try {
|
||||
const filesToRestore = Object.keys(zip.files).filter(
|
||||
(filename) => filename.startsWith("data/uploads/") && !filename.endsWith("/")
|
||||
)
|
||||
|
||||
if (filesToRestore.length > 0) {
|
||||
await mkdir(FILE_UPLOAD_PATH, { recursive: true })
|
||||
|
||||
// Extract and save each file
|
||||
let restoredFilesCount = 0
|
||||
for (const zipFilePath of filesToRestore) {
|
||||
const file = zip.file(zipFilePath)
|
||||
if (file) {
|
||||
const relativeFilePath = zipFilePath.replace("data/uploads/", "")
|
||||
const fileContent = await file.async("nodebuffer")
|
||||
|
||||
const filePath = path.join(FILE_UPLOAD_PATH, relativeFilePath)
|
||||
const fileName = path.basename(filePath)
|
||||
const fileId = path.basename(fileName, path.extname(fileName))
|
||||
const fileDir = path.dirname(filePath)
|
||||
await mkdir(fileDir, { recursive: true })
|
||||
|
||||
// Write the file
|
||||
fs.writeFileSync(filePath, fileContent)
|
||||
restoredFilesCount++
|
||||
|
||||
// Update the file record
|
||||
await prisma.file.upsert({
|
||||
where: { id: fileId },
|
||||
update: {
|
||||
path: filePath,
|
||||
},
|
||||
create: {
|
||||
id: relativeFilePath,
|
||||
path: filePath,
|
||||
filename: fileName,
|
||||
mimetype: "application/octet-stream",
|
||||
},
|
||||
})
|
||||
}
|
||||
}
|
||||
}
|
||||
} catch (error) {
|
||||
console.error("Error restoring uploaded files:", error)
|
||||
return {
|
||||
success: false,
|
||||
error: `Error restoring uploaded files: ${error instanceof Error ? error.message : String(error)}`,
|
||||
}
|
||||
}
|
||||
|
||||
return { success: true, message: `Restore completed successfully` }
|
||||
} catch (error) {
|
||||
return { success: false, error: "Failed to restore backup" }
|
||||
console.error("Error restoring from backup:", error)
|
||||
return {
|
||||
success: false,
|
||||
error: `Error restoring from backup: ${error instanceof Error ? error.message : String(error)}`,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
async function clearAllTables() {
|
||||
// Delete in reverse order to handle foreign key constraints
|
||||
for (const { model } of [...MODEL_BACKUP].reverse()) {
|
||||
try {
|
||||
await model.deleteMany({})
|
||||
} catch (error) {
|
||||
console.error(`Error clearing table:`, error)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
async function restoreModelFromJSON(model: any, jsonContent: string, idField: string): Promise<number> {
|
||||
if (!jsonContent) return 0
|
||||
|
||||
try {
|
||||
const records = JSON.parse(jsonContent)
|
||||
|
||||
if (!records || records.length === 0) {
|
||||
return 0
|
||||
}
|
||||
|
||||
let insertedCount = 0
|
||||
for (const rawRecord of records) {
|
||||
const record = processRowData(rawRecord)
|
||||
|
||||
try {
|
||||
// Skip records that don't have the required ID field
|
||||
if (record[idField] === undefined) {
|
||||
console.warn(`Skipping record missing required ID field '${idField}'`)
|
||||
continue
|
||||
}
|
||||
|
||||
await model.upsert({
|
||||
where: { [idField]: record[idField] },
|
||||
update: record,
|
||||
create: record,
|
||||
})
|
||||
insertedCount++
|
||||
} catch (error) {
|
||||
console.error(`Error upserting record:`, error)
|
||||
}
|
||||
}
|
||||
|
||||
return insertedCount
|
||||
} catch (error) {
|
||||
console.error(`Error parsing JSON content:`, error)
|
||||
return 0
|
||||
}
|
||||
}
|
||||
|
||||
function processRowData(row: Record<string, any>): Record<string, any> {
|
||||
const processedRow: Record<string, any> = {}
|
||||
|
||||
for (const [key, value] of Object.entries(row)) {
|
||||
if (value === "" || value === "null" || value === undefined) {
|
||||
processedRow[key] = null
|
||||
continue
|
||||
}
|
||||
|
||||
// Try to parse JSON for object fields
|
||||
if (typeof value === "string" && (value.startsWith("{") || value.startsWith("["))) {
|
||||
try {
|
||||
processedRow[key] = JSON.parse(value)
|
||||
continue
|
||||
} catch (e) {
|
||||
// Not valid JSON, continue with normal processing
|
||||
}
|
||||
}
|
||||
|
||||
// Handle dates (checking for ISO date format)
|
||||
if (typeof value === "string" && /^\d{4}-\d{2}-\d{2}T\d{2}:\d{2}:\d{2}(\.\d{3})?Z?$/.test(value)) {
|
||||
processedRow[key] = new Date(value)
|
||||
continue
|
||||
}
|
||||
|
||||
// Handle numbers
|
||||
if (typeof value === "string" && !isNaN(Number(value)) && key !== "id" && !key.endsWith("Code")) {
|
||||
// Convert numbers but preserving string IDs
|
||||
processedRow[key] = Number(value)
|
||||
continue
|
||||
}
|
||||
|
||||
// Default: keep as is
|
||||
processedRow[key] = value
|
||||
}
|
||||
|
||||
return { success: true }
|
||||
return processedRow
|
||||
}
|
||||
|
||||
@@ -1,10 +1,13 @@
|
||||
import { DATABASE_FILE } from "@/lib/db"
|
||||
import { FILE_UPLOAD_PATH } from "@/lib/files"
|
||||
import { MODEL_BACKUP } from "@/models/backups"
|
||||
import fs, { readdirSync } from "fs"
|
||||
import JSZip from "jszip"
|
||||
import { NextResponse } from "next/server"
|
||||
import path from "path"
|
||||
|
||||
const MAX_FILE_SIZE = 64 * 1024 * 1024 // 64MB
|
||||
const BACKUP_VERSION = "1.0"
|
||||
|
||||
export async function GET(request: Request) {
|
||||
try {
|
||||
const zip = new JSZip()
|
||||
@@ -14,8 +17,29 @@ export async function GET(request: Request) {
|
||||
return new NextResponse("Internal Server Error", { status: 500 })
|
||||
}
|
||||
|
||||
const databaseFile = fs.readFileSync(DATABASE_FILE)
|
||||
rootFolder.file("database.sqlite", databaseFile)
|
||||
// Add metadata with version information
|
||||
rootFolder.file(
|
||||
"metadata.json",
|
||||
JSON.stringify(
|
||||
{
|
||||
version: BACKUP_VERSION,
|
||||
timestamp: new Date().toISOString(),
|
||||
models: MODEL_BACKUP.map((m) => m.filename),
|
||||
},
|
||||
null,
|
||||
2
|
||||
)
|
||||
)
|
||||
|
||||
// Backup models
|
||||
for (const { filename, model } of MODEL_BACKUP) {
|
||||
try {
|
||||
const jsonContent = await tableToJSON(model)
|
||||
rootFolder.file(filename, jsonContent)
|
||||
} catch (error) {
|
||||
console.error(`Error exporting table ${filename}:`, error)
|
||||
}
|
||||
}
|
||||
|
||||
const uploadsFolder = rootFolder.folder("uploads")
|
||||
if (!uploadsFolder) {
|
||||
@@ -25,7 +49,23 @@ export async function GET(request: Request) {
|
||||
|
||||
const uploadedFiles = getAllFilePaths(FILE_UPLOAD_PATH)
|
||||
uploadedFiles.forEach((file) => {
|
||||
uploadsFolder.file(file.replace(FILE_UPLOAD_PATH, ""), fs.readFileSync(file))
|
||||
try {
|
||||
// Check file size before reading
|
||||
const stats = fs.statSync(file)
|
||||
if (stats.size > MAX_FILE_SIZE) {
|
||||
console.warn(
|
||||
`Skipping large file ${file} (${Math.round(stats.size / 1024 / 1024)}MB > ${
|
||||
MAX_FILE_SIZE / 1024 / 1024
|
||||
}MB limit)`
|
||||
)
|
||||
return
|
||||
}
|
||||
|
||||
const fileContent = fs.readFileSync(file)
|
||||
uploadsFolder.file(file.replace(FILE_UPLOAD_PATH, ""), fileContent)
|
||||
} catch (error) {
|
||||
console.error(`Error reading file ${file}:`, error)
|
||||
}
|
||||
})
|
||||
const archive = await zip.generateAsync({ type: "blob" })
|
||||
|
||||
@@ -60,3 +100,13 @@ function getAllFilePaths(dirPath: string): string[] {
|
||||
readDirectory(dirPath)
|
||||
return filePaths
|
||||
}
|
||||
|
||||
async function tableToJSON(model: any): Promise<string> {
|
||||
const data = await model.findMany()
|
||||
|
||||
if (!data || data.length === 0) {
|
||||
return "[]"
|
||||
}
|
||||
|
||||
return JSON.stringify(data, null, 2)
|
||||
}
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
import { FormError } from "@/components/forms/error"
|
||||
import { Button } from "@/components/ui/button"
|
||||
import { Card } from "@/components/ui/card"
|
||||
import { Download } from "lucide-react"
|
||||
import { Download, Loader2 } from "lucide-react"
|
||||
import Link from "next/link"
|
||||
import { useActionState } from "react"
|
||||
import { restoreBackupAction } from "./actions"
|
||||
@@ -18,38 +18,48 @@ export default function BackupSettingsPage() {
|
||||
<div className="flex flex-row gap-4">
|
||||
<Link href="/settings/backups/data">
|
||||
<Button>
|
||||
<Download /> Download data directory
|
||||
<Download /> Download Data Archive
|
||||
</Button>
|
||||
</Link>
|
||||
</div>
|
||||
<div className="text-sm text-muted-foreground max-w-xl">
|
||||
The archive consists of all uploaded files and the SQLite database. You can view the contents of the database
|
||||
using any SQLite viewer.
|
||||
Inside the archive you will find all the uploaded files, as well as JSON files for transactions, categories,
|
||||
projects, fields, currencies, and settings. You can view, edit or migrate your data to another service.
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<Card className="flex flex-col gap-4 mt-16 p-5 bg-red-100 max-w-xl">
|
||||
<h2 className="text-xl font-semibold">How to restore from a backup</h2>
|
||||
<div className="text-md">
|
||||
This feature doesn't work automatically yet. Use your docker deployment with backup archive to manually put
|
||||
database.sqlite and uploaded files into the paths specified in DATABASE_URL and UPLOAD_PATH
|
||||
</div>
|
||||
{/* <form action={restoreBackup}>
|
||||
<label>
|
||||
<input type="file" name="file" />
|
||||
</label>
|
||||
<Button type="submit" variant="destructive" disabled={restorePending}>
|
||||
{restorePending ? (
|
||||
<>
|
||||
<Loader2 className="animate-spin" /> Uploading new database...
|
||||
</>
|
||||
) : (
|
||||
"Restore"
|
||||
)}
|
||||
</Button>
|
||||
</form> */}
|
||||
<Card className="flex flex-col gap-2 mt-16 p-5 bg-red-100 max-w-xl">
|
||||
<h2 className="text-xl font-semibold">Restore from a backup</h2>
|
||||
<p className="text-sm text-muted-foreground">
|
||||
⚠️ This action will delete all existing data from your current database and remove all uploaded files. Be
|
||||
careful and make a backup first!
|
||||
</p>
|
||||
<form action={restoreBackup}>
|
||||
<div className="flex flex-col gap-4 pt-4">
|
||||
<input type="hidden" name="removeExistingData" value="true" />
|
||||
<label>
|
||||
<input type="file" name="file" />
|
||||
</label>
|
||||
<Button type="submit" variant="destructive" disabled={restorePending}>
|
||||
{restorePending ? (
|
||||
<>
|
||||
<Loader2 className="animate-spin" /> Restoring from backup...
|
||||
</>
|
||||
) : (
|
||||
"Delete existing data and restore from backup"
|
||||
)}
|
||||
</Button>
|
||||
</div>
|
||||
</form>
|
||||
{restoreState?.error && <FormError>{restoreState.error}</FormError>}
|
||||
</Card>
|
||||
|
||||
{restoreState?.success && (
|
||||
<Card className="flex flex-col gap-2 p-5 bg-green-100 max-w-xl">
|
||||
<h2 className="text-xl font-semibold">Backup restored successfully</h2>
|
||||
<p className="text-sm text-muted-foreground">You can now continue using the app.</p>
|
||||
</Card>
|
||||
)}
|
||||
</div>
|
||||
)
|
||||
}
|
||||
|
||||
@@ -1,23 +0,0 @@
|
||||
import { NextResponse } from "next/server"
|
||||
|
||||
export async function POST(request: Request) {
|
||||
try {
|
||||
const formData = await request.formData()
|
||||
const file = formData.get("file") as File
|
||||
|
||||
if (!file) {
|
||||
return new NextResponse("No file provided", { status: 400 })
|
||||
}
|
||||
|
||||
const fileBuffer = await file.arrayBuffer()
|
||||
const fileData = Buffer.from(fileBuffer)
|
||||
|
||||
// TODO: Implement restore
|
||||
// fs.writeFileSync(DATABASE_FILE, fileData)
|
||||
|
||||
return new NextResponse("File restored", { status: 200 })
|
||||
} catch (error) {
|
||||
console.error("Error restoring from backup:", error)
|
||||
return new NextResponse("Internal Server Error", { status: 500 })
|
||||
}
|
||||
}
|
||||
46
models/backups.ts
Normal file
46
models/backups.ts
Normal file
@@ -0,0 +1,46 @@
|
||||
import { prisma } from "@/lib/db"
|
||||
|
||||
type ModelEntry = {
|
||||
filename: string
|
||||
model: any
|
||||
idField: string
|
||||
}
|
||||
|
||||
// Ordering is important here
|
||||
export const MODEL_BACKUP: ModelEntry[] = [
|
||||
{
|
||||
filename: "settings.json",
|
||||
model: prisma.setting,
|
||||
idField: "code",
|
||||
},
|
||||
{
|
||||
filename: "currencies.json",
|
||||
model: prisma.currency,
|
||||
idField: "code",
|
||||
},
|
||||
{
|
||||
filename: "categories.json",
|
||||
model: prisma.category,
|
||||
idField: "code",
|
||||
},
|
||||
{
|
||||
filename: "projects.json",
|
||||
model: prisma.project,
|
||||
idField: "code",
|
||||
},
|
||||
{
|
||||
filename: "fields.json",
|
||||
model: prisma.field,
|
||||
idField: "code",
|
||||
},
|
||||
{
|
||||
filename: "files.json",
|
||||
model: prisma.file,
|
||||
idField: "id",
|
||||
},
|
||||
{
|
||||
filename: "transactions.json",
|
||||
model: prisma.transaction,
|
||||
idField: "id",
|
||||
},
|
||||
]
|
||||
@@ -16,7 +16,7 @@ export type ExportImportFieldSettings = {
|
||||
import?: (value: any) => Promise<any>
|
||||
}
|
||||
|
||||
export const exportImportFieldsMapping: Record<string, ExportImportFieldSettings> = {
|
||||
export const EXPORT_AND_IMPORT_FIELD_MAP: Record<string, ExportImportFieldSettings> = {
|
||||
name: {
|
||||
code: "name",
|
||||
type: "string",
|
||||
|
||||
@@ -6,7 +6,7 @@ const nextConfig: NextConfig = {
|
||||
},
|
||||
experimental: {
|
||||
serverActions: {
|
||||
bodySizeLimit: "100mb",
|
||||
bodySizeLimit: "64mb",
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "taxhacker",
|
||||
"version": "0.2.1",
|
||||
"version": "0.3.0",
|
||||
"private": true,
|
||||
"type": "module",
|
||||
"scripts": {
|
||||
|
||||
Reference in New Issue
Block a user