fix serve recipe: filter output to Development Server start + [200] requests only

refactor: unify CSS color variables across public and admin

- Replace old variable structure with new standardized naming:
  - Background: --bg-primary, --bg-secondary, --bg-tertiary, --bg-active
  - Text: --text-primary, --text-secondary, --text-tertiary
  - Border: --border-primary, --border-secondary
  - Status: --success, --error, --warning
  - Accent: --accent-primary, --accent-secondary, --accent-foreground, --accent-muted
- Remove admin-specific color variables (--admin-*)
- Update all CSS files to use shared variables:
  - variables.css, common.css, main.css, admin.css
  - tfe.css, search.css, apropos.css, system.css, colors.css
This commit is contained in:
Pontoporeia
2026-04-02 14:11:11 +02:00
parent ba7814c6dc
commit bf2594112b
15 changed files with 536 additions and 1656 deletions

View File

@@ -0,0 +1,167 @@
### User
for a low traffic website, using a csv vs a sqlite or a nosql db, which would make sense ? it would be a php website.
### Assistant
Short answer:**use SQLite**for almost all low-traffic PHP sites unless your data is literally a tiny, read-only list best kept as a CSV. CSVs are fragile; NoSQL is usually overkill and adds operational cost.
Heres a practical breakdown to help you choose.
# When to pick each
**SQLite (recommended)**
- Best for low-traffic sites with structured data (users, posts, settings, relations).
- ACID transactions, SQL queries, indexes, easy backups (single file).
- Very simple to use from PHP via PDO; no separate server to manage.
- Handles many concurrent reads well; writes are serialized but fine for low write volume. Enable WAL for better concurrency.
- Put the`.sqlite`file outside webroot and set tight file permissions.
**CSV**
- Accept if data is: tiny, mostly read-only, and never queried/joined (e.g., a static list of countries).
- Pros: human-readable, editable in Excel.
- Cons: no indexes, no transactions, parsing edge cases, concurrency problems, security issues if stored in webroot.
- If you use CSV writes, you must handle locking and validation yourself.
**NoSQL (MongoDB, CouchDB, etc.)**
- Use only if you already need document DB features (flexible schema, distributed clustering) or will scale horizontally soon.
- Adds server/process to manage, drivers, extra operational complexity — usually overkill for low traffic.
**MySQL/Postgres**
- Good if you expect to scale or need multi-user DB server now; slightly more operational overhead than SQLite but easier to scale later.
# Practical considerations
- **Concurrency:**SQLite serializes writes. For low traffic or mostly-read sites this is fine. Turn on WAL mode (`PRAGMA journal_mode = WAL`) to improve concurrency.
- **Backups:**SQLite = copy the file (use`VACUUM`or use`sqlite3`online backup API). CSV = copy file but risk partial writes. NoSQL = backup routines/replication needed.
- **Security:**Never keep your DB/CSV in webroot. Correct file permissions. Use prepared statements to avoid injection.
- **Querying & indexing:**CSV has none — searching/joins are O(n). SQLite supports indexes and complex queries.
- **Migrations:**SQLite supports schema migration (use a simple migration script). CSV makes schema evolution painful.
- **Tooling:**PHP has native PDO for SQLite. Many tools can open SQLite files. CSV editing via spreadsheets is easy but risky.
# Minimal PHP examples
**SQLite (PDO) — recommended**
```
PHP// open$pdo = new PDO('sqlite:/path/outside_webroot/app.db');$pdo->setAttribute(PDO::ATTR_ERRMODE, PDO::ERRMODE_EXCEPTION);// create$pdo->exec("CREATE TABLE IF NOT EXISTS items (id INTEGER PRIMARY KEY, name TEXT, created_at INTEGER)");// insert (prepared)$stmt = $pdo->prepare('INSERT INTO items (name, created_at) VALUES (:name, :ts)');$stmt->execute([':name' => 'Example', ':ts' => time()]);// queryforeach ($pdo->query('SELECT * FROM items ORDER BY id DESC') as $row) { var_dump($row);}
```
Tip: after creating DB run`PRAGMA journal_mode = WAL;`once to set WAL mode if you expect concurrent reads.
**CSV (with locking) — only for very simple data**
```
PHP$csvFile = '/path/outside_webroot/data.csv';// readif (($h = fopen($csvFile, 'r')) !== false) { while (($row = fgetcsv($h)) !== false) { // process $row array } fclose($h);}// append safelyif (($h = fopen($csvFile, 'a')) !== false) { if (flock($h, LOCK_EX)) { fputcsv($h, ['id', 'value', time()]); fflush($h); flock($h, LOCK_UN); } fclose($h);}
```
# Rule of thumb
- If you need queries, joins, indexes, transactions →**SQLite**.
- If its a single tiny static table you occasionally open in Excel →**CSV**.
- If you need scaling, replication, distributed features, or already use NoSQL → consider**NoSQL**(but expect extra ops work).
If you want, I can:
- give a small migration plan from CSV → SQLite, or
- provide a tiny PDO wrapper and migration scripts for your PHP project.
Which would you like next?
### User
this is basically to make a libary for theses of a school. how could I implement a export import of the sqlite to csv ?
### Assistant
Nice — good fit for SQLite. Below are ready-to-use options (PHP functions + CLI commands) to**export**and**import**CSVs safely and efficiently for your school theses library. I include streaming (no huge memory use), header handling, transactions, column mapping, basic validation, and tips for Excel compatibility.
# Quick notes before code
- Put DB file outside webroot and set tight permissions.
- Always back up DB before imports.
- Use transactions for imports for speed and atomicity.
- For Excel, prepend a UTF-8 BOM (`\xEF\xBB\xBF`) when exporting so Excel recognizes UTF-8.
- For large imports, process in batches and use prepared statements.
- Validate rows (required columns, types, length) to avoid corrupting DB.
# 1) Export a table to CSV (streaming, low memory)
```
PHP<?phpfunction exportTableToCsv(PDO $pdo, string $table, string $csvPath, bool $includeHeader = true, string $bom = "\xEF\xBB\xBF") { $stmt = $pdo->prepare("SELECT * FROM \"$table\""); $stmt->execute(); $out = fopen($csvPath, 'w'); if ($out === false) { throw new RuntimeException("Cannot open CSV file for writing: $csvPath"); } // Write BOM for Excel/Windows compatibility (optional) if ($bom !== '') { fwrite($out, $bom); } // Write header if ($includeHeader) { $columnCount = $stmt->columnCount(); $headers = []; for ($i = 0; $i < $columnCount; $i++) { $meta = $stmt->getColumnMeta($i); $headers[] = $meta['name'] ?? "col$i"; } fputcsv($out, $headers); } // Stream rows while ($row = $stmt->fetch(PDO::FETCH_ASSOC)) { fputcsv($out, array_values($row)); } fclose($out);}
```
Usage:
```
PHP$pdo = new PDO('sqlite:/path/outside_webroot/library.db');$pdo->setAttribute(PDO::ATTR_ERRMODE, PDO::ERRMODE_EXCEPTION);exportTableToCsv($pdo, 'theses', '/tmp/theses_export.csv');
```
# 2) Export an arbitrary query to CSV
```
PHPfunction exportQueryToCsv(PDO $pdo, string $sql, array $params, string $csvPath, bool $includeHeader = true, string $bom = "\xEF\xBB\xBF") { $stmt = $pdo->prepare($sql); $stmt->execute($params); $out = fopen($csvPath, 'w'); if ($out === false) throw new RuntimeException("Cannot open CSV file: $csvPath"); if ($bom !== '') fwrite($out, $bom); if ($includeHeader) { $columnCount = $stmt->columnCount(); $headers = []; for ($i = 0; $i < $columnCount; $i++) { $meta = $stmt->getColumnMeta($i); $headers[] = $meta['name'] ?? "col$i"; } fputcsv($out, $headers); } while ($row = $stmt->fetch(PDO::FETCH_NUM)) { fputcsv($out, $row); } fclose($out);}
```
# 3) Import CSV into a table (stream + prepared insert, with header mapping)
```
PHP<?phpfunction importCsvToTable(PDO $pdo, string $csvPath, string $table, array $columnMap = null, int $batchSize = 500, string $encoding = 'UTF-8') { if (!file_exists($csvPath)) { throw new RuntimeException("CSV file not found: $csvPath"); } $h = fopen($csvPath, 'r'); if ($h === false) throw new RuntimeException("Cannot open CSV: $csvPath"); // Read header row $header = fgetcsv($h); if ($header === false) { fclose($h); throw new RuntimeException("CSV is empty or unreadable"); } // Optional: strip BOM from first header cell $header[0] = preg_replace('/^\x{FEFF}/u', '', $header[0]); // Map incoming CSV columns to DB columns. // $columnMap can be: null (assume header names match table columns) // or an associative array: ['CSV Column Name' => 'db_column'] $map = []; if ($columnMap === null) { foreach ($header as $col) $map[$col] = $col; } else { foreach ($header as $col) { if (isset($columnMap[$col])) $map[$col] = $columnMap[$col]; else $map[$col] = null; // skip if not mapped } } // Determine DB columns we will insert (in order) $dbCols = array_values(array_filter($map, fn($v) => $v !== null)); if (count($dbCols) === 0) { fclose($h); throw new RuntimeException("No CSV columns map to DB columns."); } // Prepare insert statement $placeholders = implode(',', array_fill(0, count($dbCols), '?')); $colList = implode(',', array_map(fn($c) => "\"$c\"", $dbCols)); $insertSql = "INSERT INTO \"$table\" ($colList) VALUES ($placeholders)"; $insertStmt = $pdo->prepare($insertSql); // Read rows in batches inside a transaction $rowCount = 0; $batch = []; $pdo->beginTransaction(); try { while (($row = fgetcsv($h)) !== false) { $assoc = array_combine($header, $row); // Build values in dbCols order; do simple validation (you can extend) $values = []; foreach ($map as $csvCol => $dbCol) { if ($dbCol === null) continue; } foreach ($dbCols as $dbCol) { // find the CSV column that maps to this dbCol (reverse lookup) $csvKey = array_search($dbCol, $map, true); $values[] = $assoc[$csvKey] ?? null; } // Optional: basic validation example (skip blank required rows) $allEmpty = true; foreach ($values as $v) { if (trim((string)$v) !== '') { $allEmpty = false; break; } } if ($allEmpty) continue; $insertStmt->execute($values); $rowCount++; if ($rowCount % $batchSize === 0) { $pdo->commit(); $pdo->beginTransaction(); } } $pdo->commit(); } catch (Exception $e) { $pdo->rollBack(); fclose($h); throw $e; } fclose($h); return $rowCount;}
```
Usage with column mapping:
```
PHP$colMap = [ 'Title' => 'title', 'Author' => 'author', 'Year' => 'year', 'Supervisor' => 'supervisor'];$count = importCsvToTable($pdo, '/tmp/theses_import.csv', 'theses', $colMap);echo "Imported $count rows\n";
```
# 4) Tips & extras
**Detect table columns automatically**
```
PHPfunction getTableColumns(PDO $pdo, string $table): array { $cols = []; $stmt = $pdo->query("PRAGMA table_info(\"$table\")"); foreach ($stmt->fetchAll(PDO::FETCH_ASSOC) as $col) { $cols[] = $col['name']; } return $cols;}
```
**Speed optimizations for large imports**
- Wrap inserts in a transaction (done above).
- Turn off synchronous temporarily for import:`PRAGMA synchronous = OFF;`then set back to`NORMAL`or`FULL`afterwards. (Be careful — power loss can cause corruption during import.)
- Use`PRAGMA journal_mode = WAL;`for better concurrency.
**CSV dialect**
- Use`fgetcsv`/`fputcsv`which handle quoting/commas.
- If you must support Excel-generated CSVs, handle`\r\n`line endings and possibly semicolon delimiters (region dependent).
**Conflict handling**
- If`theses`table has unique constraints (e.g.,`unique(identifier)`), you can use`INSERT OR REPLACE`or`INSERT OR IGNORE`depending on desired behavior:```
SQLINSERT OR REPLACE INTO "theses" (col1, col2) VALUES (?, ?)
```
**CSV created by CLI (SQLite)**
- If you have shell access, you can quickly export:```
Bashsqlite3 library.db -header -csv "SELECT * FROM theses;" > theses.csv
```
- Import via sqlite3 CLI (beware: requires exact columns & table prepared):```
Bashsqlite3 library.db.mode csv.import /path/to/theses.csv theses
```
# 5) Example end-to-end flow for your theses library
- Create`theses`schema in SQLite with fields:`id, title, author, year, department, supervisor, abstract, filepath, created_at`.
- Provide an admin page to**Export**: call`exportTableToCsv()`and stream the CSV to browser with correct headers and BOM.
- Provide an admin**Import**form (file upload). On upload:
- Save uploaded CSV to a temp folder outside webroot.
- Make a backup copy of DB.
- Run`importCsvToTable()`with a column map (allow admins to map CSV columns to DB fields if header names differ).
- Show import summary: rows imported, rows skipped, errors.
- For repeatable imports, keep a small import-log table to record timestamp, filename, rows imported, and admin user.
If you want, I can:
- produce a full single-file PHP admin script (upload UI + export button + import with mapping UI + summary) tailored to your`theses`schema, or
- give a tiny migration script to convert an existing CSV folder into the`theses`SQLite DB.
Which would you like me to generate now?