mirror of
https://github.com/community-scripts/ProxmoxVE.git
synced 2026-02-28 07:55:55 +01:00
Compare commits
29 Commits
fix/node-d
...
adgu_fix
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
40aa06940c | ||
|
|
117786376a | ||
|
|
c5a635cdd7 | ||
|
|
165e3f22cd | ||
|
|
2561a50d05 | ||
|
|
6db5479b26 | ||
|
|
e74ddff49a | ||
|
|
80132b0332 | ||
|
|
83a19adbb4 | ||
|
|
6b196a7c81 | ||
|
|
30082a1ba7 | ||
|
|
7741caa6ba | ||
|
|
a3841d3cef | ||
|
|
89cdabd040 | ||
|
|
cbb82812b2 | ||
|
|
1c463369c7 | ||
|
|
1a1e707db3 | ||
|
|
95a63e66a8 | ||
|
|
b09f2db2a9 | ||
|
|
340d999b2b | ||
|
|
8c9874c4aa | ||
|
|
e9b98c2ab0 | ||
|
|
dd46dd2d87 | ||
|
|
719761de6c | ||
|
|
765cd6d123 | ||
|
|
26915e3979 | ||
|
|
34343bc61c | ||
|
|
7484c58175 | ||
|
|
dd4a15fff1 |
35
CHANGELOG.md
35
CHANGELOG.md
@@ -407,10 +407,31 @@ Exercise vigilance regarding copycat or coat-tailing sites that seek to exploit
|
|||||||
|
|
||||||
</details>
|
</details>
|
||||||
|
|
||||||
## 2026-02-25
|
## 2026-02-26
|
||||||
|
|
||||||
### 🚀 Updated Scripts
|
### 🚀 Updated Scripts
|
||||||
|
|
||||||
|
- #### ✨ New Features
|
||||||
|
|
||||||
|
- [QOL] Immich: add warning regarding library compilation time [@vhsdream](https://github.com/vhsdream) ([#12345](https://github.com/community-scripts/ProxmoxVE/pull/12345))
|
||||||
|
|
||||||
|
## 2026-02-25
|
||||||
|
|
||||||
|
### 🆕 New Scripts
|
||||||
|
|
||||||
|
- Zerobyte ([#12321](https://github.com/community-scripts/ProxmoxVE/pull/12321))
|
||||||
|
|
||||||
|
### 🚀 Updated Scripts
|
||||||
|
|
||||||
|
- #### 🐞 Bug Fixes
|
||||||
|
|
||||||
|
- fix: overseer migration [@CrazyWolf13](https://github.com/CrazyWolf13) ([#12340](https://github.com/community-scripts/ProxmoxVE/pull/12340))
|
||||||
|
- add: vikunja: daemon reload [@CrazyWolf13](https://github.com/CrazyWolf13) ([#12323](https://github.com/community-scripts/ProxmoxVE/pull/12323))
|
||||||
|
- opnsense-VM: Use ip link to verify bridge existence [@MickLesk](https://github.com/MickLesk) ([#12329](https://github.com/community-scripts/ProxmoxVE/pull/12329))
|
||||||
|
- wger: Use $http_host for proxy Host header [@MickLesk](https://github.com/MickLesk) ([#12327](https://github.com/community-scripts/ProxmoxVE/pull/12327))
|
||||||
|
- Passbolt: Update Nginx config `client_max_body_size` [@tremor021](https://github.com/tremor021) ([#12313](https://github.com/community-scripts/ProxmoxVE/pull/12313))
|
||||||
|
- Zammad: configure Elasticsearch before zammad start [@MickLesk](https://github.com/MickLesk) ([#12308](https://github.com/community-scripts/ProxmoxVE/pull/12308))
|
||||||
|
|
||||||
- #### 🔧 Refactor
|
- #### 🔧 Refactor
|
||||||
|
|
||||||
- OpenProject: Various fixes [@tremor021](https://github.com/tremor021) ([#12246](https://github.com/community-scripts/ProxmoxVE/pull/12246))
|
- OpenProject: Various fixes [@tremor021](https://github.com/tremor021) ([#12246](https://github.com/community-scripts/ProxmoxVE/pull/12246))
|
||||||
@@ -421,6 +442,18 @@ Exercise vigilance regarding copycat or coat-tailing sites that seek to exploit
|
|||||||
|
|
||||||
- Fix detection of ssh keys [@1-tempest](https://github.com/1-tempest) ([#12230](https://github.com/community-scripts/ProxmoxVE/pull/12230))
|
- Fix detection of ssh keys [@1-tempest](https://github.com/1-tempest) ([#12230](https://github.com/community-scripts/ProxmoxVE/pull/12230))
|
||||||
|
|
||||||
|
- #### ✨ New Features
|
||||||
|
|
||||||
|
- tools.func: Improve GitHub/Codeberg API error handling and error output [@MickLesk](https://github.com/MickLesk) ([#12330](https://github.com/community-scripts/ProxmoxVE/pull/12330))
|
||||||
|
|
||||||
|
- #### 🔧 Refactor
|
||||||
|
|
||||||
|
- core: remove duplicate traps, consolidate error handling and harden signal traps [@MickLesk](https://github.com/MickLesk) ([#12316](https://github.com/community-scripts/ProxmoxVE/pull/12316))
|
||||||
|
|
||||||
|
### 📂 Github
|
||||||
|
|
||||||
|
- github: improvements for node drift wf [@MickLesk](https://github.com/MickLesk) ([#12309](https://github.com/community-scripts/ProxmoxVE/pull/12309))
|
||||||
|
|
||||||
## 2026-02-24
|
## 2026-02-24
|
||||||
|
|
||||||
### 🚀 Updated Scripts
|
### 🚀 Updated Scripts
|
||||||
|
|||||||
6
ct/headers/zerobyte
Normal file
6
ct/headers/zerobyte
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
_____ __ __
|
||||||
|
/__ / ___ _________ / /_ __ __/ /____
|
||||||
|
/ / / _ \/ ___/ __ \/ __ \/ / / / __/ _ \
|
||||||
|
/ /__/ __/ / / /_/ / /_/ / /_/ / /_/ __/
|
||||||
|
/____/\___/_/ \____/_.___/\__, /\__/\___/
|
||||||
|
/____/
|
||||||
@@ -97,7 +97,7 @@ EOF
|
|||||||
if [[ -f ~/.immich_library_revisions ]]; then
|
if [[ -f ~/.immich_library_revisions ]]; then
|
||||||
libraries=("libjxl" "libheif" "libraw" "imagemagick" "libvips")
|
libraries=("libjxl" "libheif" "libraw" "imagemagick" "libvips")
|
||||||
cd "$BASE_DIR"
|
cd "$BASE_DIR"
|
||||||
msg_info "Checking for updates to custom image-processing libraries"
|
msg_warn "Checking for updates to custom image-processing libraries (recompile time: 2-15min per library)"
|
||||||
$STD git pull
|
$STD git pull
|
||||||
for library in "${libraries[@]}"; do
|
for library in "${libraries[@]}"; do
|
||||||
compile_"$library"
|
compile_"$library"
|
||||||
|
|||||||
@@ -28,7 +28,7 @@ function update_script() {
|
|||||||
exit
|
exit
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [[ -f "$HOME/.overseerr" ]] && [[ "$(cat "$HOME/.overseerr")" == "1.34.0" ]]; then
|
if [[ -f "$HOME/.overseerr" ]] && [[ "$(printf '%s\n' "1.34.0" "$(cat "$HOME/.overseerr")" | sort -V | head -n1)" == "1.35.0" ]]; then
|
||||||
echo
|
echo
|
||||||
echo "━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━"
|
echo "━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━"
|
||||||
echo "Overseerr v1.34.0 detected."
|
echo "Overseerr v1.34.0 detected."
|
||||||
|
|||||||
@@ -65,6 +65,7 @@ function update_script() {
|
|||||||
msg_ok "Stopped Service"
|
msg_ok "Stopped Service"
|
||||||
|
|
||||||
fetch_and_deploy_gh_release "vikunja" "go-vikunja/vikunja" "binary"
|
fetch_and_deploy_gh_release "vikunja" "go-vikunja/vikunja" "binary"
|
||||||
|
$STD systemctl daemon-reload
|
||||||
|
|
||||||
msg_info "Starting Service"
|
msg_info "Starting Service"
|
||||||
systemctl start vikunja
|
systemctl start vikunja
|
||||||
|
|||||||
71
ct/zerobyte.sh
Normal file
71
ct/zerobyte.sh
Normal file
@@ -0,0 +1,71 @@
|
|||||||
|
#!/usr/bin/env bash
|
||||||
|
source <(curl -fsSL https://raw.githubusercontent.com/community-scripts/ProxmoxVE/main/misc/build.func)
|
||||||
|
# Copyright (c) 2021-2026 community-scripts ORG
|
||||||
|
# Author: community-scripts
|
||||||
|
# License: MIT | https://github.com/community-scripts/ProxmoxVE/raw/main/LICENSE
|
||||||
|
# Source: https://github.com/nicotsx/zerobyte
|
||||||
|
|
||||||
|
APP="Zerobyte"
|
||||||
|
var_tags="${var_tags:-backup;encryption;restic}"
|
||||||
|
var_cpu="${var_cpu:-2}"
|
||||||
|
var_ram="${var_ram:-6144}"
|
||||||
|
var_disk="${var_disk:-10}"
|
||||||
|
var_os="${var_os:-debian}"
|
||||||
|
var_version="${var_version:-13}"
|
||||||
|
var_unprivileged="${var_unprivileged:-1}"
|
||||||
|
|
||||||
|
header_info "$APP"
|
||||||
|
variables
|
||||||
|
color
|
||||||
|
catch_errors
|
||||||
|
|
||||||
|
function update_script() {
|
||||||
|
header_info
|
||||||
|
check_container_storage
|
||||||
|
check_container_resources
|
||||||
|
|
||||||
|
if [[ ! -d /opt/zerobyte ]]; then
|
||||||
|
msg_error "No ${APP} Installation Found!"
|
||||||
|
exit
|
||||||
|
fi
|
||||||
|
|
||||||
|
if check_for_gh_release "zerobyte" "nicotsx/zerobyte"; then
|
||||||
|
msg_info "Stopping Service"
|
||||||
|
systemctl stop zerobyte
|
||||||
|
msg_ok "Stopped Service"
|
||||||
|
|
||||||
|
msg_info "Backing up Configuration"
|
||||||
|
cp /opt/zerobyte/.env /opt/zerobyte.env.bak
|
||||||
|
msg_ok "Backed up Configuration"
|
||||||
|
|
||||||
|
NODE_VERSION="24" setup_nodejs
|
||||||
|
CLEAN_INSTALL=1 fetch_and_deploy_gh_release "zerobyte" "nicotsx/zerobyte" "tarball"
|
||||||
|
|
||||||
|
msg_info "Building Zerobyte"
|
||||||
|
export NODE_OPTIONS="--max-old-space-size=3072"
|
||||||
|
cd /opt/zerobyte
|
||||||
|
$STD bun install
|
||||||
|
$STD node ./node_modules/vite/bin/vite.js build
|
||||||
|
msg_ok "Built Zerobyte"
|
||||||
|
|
||||||
|
msg_info "Restoring Configuration"
|
||||||
|
cp /opt/zerobyte.env.bak /opt/zerobyte/.env
|
||||||
|
rm -f /opt/zerobyte.env.bak
|
||||||
|
msg_ok "Restored Configuration"
|
||||||
|
|
||||||
|
msg_info "Starting Service"
|
||||||
|
systemctl start zerobyte
|
||||||
|
msg_ok "Started Service"
|
||||||
|
msg_ok "Updated successfully!"
|
||||||
|
fi
|
||||||
|
exit
|
||||||
|
}
|
||||||
|
|
||||||
|
start
|
||||||
|
build_container
|
||||||
|
description
|
||||||
|
|
||||||
|
msg_ok "Completed Successfully!\n"
|
||||||
|
echo -e "${CREATING}${GN}${APP} setup has been successfully initialized!${CL}"
|
||||||
|
echo -e "${INFO}${YW} Access it using the following URL:${CL}"
|
||||||
|
echo -e "${TAB}${GATEWAY}${BGN}http://${IP}:4096${CL}"
|
||||||
@@ -1,5 +1,5 @@
|
|||||||
{
|
{
|
||||||
"generated": "2026-02-25T06:25:10Z",
|
"generated": "2026-02-26T06:22:43Z",
|
||||||
"versions": [
|
"versions": [
|
||||||
{
|
{
|
||||||
"slug": "2fauth",
|
"slug": "2fauth",
|
||||||
@@ -39,9 +39,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "ampache",
|
"slug": "ampache",
|
||||||
"repo": "ampache/ampache",
|
"repo": "ampache/ampache",
|
||||||
"version": "7.9.0",
|
"version": "7.9.1",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-19T07:01:25Z"
|
"date": "2026-02-25T08:52:58Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "argus",
|
"slug": "argus",
|
||||||
@@ -151,9 +151,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "booklore",
|
"slug": "booklore",
|
||||||
"repo": "booklore-app/BookLore",
|
"repo": "booklore-app/BookLore",
|
||||||
"version": "v2.0.1",
|
"version": "v2.0.2",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-24T04:15:33Z"
|
"date": "2026-02-25T19:59:20Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "bookstack",
|
"slug": "bookstack",
|
||||||
@@ -242,9 +242,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "cosmos",
|
"slug": "cosmos",
|
||||||
"repo": "azukaar/Cosmos-Server",
|
"repo": "azukaar/Cosmos-Server",
|
||||||
"version": "v0.20.2",
|
"version": "v0.21.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-01-24T00:12:39Z"
|
"date": "2026-02-25T17:26:37Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "cronicle",
|
"slug": "cronicle",
|
||||||
@@ -270,16 +270,16 @@
|
|||||||
{
|
{
|
||||||
"slug": "databasus",
|
"slug": "databasus",
|
||||||
"repo": "databasus/databasus",
|
"repo": "databasus/databasus",
|
||||||
"version": "v3.16.2",
|
"version": "v3.16.3",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-22T21:10:12Z"
|
"date": "2026-02-25T19:57:26Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "dawarich",
|
"slug": "dawarich",
|
||||||
"repo": "Freika/dawarich",
|
"repo": "Freika/dawarich",
|
||||||
"version": "1.2.0",
|
"version": "1.3.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-15T22:33:56Z"
|
"date": "2026-02-25T19:30:25Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "discopanel",
|
"slug": "discopanel",
|
||||||
@@ -452,9 +452,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "gitea-mirror",
|
"slug": "gitea-mirror",
|
||||||
"repo": "RayLabsHQ/gitea-mirror",
|
"repo": "RayLabsHQ/gitea-mirror",
|
||||||
"version": "v3.9.4",
|
"version": "v3.9.5",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-24T06:17:56Z"
|
"date": "2026-02-26T05:32:12Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "glance",
|
"slug": "glance",
|
||||||
@@ -606,16 +606,16 @@
|
|||||||
{
|
{
|
||||||
"slug": "invoiceninja",
|
"slug": "invoiceninja",
|
||||||
"repo": "invoiceninja/invoiceninja",
|
"repo": "invoiceninja/invoiceninja",
|
||||||
"version": "v5.12.66",
|
"version": "v5.12.68",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-24T09:12:50Z"
|
"date": "2026-02-25T19:38:19Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "jackett",
|
"slug": "jackett",
|
||||||
"repo": "Jackett/Jackett",
|
"repo": "Jackett/Jackett",
|
||||||
"version": "v0.24.1205",
|
"version": "v0.24.1218",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-25T05:49:14Z"
|
"date": "2026-02-26T05:55:11Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "jellystat",
|
"slug": "jellystat",
|
||||||
@@ -627,9 +627,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "joplin-server",
|
"slug": "joplin-server",
|
||||||
"repo": "laurent22/joplin",
|
"repo": "laurent22/joplin",
|
||||||
"version": "v3.5.12",
|
"version": "v3.5.13",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-01-17T14:20:33Z"
|
"date": "2026-02-25T21:19:11Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "jotty",
|
"slug": "jotty",
|
||||||
@@ -669,9 +669,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "kimai",
|
"slug": "kimai",
|
||||||
"repo": "kimai/kimai",
|
"repo": "kimai/kimai",
|
||||||
"version": "2.49.0",
|
"version": "2.50.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-15T20:40:19Z"
|
"date": "2026-02-25T20:13:51Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "kitchenowl",
|
"slug": "kitchenowl",
|
||||||
@@ -711,9 +711,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "kubo",
|
"slug": "kubo",
|
||||||
"repo": "ipfs/kubo",
|
"repo": "ipfs/kubo",
|
||||||
"version": "v0.39.0",
|
"version": "v0.40.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2025-11-27T03:47:38Z"
|
"date": "2026-02-25T23:16:17Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "kutt",
|
"slug": "kutt",
|
||||||
@@ -823,9 +823,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "manyfold",
|
"slug": "manyfold",
|
||||||
"repo": "manyfold3d/manyfold",
|
"repo": "manyfold3d/manyfold",
|
||||||
"version": "v0.132.1",
|
"version": "v0.133.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-09T22:02:28Z"
|
"date": "2026-02-25T10:40:26Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "mealie",
|
"slug": "mealie",
|
||||||
@@ -995,6 +995,13 @@
|
|||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-03T09:00:43Z"
|
"date": "2026-02-03T09:00:43Z"
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"slug": "openproject",
|
||||||
|
"repo": "jemalloc/jemalloc",
|
||||||
|
"version": "5.3.0",
|
||||||
|
"pinned": false,
|
||||||
|
"date": "2022-05-06T19:14:21Z"
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"slug": "ots",
|
"slug": "ots",
|
||||||
"repo": "Luzifer/ots",
|
"repo": "Luzifer/ots",
|
||||||
@@ -1159,9 +1166,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "prometheus",
|
"slug": "prometheus",
|
||||||
"repo": "prometheus/prometheus",
|
"repo": "prometheus/prometheus",
|
||||||
"version": "v3.9.1",
|
"version": "v3.10.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-01-07T17:05:53Z"
|
"date": "2026-02-26T01:19:51Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "prometheus-alertmanager",
|
"slug": "prometheus-alertmanager",
|
||||||
@@ -1257,9 +1264,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "radicale",
|
"slug": "radicale",
|
||||||
"repo": "Kozea/Radicale",
|
"repo": "Kozea/Radicale",
|
||||||
"version": "v3.6.0",
|
"version": "v3.6.1",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-01-10T06:56:46Z"
|
"date": "2026-02-24T06:36:23Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "rclone",
|
"slug": "rclone",
|
||||||
@@ -1285,9 +1292,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "recyclarr",
|
"slug": "recyclarr",
|
||||||
"repo": "recyclarr/recyclarr",
|
"repo": "recyclarr/recyclarr",
|
||||||
"version": "v8.3.1",
|
"version": "v8.3.2",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-25T01:01:31Z"
|
"date": "2026-02-25T22:39:51Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "reitti",
|
"slug": "reitti",
|
||||||
@@ -1383,9 +1390,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "signoz",
|
"slug": "signoz",
|
||||||
"repo": "SigNoz/signoz-otel-collector",
|
"repo": "SigNoz/signoz-otel-collector",
|
||||||
"version": "v0.144.1",
|
"version": "v0.144.2",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-25T05:57:17Z"
|
"date": "2026-02-26T05:57:26Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "silverbullet",
|
"slug": "silverbullet",
|
||||||
@@ -1593,9 +1600,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "tunarr",
|
"slug": "tunarr",
|
||||||
"repo": "chrisbenincasa/tunarr",
|
"repo": "chrisbenincasa/tunarr",
|
||||||
"version": "v1.1.16",
|
"version": "v1.1.17",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-23T21:24:47Z"
|
"date": "2026-02-25T19:56:36Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "uhf",
|
"slug": "uhf",
|
||||||
@@ -1656,9 +1663,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "vikunja",
|
"slug": "vikunja",
|
||||||
"repo": "go-vikunja/vikunja",
|
"repo": "go-vikunja/vikunja",
|
||||||
"version": "v1.1.0",
|
"version": "v2.0.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-09T10:34:29Z"
|
"date": "2026-02-25T13:58:47Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "wallabag",
|
"slug": "wallabag",
|
||||||
@@ -1772,6 +1779,13 @@
|
|||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-24T15:15:46Z"
|
"date": "2026-02-24T15:15:46Z"
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"slug": "zerobyte",
|
||||||
|
"repo": "restic/restic",
|
||||||
|
"version": "v0.18.1",
|
||||||
|
"pinned": false,
|
||||||
|
"date": "2025-09-21T18:24:38Z"
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"slug": "zigbee2mqtt",
|
"slug": "zigbee2mqtt",
|
||||||
"repo": "Koenkk/zigbee2mqtt",
|
"repo": "Koenkk/zigbee2mqtt",
|
||||||
|
|||||||
@@ -51,6 +51,10 @@
|
|||||||
{
|
{
|
||||||
"text": "Logs: `/var/log/immich`",
|
"text": "Logs: `/var/log/immich`",
|
||||||
"type": "info"
|
"type": "info"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"text": "During first install, 5 custom libraries need to be compiled from source. Depending on your CPU, this can take anywhere between 15 minutes and 2 hours. Please be patient. Touch grass or something.",
|
||||||
|
"type": "warning"
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
|
|||||||
40
frontend/public/json/zerobyte.json
Normal file
40
frontend/public/json/zerobyte.json
Normal file
@@ -0,0 +1,40 @@
|
|||||||
|
{
|
||||||
|
"name": "Zerobyte",
|
||||||
|
"slug": "zerobyte",
|
||||||
|
"categories": [
|
||||||
|
7
|
||||||
|
],
|
||||||
|
"date_created": "2026-02-25",
|
||||||
|
"type": "ct",
|
||||||
|
"updateable": true,
|
||||||
|
"privileged": false,
|
||||||
|
"interface_port": 4096,
|
||||||
|
"documentation": "https://github.com/nicotsx/zerobyte#readme",
|
||||||
|
"website": "https://github.com/nicotsx/zerobyte",
|
||||||
|
"logo": "https://cdn.jsdelivr.net/gh/selfhst/icons@main/webp/zerobyte.webp",
|
||||||
|
"config_path": "/opt/zerobyte/.env",
|
||||||
|
"description": "Zerobyte is a backup automation tool built on top of Restic that provides a modern web interface to schedule, manage, and monitor encrypted backups across multiple storage backends including NFS, SMB, WebDAV, SFTP, S3, and local directories.",
|
||||||
|
"install_methods": [
|
||||||
|
{
|
||||||
|
"type": "default",
|
||||||
|
"script": "ct/zerobyte.sh",
|
||||||
|
"resources": {
|
||||||
|
"cpu": 2,
|
||||||
|
"ram": 6144,
|
||||||
|
"hdd": 10,
|
||||||
|
"os": "Debian",
|
||||||
|
"version": "13"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"default_credentials": {
|
||||||
|
"username": null,
|
||||||
|
"password": null
|
||||||
|
},
|
||||||
|
"notes": [
|
||||||
|
{
|
||||||
|
"text": "For remote mount support (NFS, SMB, WebDAV, SFTP), enable FUSE device passthrough on the LXC container. (FUSE is pre-configured)",
|
||||||
|
"type": "info"
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
@@ -154,7 +154,7 @@ sed -i -e "/^#shared_preload/s/^#//;/^shared_preload/s/''/'vchord.so'/" /etc/pos
|
|||||||
systemctl restart postgresql.service
|
systemctl restart postgresql.service
|
||||||
PG_DB_NAME="immich" PG_DB_USER="immich" PG_DB_GRANT_SUPERUSER="true" PG_DB_SKIP_ALTER_ROLE="true" setup_postgresql_db
|
PG_DB_NAME="immich" PG_DB_USER="immich" PG_DB_GRANT_SUPERUSER="true" PG_DB_SKIP_ALTER_ROLE="true" setup_postgresql_db
|
||||||
|
|
||||||
msg_info "Compiling Custom Photo-processing Library (extreme patience)"
|
msg_warn "Compiling Custom Photo-processing Libraries (can take anywhere from 15min to 2h)"
|
||||||
LD_LIBRARY_PATH=/usr/local/lib
|
LD_LIBRARY_PATH=/usr/local/lib
|
||||||
export LD_RUN_PATH=/usr/local/lib
|
export LD_RUN_PATH=/usr/local/lib
|
||||||
STAGING_DIR=/opt/staging
|
STAGING_DIR=/opt/staging
|
||||||
|
|||||||
@@ -44,6 +44,8 @@ echo passbolt-ce-server passbolt/nginx-domain string $LOCAL_IP | debconf-set-sel
|
|||||||
echo passbolt-ce-server passbolt/nginx-certificate-file string /etc/ssl/passbolt/passbolt.crt | debconf-set-selections
|
echo passbolt-ce-server passbolt/nginx-certificate-file string /etc/ssl/passbolt/passbolt.crt | debconf-set-selections
|
||||||
echo passbolt-ce-server passbolt/nginx-certificate-key-file string /etc/ssl/passbolt/passbolt.key | debconf-set-selections
|
echo passbolt-ce-server passbolt/nginx-certificate-key-file string /etc/ssl/passbolt/passbolt.key | debconf-set-selections
|
||||||
$STD apt install -y --no-install-recommends passbolt-ce-server
|
$STD apt install -y --no-install-recommends passbolt-ce-server
|
||||||
|
sed -i 's/client_max_body_size[[:space:]]\+[0-9]\+M;/client_max_body_size 15M;/' /etc/nginx/sites-enabled/nginx-passbolt.conf
|
||||||
|
systemctl reload nginx
|
||||||
msg_ok "Setup Passbolt"
|
msg_ok "Setup Passbolt"
|
||||||
|
|
||||||
motd_ssh
|
motd_ssh
|
||||||
|
|||||||
@@ -164,7 +164,7 @@ server {
|
|||||||
|
|
||||||
location / {
|
location / {
|
||||||
proxy_pass http://127.0.0.1:8000;
|
proxy_pass http://127.0.0.1:8000;
|
||||||
proxy_set_header Host $host;
|
proxy_set_header Host $http_host;
|
||||||
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
||||||
proxy_set_header X-Forwarded-Proto $scheme;
|
proxy_set_header X-Forwarded-Proto $scheme;
|
||||||
proxy_redirect off;
|
proxy_redirect off;
|
||||||
|
|||||||
@@ -28,12 +28,23 @@ setup_deb822_repo \
|
|||||||
"stable" \
|
"stable" \
|
||||||
"main"
|
"main"
|
||||||
$STD apt install -y elasticsearch
|
$STD apt install -y elasticsearch
|
||||||
sed -i 's/^-Xms.*/-Xms2g/' /etc/elasticsearch/jvm.options
|
sed -i 's/^#\{0,2\} *-Xms[0-9]*g.*/-Xms2g/' /etc/elasticsearch/jvm.options
|
||||||
sed -i 's/^-Xmx.*/-Xmx2g/' /etc/elasticsearch/jvm.options
|
sed -i 's/^#\{0,2\} *-Xmx[0-9]*g.*/-Xmx2g/' /etc/elasticsearch/jvm.options
|
||||||
|
cat <<EOF >>/etc/elasticsearch/elasticsearch.yml
|
||||||
|
discovery.type: single-node
|
||||||
|
xpack.security.enabled: false
|
||||||
|
bootstrap.memory_lock: false
|
||||||
|
EOF
|
||||||
$STD /usr/share/elasticsearch/bin/elasticsearch-plugin install ingest-attachment -b
|
$STD /usr/share/elasticsearch/bin/elasticsearch-plugin install ingest-attachment -b
|
||||||
systemctl daemon-reload
|
systemctl daemon-reload
|
||||||
systemctl enable -q elasticsearch
|
systemctl enable -q elasticsearch
|
||||||
systemctl restart -q elasticsearch
|
systemctl restart -q elasticsearch
|
||||||
|
for i in $(seq 1 30); do
|
||||||
|
if curl -s http://localhost:9200 >/dev/null 2>&1; then
|
||||||
|
break
|
||||||
|
fi
|
||||||
|
sleep 2
|
||||||
|
done
|
||||||
msg_ok "Setup Elasticsearch"
|
msg_ok "Setup Elasticsearch"
|
||||||
|
|
||||||
msg_info "Installing Zammad"
|
msg_info "Installing Zammad"
|
||||||
|
|||||||
96
install/zerobyte-install.sh
Normal file
96
install/zerobyte-install.sh
Normal file
@@ -0,0 +1,96 @@
|
|||||||
|
#!/usr/bin/env bash
|
||||||
|
|
||||||
|
# Copyright (c) 2021-2026 community-scripts ORG
|
||||||
|
# Author: community-scripts
|
||||||
|
# License: MIT | https://github.com/community-scripts/ProxmoxVE/raw/main/LICENSE
|
||||||
|
# Source: https://github.com/nicotsx/zerobyte
|
||||||
|
|
||||||
|
source /dev/stdin <<<"$FUNCTIONS_FILE_PATH"
|
||||||
|
color
|
||||||
|
verb_ip6
|
||||||
|
catch_errors
|
||||||
|
setting_up_container
|
||||||
|
network_check
|
||||||
|
update_os
|
||||||
|
|
||||||
|
msg_info "Installing Dependencies"
|
||||||
|
echo "davfs2 davfs2/suid_file boolean false" | debconf-set-selections
|
||||||
|
$STD apt-get install -y \
|
||||||
|
bzip2 \
|
||||||
|
fuse3 \
|
||||||
|
sshfs \
|
||||||
|
davfs2 \
|
||||||
|
openssh-client
|
||||||
|
msg_ok "Installed Dependencies"
|
||||||
|
|
||||||
|
fetch_and_deploy_gh_release "restic" "restic/restic" "singlefile" "latest" "/usr/local/bin" "restic_*_linux_amd64.bz2"
|
||||||
|
mv /usr/local/bin/restic /usr/local/bin/restic.bz2
|
||||||
|
bzip2 -d /usr/local/bin/restic.bz2
|
||||||
|
chmod +x /usr/local/bin/restic
|
||||||
|
|
||||||
|
fetch_and_deploy_gh_release "rclone" "rclone/rclone" "prebuild" "latest" "/opt/rclone" "rclone-*-linux-amd64.zip"
|
||||||
|
ln -sf /opt/rclone/rclone /usr/local/bin/rclone
|
||||||
|
|
||||||
|
fetch_and_deploy_gh_release "shoutrrr" "nicholas-fedor/shoutrrr" "prebuild" "latest" "/opt/shoutrrr" "shoutrrr_linux_amd64_*.tar.gz"
|
||||||
|
ln -sf /opt/shoutrrr/shoutrrr /usr/local/bin/shoutrrr
|
||||||
|
|
||||||
|
msg_info "Installing Bun"
|
||||||
|
export BUN_INSTALL="/root/.bun"
|
||||||
|
curl -fsSL https://bun.sh/install | $STD bash
|
||||||
|
ln -sf /root/.bun/bin/bun /usr/local/bin/bun
|
||||||
|
ln -sf /root/.bun/bin/bunx /usr/local/bin/bunx
|
||||||
|
msg_ok "Installed Bun"
|
||||||
|
|
||||||
|
NODE_VERSION="24" setup_nodejs
|
||||||
|
fetch_and_deploy_gh_release "zerobyte" "nicotsx/zerobyte" "tarball"
|
||||||
|
|
||||||
|
msg_info "Building Zerobyte (Patience)"
|
||||||
|
cd /opt/zerobyte
|
||||||
|
export VITE_RESTIC_VERSION=$(cat ~/.restic)
|
||||||
|
export VITE_RCLONE_VERSION=$(cat ~/.rclone)
|
||||||
|
export VITE_SHOUTRRR_VERSION=$(cat ~/.shoutrrr)
|
||||||
|
export NODE_OPTIONS="--max-old-space-size=3072"
|
||||||
|
$STD bun install
|
||||||
|
$STD node ./node_modules/vite/bin/vite.js build
|
||||||
|
msg_ok "Built Zerobyte"
|
||||||
|
|
||||||
|
msg_info "Configuring Zerobyte"
|
||||||
|
mkdir -p /var/lib/zerobyte/{data,restic/cache,repositories,volumes}
|
||||||
|
APP_SECRET=$(openssl rand -hex 32)
|
||||||
|
cat <<EOF >/opt/zerobyte/.env
|
||||||
|
BASE_URL=http://${LOCAL_IP}:4096
|
||||||
|
APP_SECRET=${APP_SECRET}
|
||||||
|
PORT=4096
|
||||||
|
ZEROBYTE_DATABASE_URL=/var/lib/zerobyte/data/zerobyte.db
|
||||||
|
RESTIC_CACHE_DIR=/var/lib/zerobyte/restic/cache
|
||||||
|
ZEROBYTE_REPOSITORIES_DIR=/var/lib/zerobyte/repositories
|
||||||
|
ZEROBYTE_VOLUMES_DIR=/var/lib/zerobyte/volumes
|
||||||
|
MIGRATIONS_PATH=/opt/zerobyte/app/drizzle
|
||||||
|
NODE_ENV=production
|
||||||
|
EOF
|
||||||
|
msg_ok "Configured Zerobyte"
|
||||||
|
|
||||||
|
msg_info "Creating Service"
|
||||||
|
cat <<EOF >/etc/systemd/system/zerobyte.service
|
||||||
|
[Unit]
|
||||||
|
Description=Zerobyte Backup Automation
|
||||||
|
After=network.target
|
||||||
|
|
||||||
|
[Service]
|
||||||
|
Type=simple
|
||||||
|
User=root
|
||||||
|
WorkingDirectory=/opt/zerobyte
|
||||||
|
EnvironmentFile=/opt/zerobyte/.env
|
||||||
|
ExecStart=/usr/local/bin/bun .output/server/index.mjs
|
||||||
|
Restart=on-failure
|
||||||
|
RestartSec=5
|
||||||
|
|
||||||
|
[Install]
|
||||||
|
WantedBy=multi-user.target
|
||||||
|
EOF
|
||||||
|
systemctl enable -q --now zerobyte
|
||||||
|
msg_ok "Created Service"
|
||||||
|
|
||||||
|
motd_ssh
|
||||||
|
customize
|
||||||
|
cleanup_lxc
|
||||||
@@ -312,7 +312,8 @@ json_escape() {
|
|||||||
s=${s//$'\r'/}
|
s=${s//$'\r'/}
|
||||||
s=${s//$'\t'/\\t}
|
s=${s//$'\t'/\\t}
|
||||||
# Remove any remaining control characters (0x00-0x1F except those already handled)
|
# Remove any remaining control characters (0x00-0x1F except those already handled)
|
||||||
s=$(printf '%s' "$s" | tr -d '\000-\010\013\014\016-\037')
|
# Also remove DEL (0x7F) and invalid high bytes that break JSON parsers
|
||||||
|
s=$(printf '%s' "$s" | tr -d '\000-\010\013\014\016-\037\177')
|
||||||
printf '%s' "$s"
|
printf '%s' "$s"
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -982,7 +983,7 @@ EOF
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
# All 3 attempts failed — do NOT set POST_UPDATE_DONE=true.
|
# All 3 attempts failed — do NOT set POST_UPDATE_DONE=true.
|
||||||
# This allows the EXIT trap (api_exit_script) to retry with 3 fresh attempts.
|
# This allows the EXIT trap (on_exit in error_handler.func) to retry.
|
||||||
# No infinite loop risk: EXIT trap fires exactly once.
|
# No infinite loop risk: EXIT trap fires exactly once.
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -4098,6 +4098,12 @@ EOF'
|
|||||||
|
|
||||||
# Installation failed?
|
# Installation failed?
|
||||||
if [[ $install_exit_code -ne 0 ]]; then
|
if [[ $install_exit_code -ne 0 ]]; then
|
||||||
|
# Prevent job-control signals from suspending the script during recovery.
|
||||||
|
# In non-interactive shells (bash -c), background processes (spinner) can
|
||||||
|
# trigger terminal-related signals that stop the entire process group.
|
||||||
|
# TSTP = Ctrl+Z, TTIN = bg read from tty, TTOU = bg write to tty (tostop)
|
||||||
|
trap '' TSTP TTIN TTOU
|
||||||
|
|
||||||
msg_error "Installation failed in container ${CTID} (exit code: ${install_exit_code})"
|
msg_error "Installation failed in container ${CTID} (exit code: ${install_exit_code})"
|
||||||
|
|
||||||
# Copy install log from container BEFORE API call so get_error_text() can read it
|
# Copy install log from container BEFORE API call so get_error_text() can read it
|
||||||
@@ -4173,7 +4179,12 @@ EOF'
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
# Report failure to telemetry API (now with log available on host)
|
# Report failure to telemetry API (now with log available on host)
|
||||||
|
# NOTE: Do NOT use msg_info/spinner here — the background spinner process
|
||||||
|
# causes SIGTSTP in non-interactive shells (bash -c "$(curl ...)"), which
|
||||||
|
# stops the entire process group and prevents the recovery dialog from appearing.
|
||||||
|
$STD echo -e "${TAB}⏳ Reporting failure to telemetry..."
|
||||||
post_update_to_api "failed" "$install_exit_code"
|
post_update_to_api "failed" "$install_exit_code"
|
||||||
|
$STD echo -e "${TAB}${CM:-✔} Failure reported"
|
||||||
|
|
||||||
# Defense-in-depth: Ensure error handling stays disabled during recovery.
|
# Defense-in-depth: Ensure error handling stays disabled during recovery.
|
||||||
# Some functions (e.g. silent/$STD) unconditionally re-enable set -Eeuo pipefail
|
# Some functions (e.g. silent/$STD) unconditionally re-enable set -Eeuo pipefail
|
||||||
@@ -4537,8 +4548,12 @@ EOF'
|
|||||||
|
|
||||||
# Force one final status update attempt after cleanup
|
# Force one final status update attempt after cleanup
|
||||||
# This ensures status is updated even if the first attempt failed (e.g., HTTP 400)
|
# This ensures status is updated even if the first attempt failed (e.g., HTTP 400)
|
||||||
|
$STD echo -e "${TAB}⏳ Finalizing telemetry report..."
|
||||||
post_update_to_api "failed" "$install_exit_code" "force"
|
post_update_to_api "failed" "$install_exit_code" "force"
|
||||||
|
$STD echo -e "${TAB}${CM:-✔} Telemetry finalized"
|
||||||
|
|
||||||
|
# Restore default job-control signal handling before exit
|
||||||
|
trap - TSTP TTIN TTOU
|
||||||
exit $install_exit_code
|
exit $install_exit_code
|
||||||
fi
|
fi
|
||||||
|
|
||||||
@@ -5608,44 +5623,21 @@ ensure_log_on_host() {
|
|||||||
fi
|
fi
|
||||||
}
|
}
|
||||||
|
|
||||||
# ------------------------------------------------------------------------------
|
# ==============================================================================
|
||||||
# api_exit_script()
|
# TRAP MANAGEMENT
|
||||||
|
# ==============================================================================
|
||||||
|
# All traps (ERR, EXIT, INT, TERM, HUP) are set by catch_errors() in
|
||||||
|
# error_handler.func — called at the top of this file after sourcing.
|
||||||
#
|
#
|
||||||
# - Exit trap handler for reporting to API telemetry
|
# Do NOT set duplicate traps here. The handlers in error_handler.func
|
||||||
# - Captures exit code and reports to PocketBase using centralized error descriptions
|
# (on_exit, on_interrupt, on_terminate, on_hangup, error_handler) already:
|
||||||
# - Uses explain_exit_code() from api.func for consistent error messages
|
# - Send telemetry via post_update_to_api / _send_abort_telemetry
|
||||||
# - ALWAYS sends telemetry FIRST before log collection to prevent pct pull
|
# - Stop orphaned containers via _stop_container_if_installing
|
||||||
# hangs from blocking status updates (container may be dead/unresponsive)
|
# - Collect logs via ensure_log_on_host
|
||||||
# - For non-zero exit codes: posts "failed" status
|
# - Clean up lock files and spinner processes
|
||||||
# - For zero exit codes where post_update_to_api was never called:
|
#
|
||||||
# catches orphaned "installing" records (e.g., script exited cleanly
|
# Previously, inline traps here overwrote catch_errors() traps, causing:
|
||||||
# but description() was never reached)
|
# - error_handler() never fired (no error output, no cleanup dialog)
|
||||||
# ------------------------------------------------------------------------------
|
# - on_hangup() never fired (SSH disconnect → stuck records)
|
||||||
api_exit_script() {
|
# - Duplicated logic in two places (hard to debug)
|
||||||
local exit_code=$?
|
# ==============================================================================
|
||||||
if [ $exit_code -ne 0 ]; then
|
|
||||||
# ALWAYS send telemetry FIRST - ensure status is reported even if
|
|
||||||
# ensure_log_on_host hangs (e.g. pct pull on dead container)
|
|
||||||
post_update_to_api "failed" "$exit_code" 2>/dev/null || true
|
|
||||||
# Best-effort log collection (non-critical after telemetry is sent)
|
|
||||||
if declare -f ensure_log_on_host >/dev/null 2>&1; then
|
|
||||||
ensure_log_on_host 2>/dev/null || true
|
|
||||||
fi
|
|
||||||
# Stop orphaned container if we're in the install phase
|
|
||||||
if [[ "${CONTAINER_INSTALLING:-}" == "true" && -n "${CTID:-}" ]] && command -v pct &>/dev/null; then
|
|
||||||
pct stop "$CTID" 2>/dev/null || true
|
|
||||||
fi
|
|
||||||
elif [[ "${POST_TO_API_DONE:-}" == "true" && "${POST_UPDATE_DONE:-}" != "true" ]]; then
|
|
||||||
# Script exited with 0 but never sent a completion status
|
|
||||||
# exit_code=0 is never an error — report as success
|
|
||||||
post_update_to_api "done" "0"
|
|
||||||
fi
|
|
||||||
}
|
|
||||||
|
|
||||||
if command -v pveversion >/dev/null 2>&1; then
|
|
||||||
trap 'api_exit_script' EXIT
|
|
||||||
fi
|
|
||||||
trap 'local _ec=$?; if [[ $_ec -ne 0 ]]; then post_update_to_api "failed" "$_ec" 2>/dev/null || true; if declare -f ensure_log_on_host &>/dev/null; then ensure_log_on_host 2>/dev/null || true; fi; fi' ERR
|
|
||||||
trap 'post_update_to_api "failed" "129" 2>/dev/null || true; if [[ -n "${CTID:-}" ]] && command -v pct &>/dev/null; then pct stop "$CTID" 2>/dev/null || true; fi; exit 129' SIGHUP
|
|
||||||
trap 'post_update_to_api "failed" "130" 2>/dev/null || true; if [[ -n "${CTID:-}" ]] && command -v pct &>/dev/null; then pct stop "$CTID" 2>/dev/null || true; fi; exit 130' SIGINT
|
|
||||||
trap 'post_update_to_api "failed" "143" 2>/dev/null || true; if [[ -n "${CTID:-}" ]] && command -v pct &>/dev/null; then pct stop "$CTID" 2>/dev/null || true; fi; exit 143' SIGTERM
|
|
||||||
|
|||||||
@@ -607,6 +607,7 @@ stop_spinner() {
|
|||||||
|
|
||||||
unset SPINNER_PID SPINNER_MSG
|
unset SPINNER_PID SPINNER_MSG
|
||||||
stty sane 2>/dev/null || true
|
stty sane 2>/dev/null || true
|
||||||
|
stty -tostop 2>/dev/null || true
|
||||||
}
|
}
|
||||||
|
|
||||||
# ==============================================================================
|
# ==============================================================================
|
||||||
|
|||||||
@@ -199,11 +199,16 @@ error_handler() {
|
|||||||
return 0
|
return 0
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
# Stop spinner and restore cursor FIRST — before any output
|
||||||
|
# This prevents spinner text overlapping with error messages
|
||||||
|
if declare -f stop_spinner >/dev/null 2>&1; then
|
||||||
|
stop_spinner 2>/dev/null || true
|
||||||
|
fi
|
||||||
|
printf "\e[?25h"
|
||||||
|
|
||||||
local explanation
|
local explanation
|
||||||
explanation="$(explain_exit_code "$exit_code")"
|
explanation="$(explain_exit_code "$exit_code")"
|
||||||
|
|
||||||
printf "\e[?25h"
|
|
||||||
|
|
||||||
# ALWAYS report failure to API immediately - don't wait for container checks
|
# ALWAYS report failure to API immediately - don't wait for container checks
|
||||||
# This ensures we capture failures that occur before/after container exists
|
# This ensures we capture failures that occur before/after container exists
|
||||||
if declare -f post_update_to_api &>/dev/null; then
|
if declare -f post_update_to_api &>/dev/null; then
|
||||||
@@ -359,9 +364,39 @@ _send_abort_telemetry() {
|
|||||||
command -v curl &>/dev/null || return 0
|
command -v curl &>/dev/null || return 0
|
||||||
[[ "${DIAGNOSTICS:-no}" == "no" ]] && return 0
|
[[ "${DIAGNOSTICS:-no}" == "no" ]] && return 0
|
||||||
[[ -z "${RANDOM_UUID:-}" ]] && return 0
|
[[ -z "${RANDOM_UUID:-}" ]] && return 0
|
||||||
curl -fsS -m 5 -X POST "${TELEMETRY_URL:-https://telemetry.community-scripts.org/telemetry}" \
|
|
||||||
-H "Content-Type: application/json" \
|
# Collect last 20 log lines for error diagnosis (best-effort)
|
||||||
-d "{\"random_id\":\"${RANDOM_UUID}\",\"execution_id\":\"${EXECUTION_ID:-${RANDOM_UUID}}\",\"type\":\"${TELEMETRY_TYPE:-lxc}\",\"nsapp\":\"${NSAPP:-${app:-unknown}}\",\"status\":\"failed\",\"exit_code\":${exit_code}}" &>/dev/null || true
|
local error_text=""
|
||||||
|
if [[ -n "${INSTALL_LOG:-}" && -s "${INSTALL_LOG}" ]]; then
|
||||||
|
error_text=$(tail -n 20 "$INSTALL_LOG" 2>/dev/null | sed 's/\x1b\[[0-9;]*[a-zA-Z]//g; s/\\/\\\\/g; s/"/\\"/g; s/\r//g' | tr '\n' '|' | sed 's/|$//' | tr -d '\000-\010\013\014\016-\037\177') || true
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Calculate duration if start time is available
|
||||||
|
local duration=""
|
||||||
|
if [[ -n "${DIAGNOSTICS_START_TIME:-}" ]]; then
|
||||||
|
duration=$(($(date +%s) - DIAGNOSTICS_START_TIME))
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Build JSON payload with error context
|
||||||
|
local payload
|
||||||
|
payload="{\"random_id\":\"${RANDOM_UUID}\",\"execution_id\":\"${EXECUTION_ID:-${RANDOM_UUID}}\",\"type\":\"${TELEMETRY_TYPE:-lxc}\",\"nsapp\":\"${NSAPP:-${app:-unknown}}\",\"status\":\"failed\",\"exit_code\":${exit_code}"
|
||||||
|
[[ -n "$error_text" ]] && payload="${payload},\"error\":\"${error_text}\""
|
||||||
|
[[ -n "$duration" ]] && payload="${payload},\"duration\":${duration}"
|
||||||
|
payload="${payload}}"
|
||||||
|
|
||||||
|
local api_url="${TELEMETRY_URL:-https://telemetry.community-scripts.org/telemetry}"
|
||||||
|
|
||||||
|
# 2 attempts (retry once on failure) — original had no retry
|
||||||
|
local attempt
|
||||||
|
for attempt in 1 2; do
|
||||||
|
if curl -fsS -m 5 -X POST "$api_url" \
|
||||||
|
-H "Content-Type: application/json" \
|
||||||
|
-d "$payload" &>/dev/null; then
|
||||||
|
return 0
|
||||||
|
fi
|
||||||
|
[[ $attempt -eq 1 ]] && sleep 1
|
||||||
|
done
|
||||||
|
return 0
|
||||||
}
|
}
|
||||||
|
|
||||||
# ------------------------------------------------------------------------------
|
# ------------------------------------------------------------------------------
|
||||||
@@ -437,6 +472,12 @@ on_exit() {
|
|||||||
# - Exits with code 130 (128 + SIGINT=2)
|
# - Exits with code 130 (128 + SIGINT=2)
|
||||||
# ------------------------------------------------------------------------------
|
# ------------------------------------------------------------------------------
|
||||||
on_interrupt() {
|
on_interrupt() {
|
||||||
|
# Stop spinner and restore cursor before any output
|
||||||
|
if declare -f stop_spinner >/dev/null 2>&1; then
|
||||||
|
stop_spinner 2>/dev/null || true
|
||||||
|
fi
|
||||||
|
printf "\e[?25h" 2>/dev/null || true
|
||||||
|
|
||||||
_send_abort_telemetry "130"
|
_send_abort_telemetry "130"
|
||||||
_stop_container_if_installing
|
_stop_container_if_installing
|
||||||
if declare -f msg_error >/dev/null 2>&1; then
|
if declare -f msg_error >/dev/null 2>&1; then
|
||||||
@@ -456,6 +497,12 @@ on_interrupt() {
|
|||||||
# - Exits with code 143 (128 + SIGTERM=15)
|
# - Exits with code 143 (128 + SIGTERM=15)
|
||||||
# ------------------------------------------------------------------------------
|
# ------------------------------------------------------------------------------
|
||||||
on_terminate() {
|
on_terminate() {
|
||||||
|
# Stop spinner and restore cursor before any output
|
||||||
|
if declare -f stop_spinner >/dev/null 2>&1; then
|
||||||
|
stop_spinner 2>/dev/null || true
|
||||||
|
fi
|
||||||
|
printf "\e[?25h" 2>/dev/null || true
|
||||||
|
|
||||||
_send_abort_telemetry "143"
|
_send_abort_telemetry "143"
|
||||||
_stop_container_if_installing
|
_stop_container_if_installing
|
||||||
if declare -f msg_error >/dev/null 2>&1; then
|
if declare -f msg_error >/dev/null 2>&1; then
|
||||||
@@ -478,6 +525,11 @@ on_terminate() {
|
|||||||
# - Exits with code 129 (128 + SIGHUP=1)
|
# - Exits with code 129 (128 + SIGHUP=1)
|
||||||
# ------------------------------------------------------------------------------
|
# ------------------------------------------------------------------------------
|
||||||
on_hangup() {
|
on_hangup() {
|
||||||
|
# Stop spinner (no cursor restore needed — terminal is already gone)
|
||||||
|
if declare -f stop_spinner >/dev/null 2>&1; then
|
||||||
|
stop_spinner 2>/dev/null || true
|
||||||
|
fi
|
||||||
|
|
||||||
_send_abort_telemetry "129"
|
_send_abort_telemetry "129"
|
||||||
_stop_container_if_installing
|
_stop_container_if_installing
|
||||||
exit 129
|
exit 129
|
||||||
|
|||||||
123
misc/tools.func
123
misc/tools.func
@@ -783,16 +783,25 @@ github_api_call() {
|
|||||||
|
|
||||||
for attempt in $(seq 1 $max_retries); do
|
for attempt in $(seq 1 $max_retries); do
|
||||||
local http_code
|
local http_code
|
||||||
http_code=$(curl -fsSL -w "%{http_code}" -o "$output_file" \
|
http_code=$(curl -sSL -w "%{http_code}" -o "$output_file" \
|
||||||
-H "Accept: application/vnd.github+json" \
|
-H "Accept: application/vnd.github+json" \
|
||||||
-H "X-GitHub-Api-Version: 2022-11-28" \
|
-H "X-GitHub-Api-Version: 2022-11-28" \
|
||||||
"${header_args[@]}" \
|
"${header_args[@]}" \
|
||||||
"$url" 2>/dev/null || echo "000")
|
"$url" 2>/dev/null) || true
|
||||||
|
|
||||||
case "$http_code" in
|
case "$http_code" in
|
||||||
200)
|
200)
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
|
401)
|
||||||
|
msg_error "GitHub API authentication failed (HTTP 401)."
|
||||||
|
if [[ -n "${GITHUB_TOKEN:-}" ]]; then
|
||||||
|
msg_error "Your GITHUB_TOKEN appears to be invalid or expired."
|
||||||
|
else
|
||||||
|
msg_error "The repository may require authentication. Try: export GITHUB_TOKEN=\"ghp_your_token\""
|
||||||
|
fi
|
||||||
|
return 1
|
||||||
|
;;
|
||||||
403)
|
403)
|
||||||
# Rate limit - check if we can retry
|
# Rate limit - check if we can retry
|
||||||
if [[ $attempt -lt $max_retries ]]; then
|
if [[ $attempt -lt $max_retries ]]; then
|
||||||
@@ -801,11 +810,22 @@ github_api_call() {
|
|||||||
retry_delay=$((retry_delay * 2))
|
retry_delay=$((retry_delay * 2))
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
msg_error "GitHub API rate limit exceeded. Set GITHUB_TOKEN to increase limits."
|
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
||||||
|
msg_error "To increase the limit, export a GitHub token before running the script:"
|
||||||
|
msg_error " export GITHUB_TOKEN=\"ghp_your_token_here\""
|
||||||
return 1
|
return 1
|
||||||
;;
|
;;
|
||||||
404)
|
404)
|
||||||
msg_error "GitHub API endpoint not found: $url"
|
msg_error "GitHub repository or release not found (HTTP 404): $url"
|
||||||
|
return 1
|
||||||
|
;;
|
||||||
|
000 | "")
|
||||||
|
if [[ $attempt -lt $max_retries ]]; then
|
||||||
|
sleep "$retry_delay"
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
msg_error "GitHub API connection failed (no response)."
|
||||||
|
msg_error "Check your network/DNS: curl -sSL https://api.github.com/rate_limit"
|
||||||
return 1
|
return 1
|
||||||
;;
|
;;
|
||||||
*)
|
*)
|
||||||
@@ -813,7 +833,7 @@ github_api_call() {
|
|||||||
sleep "$retry_delay"
|
sleep "$retry_delay"
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
msg_error "GitHub API call failed with HTTP $http_code"
|
msg_error "GitHub API call failed (HTTP $http_code)."
|
||||||
return 1
|
return 1
|
||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
@@ -833,14 +853,18 @@ codeberg_api_call() {
|
|||||||
|
|
||||||
for attempt in $(seq 1 $max_retries); do
|
for attempt in $(seq 1 $max_retries); do
|
||||||
local http_code
|
local http_code
|
||||||
http_code=$(curl -fsSL -w "%{http_code}" -o "$output_file" \
|
http_code=$(curl -sSL -w "%{http_code}" -o "$output_file" \
|
||||||
-H "Accept: application/json" \
|
-H "Accept: application/json" \
|
||||||
"$url" 2>/dev/null || echo "000")
|
"$url" 2>/dev/null) || true
|
||||||
|
|
||||||
case "$http_code" in
|
case "$http_code" in
|
||||||
200)
|
200)
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
|
401)
|
||||||
|
msg_error "Codeberg API authentication failed (HTTP 401)."
|
||||||
|
return 1
|
||||||
|
;;
|
||||||
403)
|
403)
|
||||||
# Rate limit - retry
|
# Rate limit - retry
|
||||||
if [[ $attempt -lt $max_retries ]]; then
|
if [[ $attempt -lt $max_retries ]]; then
|
||||||
@@ -849,11 +873,20 @@ codeberg_api_call() {
|
|||||||
retry_delay=$((retry_delay * 2))
|
retry_delay=$((retry_delay * 2))
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
msg_error "Codeberg API rate limit exceeded."
|
msg_error "Codeberg API rate limit exceeded (HTTP 403)."
|
||||||
return 1
|
return 1
|
||||||
;;
|
;;
|
||||||
404)
|
404)
|
||||||
msg_error "Codeberg API endpoint not found: $url"
|
msg_error "Codeberg repository or release not found (HTTP 404): $url"
|
||||||
|
return 1
|
||||||
|
;;
|
||||||
|
000 | "")
|
||||||
|
if [[ $attempt -lt $max_retries ]]; then
|
||||||
|
sleep "$retry_delay"
|
||||||
|
continue
|
||||||
|
fi
|
||||||
|
msg_error "Codeberg API connection failed (no response)."
|
||||||
|
msg_error "Check your network/DNS: curl -sSL https://codeberg.org"
|
||||||
return 1
|
return 1
|
||||||
;;
|
;;
|
||||||
*)
|
*)
|
||||||
@@ -861,7 +894,7 @@ codeberg_api_call() {
|
|||||||
sleep "$retry_delay"
|
sleep "$retry_delay"
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
msg_error "Codeberg API call failed with HTTP $http_code"
|
msg_error "Codeberg API call failed (HTTP $http_code)."
|
||||||
return 1
|
return 1
|
||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
@@ -1441,7 +1474,7 @@ get_latest_github_release() {
|
|||||||
|
|
||||||
if ! github_api_call "https://api.github.com/repos/${repo}/releases/latest" "$temp_file"; then
|
if ! github_api_call "https://api.github.com/repos/${repo}/releases/latest" "$temp_file"; then
|
||||||
rm -f "$temp_file"
|
rm -f "$temp_file"
|
||||||
return 1
|
return 0
|
||||||
fi
|
fi
|
||||||
|
|
||||||
local version
|
local version
|
||||||
@@ -1449,7 +1482,8 @@ get_latest_github_release() {
|
|||||||
rm -f "$temp_file"
|
rm -f "$temp_file"
|
||||||
|
|
||||||
if [[ -z "$version" ]]; then
|
if [[ -z "$version" ]]; then
|
||||||
return 1
|
msg_error "Could not determine latest version for ${repo}"
|
||||||
|
return 0
|
||||||
fi
|
fi
|
||||||
|
|
||||||
echo "$version"
|
echo "$version"
|
||||||
@@ -1466,7 +1500,7 @@ get_latest_codeberg_release() {
|
|||||||
# Codeberg API: get all releases and pick the first non-draft/non-prerelease
|
# Codeberg API: get all releases and pick the first non-draft/non-prerelease
|
||||||
if ! codeberg_api_call "https://codeberg.org/api/v1/repos/${repo}/releases" "$temp_file"; then
|
if ! codeberg_api_call "https://codeberg.org/api/v1/repos/${repo}/releases" "$temp_file"; then
|
||||||
rm -f "$temp_file"
|
rm -f "$temp_file"
|
||||||
return 1
|
return 0
|
||||||
fi
|
fi
|
||||||
|
|
||||||
local version
|
local version
|
||||||
@@ -1480,7 +1514,8 @@ get_latest_codeberg_release() {
|
|||||||
rm -f "$temp_file"
|
rm -f "$temp_file"
|
||||||
|
|
||||||
if [[ -z "$version" ]]; then
|
if [[ -z "$version" ]]; then
|
||||||
return 1
|
msg_error "Could not determine latest version for ${repo}"
|
||||||
|
return 0
|
||||||
fi
|
fi
|
||||||
|
|
||||||
echo "$version"
|
echo "$version"
|
||||||
@@ -1567,13 +1602,34 @@ get_latest_gh_tag() {
|
|||||||
"${header_args[@]}" \
|
"${header_args[@]}" \
|
||||||
"https://api.github.com/repos/${repo}/tags?per_page=100" 2>/dev/null) || true
|
"https://api.github.com/repos/${repo}/tags?per_page=100" 2>/dev/null) || true
|
||||||
|
|
||||||
|
if [[ "$http_code" == "401" ]]; then
|
||||||
|
msg_error "GitHub API authentication failed (HTTP 401)."
|
||||||
|
if [[ -n "${GITHUB_TOKEN:-}" ]]; then
|
||||||
|
msg_error "Your GITHUB_TOKEN appears to be invalid or expired."
|
||||||
|
else
|
||||||
|
msg_error "The repository may require authentication. Try: export GITHUB_TOKEN=\"ghp_your_token\""
|
||||||
|
fi
|
||||||
|
rm -f /tmp/gh_tags.json
|
||||||
|
return 1
|
||||||
|
fi
|
||||||
|
|
||||||
if [[ "$http_code" == "403" ]]; then
|
if [[ "$http_code" == "403" ]]; then
|
||||||
msg_warn "GitHub API rate limit exceeded while fetching tags for ${repo}"
|
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
||||||
|
msg_error "To increase the limit, export a GitHub token before running the script:"
|
||||||
|
msg_error " export GITHUB_TOKEN=\"ghp_your_token_here\""
|
||||||
|
rm -f /tmp/gh_tags.json
|
||||||
|
return 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ "$http_code" == "000" || -z "$http_code" ]]; then
|
||||||
|
msg_error "GitHub API connection failed (no response)."
|
||||||
|
msg_error "Check your network/DNS: curl -sSL https://api.github.com/rate_limit"
|
||||||
rm -f /tmp/gh_tags.json
|
rm -f /tmp/gh_tags.json
|
||||||
return 1
|
return 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [[ "$http_code" != "200" ]] || [[ ! -s /tmp/gh_tags.json ]]; then
|
if [[ "$http_code" != "200" ]] || [[ ! -s /tmp/gh_tags.json ]]; then
|
||||||
|
msg_error "Unable to fetch tags for ${repo} (HTTP ${http_code})"
|
||||||
rm -f /tmp/gh_tags.json
|
rm -f /tmp/gh_tags.json
|
||||||
return 1
|
return 1
|
||||||
fi
|
fi
|
||||||
@@ -1659,6 +1715,15 @@ check_for_gh_release() {
|
|||||||
|
|
||||||
if [[ "$http_code" == "200" ]] && [[ -s /tmp/gh_check.json ]]; then
|
if [[ "$http_code" == "200" ]] && [[ -s /tmp/gh_check.json ]]; then
|
||||||
releases_json="[$(</tmp/gh_check.json)]"
|
releases_json="[$(</tmp/gh_check.json)]"
|
||||||
|
elif [[ "$http_code" == "401" ]]; then
|
||||||
|
msg_error "GitHub API authentication failed (HTTP 401)."
|
||||||
|
if [[ -n "${GITHUB_TOKEN:-}" ]]; then
|
||||||
|
msg_error "Your GITHUB_TOKEN appears to be invalid or expired."
|
||||||
|
else
|
||||||
|
msg_error "The repository may require authentication. Try: export GITHUB_TOKEN=\"ghp_your_token\""
|
||||||
|
fi
|
||||||
|
rm -f /tmp/gh_check.json
|
||||||
|
return 1
|
||||||
elif [[ "$http_code" == "403" ]]; then
|
elif [[ "$http_code" == "403" ]]; then
|
||||||
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
||||||
msg_error "To increase the limit, export a GitHub token before running the script:"
|
msg_error "To increase the limit, export a GitHub token before running the script:"
|
||||||
@@ -1679,12 +1744,26 @@ check_for_gh_release() {
|
|||||||
|
|
||||||
if [[ "$http_code" == "200" ]] && [[ -s /tmp/gh_check.json ]]; then
|
if [[ "$http_code" == "200" ]] && [[ -s /tmp/gh_check.json ]]; then
|
||||||
releases_json=$(</tmp/gh_check.json)
|
releases_json=$(</tmp/gh_check.json)
|
||||||
|
elif [[ "$http_code" == "401" ]]; then
|
||||||
|
msg_error "GitHub API authentication failed (HTTP 401)."
|
||||||
|
if [[ -n "${GITHUB_TOKEN:-}" ]]; then
|
||||||
|
msg_error "Your GITHUB_TOKEN appears to be invalid or expired."
|
||||||
|
else
|
||||||
|
msg_error "The repository may require authentication. Try: export GITHUB_TOKEN=\"ghp_your_token\""
|
||||||
|
fi
|
||||||
|
rm -f /tmp/gh_check.json
|
||||||
|
return 1
|
||||||
elif [[ "$http_code" == "403" ]]; then
|
elif [[ "$http_code" == "403" ]]; then
|
||||||
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
||||||
msg_error "To increase the limit, export a GitHub token before running the script:"
|
msg_error "To increase the limit, export a GitHub token before running the script:"
|
||||||
msg_error " export GITHUB_TOKEN=\"ghp_your_token_here\""
|
msg_error " export GITHUB_TOKEN=\"ghp_your_token_here\""
|
||||||
rm -f /tmp/gh_check.json
|
rm -f /tmp/gh_check.json
|
||||||
return 1
|
return 1
|
||||||
|
elif [[ "$http_code" == "000" || -z "$http_code" ]]; then
|
||||||
|
msg_error "GitHub API connection failed (no response)."
|
||||||
|
msg_error "Check your network/DNS: curl -sSL https://api.github.com/rate_limit"
|
||||||
|
rm -f /tmp/gh_check.json
|
||||||
|
return 1
|
||||||
else
|
else
|
||||||
msg_error "Unable to fetch releases for ${app} (HTTP ${http_code})"
|
msg_error "Unable to fetch releases for ${app} (HTTP ${http_code})"
|
||||||
rm -f /tmp/gh_check.json
|
rm -f /tmp/gh_check.json
|
||||||
@@ -2608,12 +2687,22 @@ function fetch_and_deploy_gh_release() {
|
|||||||
done
|
done
|
||||||
|
|
||||||
if ! $success; then
|
if ! $success; then
|
||||||
if [[ "$http_code" == "403" ]]; then
|
if [[ "$http_code" == "401" ]]; then
|
||||||
|
msg_error "GitHub API authentication failed (HTTP 401)."
|
||||||
|
if [[ -n "${GITHUB_TOKEN:-}" ]]; then
|
||||||
|
msg_error "Your GITHUB_TOKEN appears to be invalid or expired."
|
||||||
|
else
|
||||||
|
msg_error "The repository may require authentication. Try: export GITHUB_TOKEN=\"ghp_your_token\""
|
||||||
|
fi
|
||||||
|
elif [[ "$http_code" == "403" ]]; then
|
||||||
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
||||||
msg_error "To increase the limit, export a GitHub token before running the script:"
|
msg_error "To increase the limit, export a GitHub token before running the script:"
|
||||||
msg_error " export GITHUB_TOKEN=\"ghp_your_token_here\""
|
msg_error " export GITHUB_TOKEN=\"ghp_your_token_here\""
|
||||||
|
elif [[ "$http_code" == "000" || -z "$http_code" ]]; then
|
||||||
|
msg_error "GitHub API connection failed (no response)."
|
||||||
|
msg_error "Check your network/DNS: curl -sSL https://api.github.com/rate_limit"
|
||||||
else
|
else
|
||||||
msg_error "Failed to fetch release metadata from $api_url after $max_retries attempts (HTTP $http_code)"
|
msg_error "Failed to fetch release metadata (HTTP $http_code)"
|
||||||
fi
|
fi
|
||||||
return 1
|
return 1
|
||||||
fi
|
fi
|
||||||
|
|||||||
@@ -28,7 +28,7 @@ INSTALL_PATH="/opt/immich-proxy"
|
|||||||
CONFIG_PATH="/opt/immich-proxy/app"
|
CONFIG_PATH="/opt/immich-proxy/app"
|
||||||
DEFAULT_PORT=3000
|
DEFAULT_PORT=3000
|
||||||
|
|
||||||
# Initialize all core functions (colors, formatting, icons, STD mode)
|
# Initialize all core functions (colors, formatting, icons, $STD mode)
|
||||||
load_functions
|
load_functions
|
||||||
init_tool_telemetry "" "addon"
|
init_tool_telemetry "" "addon"
|
||||||
|
|
||||||
|
|||||||
@@ -288,8 +288,8 @@ function default_settings() {
|
|||||||
echo -e "${DGN}Using Hostname: ${BGN}${HN}${CL}"
|
echo -e "${DGN}Using Hostname: ${BGN}${HN}${CL}"
|
||||||
echo -e "${DGN}Allocated Cores: ${BGN}${CORE_COUNT}${CL}"
|
echo -e "${DGN}Allocated Cores: ${BGN}${CORE_COUNT}${CL}"
|
||||||
echo -e "${DGN}Allocated RAM: ${BGN}${RAM_SIZE}${CL}"
|
echo -e "${DGN}Allocated RAM: ${BGN}${RAM_SIZE}${CL}"
|
||||||
if ! grep -q "^iface ${BRG}" /etc/network/interfaces; then
|
if ! ip link show "${BRG}" &>/dev/null; then
|
||||||
msg_error "Bridge '${BRG}' does not exist in /etc/network/interfaces"
|
msg_error "Bridge '${BRG}' does not exist"
|
||||||
exit
|
exit
|
||||||
else
|
else
|
||||||
echo -e "${DGN}Using LAN Bridge: ${BGN}${BRG}${CL}"
|
echo -e "${DGN}Using LAN Bridge: ${BGN}${BRG}${CL}"
|
||||||
@@ -305,8 +305,8 @@ function default_settings() {
|
|||||||
if [ "$NETWORK_MODE" = "dual" ]; then
|
if [ "$NETWORK_MODE" = "dual" ]; then
|
||||||
echo -e "${DGN}Network Mode: ${BGN}Dual Interface (Firewall)${CL}"
|
echo -e "${DGN}Network Mode: ${BGN}Dual Interface (Firewall)${CL}"
|
||||||
echo -e "${DGN}Using WAN MAC Address: ${BGN}${WAN_MAC}${CL}"
|
echo -e "${DGN}Using WAN MAC Address: ${BGN}${WAN_MAC}${CL}"
|
||||||
if ! grep -q "^iface ${WAN_BRG}" /etc/network/interfaces; then
|
if ! ip link show "${WAN_BRG}" &>/dev/null; then
|
||||||
msg_error "Bridge '${WAN_BRG}' does not exist in /etc/network/interfaces"
|
msg_error "Bridge '${WAN_BRG}' does not exist"
|
||||||
exit
|
exit
|
||||||
else
|
else
|
||||||
echo -e "${DGN}Using WAN Bridge: ${BGN}${WAN_BRG}${CL}"
|
echo -e "${DGN}Using WAN Bridge: ${BGN}${WAN_BRG}${CL}"
|
||||||
@@ -424,8 +424,8 @@ function advanced_settings() {
|
|||||||
if [ -z $BRG ]; then
|
if [ -z $BRG ]; then
|
||||||
BRG="vmbr0"
|
BRG="vmbr0"
|
||||||
fi
|
fi
|
||||||
if ! grep -q "^iface ${BRG}" /etc/network/interfaces; then
|
if ! ip link show "${BRG}" &>/dev/null; then
|
||||||
msg_error "Bridge '${BRG}' does not exist in /etc/network/interfaces"
|
msg_error "Bridge '${BRG}' does not exist"
|
||||||
exit
|
exit
|
||||||
fi
|
fi
|
||||||
echo -e "${DGN}Using LAN Bridge: ${BGN}$BRG${CL}"
|
echo -e "${DGN}Using LAN Bridge: ${BGN}$BRG${CL}"
|
||||||
@@ -474,8 +474,8 @@ function advanced_settings() {
|
|||||||
if [ -z $WAN_BRG ]; then
|
if [ -z $WAN_BRG ]; then
|
||||||
WAN_BRG="vmbr1"
|
WAN_BRG="vmbr1"
|
||||||
fi
|
fi
|
||||||
if ! grep -q "^iface ${WAN_BRG}" /etc/network/interfaces; then
|
if ! ip link show "${WAN_BRG}" &>/dev/null; then
|
||||||
msg_error "WAN Bridge '${WAN_BRG}' does not exist in /etc/network/interfaces"
|
msg_error "WAN Bridge '${WAN_BRG}' does not exist"
|
||||||
exit
|
exit
|
||||||
fi
|
fi
|
||||||
echo -e "${DGN}Using WAN Bridge: ${BGN}$WAN_BRG${CL}"
|
echo -e "${DGN}Using WAN Bridge: ${BGN}$WAN_BRG${CL}"
|
||||||
|
|||||||
Reference in New Issue
Block a user