mirror of
https://github.com/community-scripts/ProxmoxVE.git
synced 2026-02-27 07:25:56 +01:00
Compare commits
8 Commits
github-act
...
ref_api
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
cfc1fc3c6b | ||
|
|
6731001360 | ||
|
|
603cba8683 | ||
|
|
7db8ddda52 | ||
|
|
6b107fc4d3 | ||
|
|
2cdeb07353 | ||
|
|
866c4062e0 | ||
|
|
10e450b72f |
152
.github/workflows/check-node-versions.yml
generated
vendored
152
.github/workflows/check-node-versions.yml
generated
vendored
@@ -13,7 +13,7 @@ permissions:
|
|||||||
jobs:
|
jobs:
|
||||||
check-node-versions:
|
check-node-versions:
|
||||||
if: github.repository == 'community-scripts/ProxmoxVE'
|
if: github.repository == 'community-scripts/ProxmoxVE'
|
||||||
runs-on: coolify-runner
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout Repository
|
- name: Checkout Repository
|
||||||
@@ -110,94 +110,22 @@ jobs:
|
|||||||
}
|
}
|
||||||
|
|
||||||
# Extract Node major from engines.node in package.json
|
# Extract Node major from engines.node in package.json
|
||||||
# Sets: ENGINES_NODE_RAW (raw string), ENGINES_MIN_MAJOR, ENGINES_IS_MINIMUM
|
# Sets: ENGINES_NODE_RAW (raw string), ENGINES_MIN_MAJOR
|
||||||
extract_engines_node() {
|
extract_engines_node() {
|
||||||
local content="$1"
|
local content="$1"
|
||||||
ENGINES_NODE_RAW=""
|
ENGINES_NODE_RAW=""
|
||||||
ENGINES_MIN_MAJOR=""
|
ENGINES_MIN_MAJOR=""
|
||||||
ENGINES_IS_MINIMUM="false"
|
|
||||||
|
|
||||||
ENGINES_NODE_RAW=$(echo "$content" | jq -r '.engines.node // empty' 2>/dev/null || echo "")
|
ENGINES_NODE_RAW=$(echo "$content" | jq -r '.engines.node // empty' 2>/dev/null || echo "")
|
||||||
if [[ -z "$ENGINES_NODE_RAW" ]]; then
|
if [[ -z "$ENGINES_NODE_RAW" ]]; then
|
||||||
return
|
return
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# Detect if constraint is a minimum (>=, ^) vs exact pinning
|
|
||||||
if [[ "$ENGINES_NODE_RAW" =~ ^(\>=|\^|\~) ]]; then
|
|
||||||
ENGINES_IS_MINIMUM="true"
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Extract the first number (major) from the constraint
|
# Extract the first number (major) from the constraint
|
||||||
# Handles: ">=24.13.1", "^22", ">=18.0.0", ">=18.15.0 <19 || ^20", etc.
|
# Handles: ">=24.13.1", "^22", ">=18.0.0", ">=18.15.0 <19 || ^20", etc.
|
||||||
ENGINES_MIN_MAJOR=$(echo "$ENGINES_NODE_RAW" | grep -oP '\d+' | head -1 || echo "")
|
ENGINES_MIN_MAJOR=$(echo "$ENGINES_NODE_RAW" | grep -oP '\d+' | head -1 || echo "")
|
||||||
}
|
}
|
||||||
|
|
||||||
# Check if our_version satisfies an engines.node constraint
|
|
||||||
# Returns 0 if satisfied, 1 if not
|
|
||||||
# Usage: version_satisfies_engines "22" ">=18.0.0" "true"
|
|
||||||
version_satisfies_engines() {
|
|
||||||
local our="$1"
|
|
||||||
local min_major="$2"
|
|
||||||
local is_minimum="$3"
|
|
||||||
|
|
||||||
if [[ -z "$min_major" || -z "$our" ]]; then
|
|
||||||
return 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [[ "$is_minimum" == "true" ]]; then
|
|
||||||
# >= or ^ constraint: our version must be >= min_major
|
|
||||||
if [[ "$our" -ge "$min_major" ]]; then
|
|
||||||
return 0
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
return 1
|
|
||||||
}
|
|
||||||
|
|
||||||
# Search for files in subdirectories via GitHub API tree
|
|
||||||
# Usage: find_repo_file "owner/repo" "branch" "filename" => sets REPLY to raw URL or empty
|
|
||||||
find_repo_file() {
|
|
||||||
local repo="$1"
|
|
||||||
local branch="$2"
|
|
||||||
local filename="$3"
|
|
||||||
REPLY=""
|
|
||||||
|
|
||||||
# Try root first (fast)
|
|
||||||
local root_url="https://raw.githubusercontent.com/${repo}/${branch}/${filename}"
|
|
||||||
if curl -sfI "$root_url" >/dev/null 2>&1; then
|
|
||||||
REPLY="$root_url"
|
|
||||||
return
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Search via GitHub API tree (recursive)
|
|
||||||
local tree_url="https://api.github.com/repos/${repo}/git/trees/${branch}?recursive=1"
|
|
||||||
local tree_json
|
|
||||||
tree_json=$(curl -sf -H "Authorization: token $GH_TOKEN" "$tree_url" 2>/dev/null || echo "")
|
|
||||||
if [[ -z "$tree_json" ]]; then
|
|
||||||
return
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Find first matching path (prefer shorter/root-level paths)
|
|
||||||
local match_path
|
|
||||||
match_path=$(echo "$tree_json" | jq -r --arg fn "$filename" \
|
|
||||||
'.tree[]? | select(.path | endswith("/" + $fn) or . == $fn) | .path' 2>/dev/null \
|
|
||||||
| sort | head -1 || echo "")
|
|
||||||
|
|
||||||
if [[ -n "$match_path" ]]; then
|
|
||||||
REPLY="https://raw.githubusercontent.com/${repo}/${branch}/${match_path}"
|
|
||||||
fi
|
|
||||||
}
|
|
||||||
|
|
||||||
# Extract Node major from .nvmrc or .node-version
|
|
||||||
# Sets: NVMRC_NODE_MAJOR
|
|
||||||
extract_nvmrc_node() {
|
|
||||||
local content="$1"
|
|
||||||
NVMRC_NODE_MAJOR=""
|
|
||||||
# .nvmrc/.node-version typically has: "v22.9.0", "22", "lts/iron", etc.
|
|
||||||
local ver
|
|
||||||
ver=$(echo "$content" | tr -d '[:space:]' | grep -oP '^v?\K[0-9]+' | head -1 || echo "")
|
|
||||||
NVMRC_NODE_MAJOR="$ver"
|
|
||||||
}
|
|
||||||
|
|
||||||
# Collect results
|
# Collect results
|
||||||
declare -a issue_scripts=()
|
declare -a issue_scripts=()
|
||||||
declare -a report_lines=()
|
declare -a report_lines=()
|
||||||
@@ -214,12 +142,8 @@ jobs:
|
|||||||
total=$((total + 1))
|
total=$((total + 1))
|
||||||
slug=$(basename "$script" | sed 's/-install\.sh$//')
|
slug=$(basename "$script" | sed 's/-install\.sh$//')
|
||||||
|
|
||||||
# Extract Source URL (GitHub only) from the "# Source:" line
|
# Extract Source URL (GitHub only)
|
||||||
# Supports both:
|
source_url=$(head -20 "$script" | grep -oP '(?<=# Source: )https://github\.com/[^\s]+' | head -1 || echo "")
|
||||||
# # Source: https://github.com/owner/repo
|
|
||||||
# # Source: https://example.com | Github: https://github.com/owner/repo
|
|
||||||
# NOTE: Must filter for "# Source:" line first to avoid matching the License URL
|
|
||||||
source_url=$(head -20 "$script" | grep -i '# Source:' | grep -oP 'https://github\.com/[^\s|]+' | head -1 || echo "")
|
|
||||||
if [[ -z "$source_url" ]]; then
|
if [[ -z "$source_url" ]]; then
|
||||||
report_lines+=("| \`$slug\` | — | — | — | — | ⏭️ No GitHub source |")
|
report_lines+=("| \`$slug\` | — | — | — | — | ⏭️ No GitHub source |")
|
||||||
continue
|
continue
|
||||||
@@ -243,23 +167,12 @@ jobs:
|
|||||||
fi
|
fi
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# Determine default branch via GitHub API (fast, single call)
|
# Fetch upstream Dockerfile
|
||||||
detected_branch=""
|
|
||||||
api_default=$(curl -sf -H "Authorization: token $GH_TOKEN" \
|
|
||||||
"https://api.github.com/repos/${repo}" 2>/dev/null \
|
|
||||||
| jq -r '.default_branch // empty' 2>/dev/null || echo "")
|
|
||||||
if [[ -n "$api_default" ]]; then
|
|
||||||
detected_branch="$api_default"
|
|
||||||
else
|
|
||||||
detected_branch="main"
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Fetch upstream Dockerfile (root + subdirectories)
|
|
||||||
df_content=""
|
df_content=""
|
||||||
find_repo_file "$repo" "$detected_branch" "Dockerfile"
|
for branch in main master dev; do
|
||||||
if [[ -n "$REPLY" ]]; then
|
df_content=$(curl -sf "https://raw.githubusercontent.com/${repo}/${branch}/Dockerfile" 2>/dev/null || echo "")
|
||||||
df_content=$(curl -sf "$REPLY" 2>/dev/null || echo "")
|
[[ -n "$df_content" ]] && break
|
||||||
fi
|
done
|
||||||
|
|
||||||
DF_NODE_MAJOR=""
|
DF_NODE_MAJOR=""
|
||||||
DF_SOURCE=""
|
DF_SOURCE=""
|
||||||
@@ -267,35 +180,19 @@ jobs:
|
|||||||
extract_dockerfile_node "$df_content"
|
extract_dockerfile_node "$df_content"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# Fetch upstream package.json (root + subdirectories)
|
# Fetch upstream package.json
|
||||||
pkg_content=""
|
pkg_content=""
|
||||||
find_repo_file "$repo" "$detected_branch" "package.json"
|
for branch in main master dev; do
|
||||||
if [[ -n "$REPLY" ]]; then
|
pkg_content=$(curl -sf "https://raw.githubusercontent.com/${repo}/${branch}/package.json" 2>/dev/null || echo "")
|
||||||
pkg_content=$(curl -sf "$REPLY" 2>/dev/null || echo "")
|
[[ -n "$pkg_content" ]] && break
|
||||||
fi
|
done
|
||||||
|
|
||||||
ENGINES_NODE_RAW=""
|
ENGINES_NODE_RAW=""
|
||||||
ENGINES_MIN_MAJOR=""
|
ENGINES_MIN_MAJOR=""
|
||||||
ENGINES_IS_MINIMUM="false"
|
|
||||||
if [[ -n "$pkg_content" ]]; then
|
if [[ -n "$pkg_content" ]]; then
|
||||||
extract_engines_node "$pkg_content"
|
extract_engines_node "$pkg_content"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# Fallback: check .nvmrc or .node-version
|
|
||||||
NVMRC_NODE_MAJOR=""
|
|
||||||
if [[ -z "$DF_NODE_MAJOR" && -z "$ENGINES_MIN_MAJOR" ]]; then
|
|
||||||
for nvmfile in .nvmrc .node-version; do
|
|
||||||
find_repo_file "$repo" "$detected_branch" "$nvmfile"
|
|
||||||
if [[ -n "$REPLY" ]]; then
|
|
||||||
nvmrc_content=$(curl -sf "$REPLY" 2>/dev/null || echo "")
|
|
||||||
if [[ -n "$nvmrc_content" ]]; then
|
|
||||||
extract_nvmrc_node "$nvmrc_content"
|
|
||||||
[[ -n "$NVMRC_NODE_MAJOR" ]] && break
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
done
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Determine upstream recommended major version
|
# Determine upstream recommended major version
|
||||||
upstream_major=""
|
upstream_major=""
|
||||||
upstream_hint=""
|
upstream_hint=""
|
||||||
@@ -306,9 +203,6 @@ jobs:
|
|||||||
elif [[ -n "$ENGINES_MIN_MAJOR" ]]; then
|
elif [[ -n "$ENGINES_MIN_MAJOR" ]]; then
|
||||||
upstream_major="$ENGINES_MIN_MAJOR"
|
upstream_major="$ENGINES_MIN_MAJOR"
|
||||||
upstream_hint="engines: $ENGINES_NODE_RAW"
|
upstream_hint="engines: $ENGINES_NODE_RAW"
|
||||||
elif [[ -n "$NVMRC_NODE_MAJOR" ]]; then
|
|
||||||
upstream_major="$NVMRC_NODE_MAJOR"
|
|
||||||
upstream_hint=".nvmrc/.node-version"
|
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# Build display values
|
# Build display values
|
||||||
@@ -320,23 +214,13 @@ jobs:
|
|||||||
if [[ "$our_version" == "dynamic" ]]; then
|
if [[ "$our_version" == "dynamic" ]]; then
|
||||||
status="🔄 Dynamic"
|
status="🔄 Dynamic"
|
||||||
elif [[ "$our_version" == "unset" ]]; then
|
elif [[ "$our_version" == "unset" ]]; then
|
||||||
if [[ -n "$upstream_major" ]]; then
|
status="⚠️ NODE_VERSION not set"
|
||||||
status="⚠️ NODE_VERSION not set (upstream=$upstream_major via $upstream_hint)"
|
|
||||||
else
|
|
||||||
status="⚠️ NODE_VERSION not set (no upstream info found)"
|
|
||||||
fi
|
|
||||||
issue_scripts+=("$slug|$our_version|$upstream_major|$upstream_hint|$repo")
|
issue_scripts+=("$slug|$our_version|$upstream_major|$upstream_hint|$repo")
|
||||||
drift_count=$((drift_count + 1))
|
drift_count=$((drift_count + 1))
|
||||||
elif [[ -n "$upstream_major" && "$our_version" != "$upstream_major" ]]; then
|
elif [[ -n "$upstream_major" && "$our_version" != "$upstream_major" ]]; then
|
||||||
# Check if engines.node is a minimum constraint that our version satisfies
|
status="🔸 Drift → upstream=$upstream_major ($upstream_hint)"
|
||||||
if [[ -z "$DF_NODE_MAJOR" && "$ENGINES_IS_MINIMUM" == "true" ]] && \
|
issue_scripts+=("$slug|$our_version|$upstream_major|$upstream_hint|$repo")
|
||||||
version_satisfies_engines "$our_version" "$ENGINES_MIN_MAJOR" "$ENGINES_IS_MINIMUM"; then
|
drift_count=$((drift_count + 1))
|
||||||
status="✅ (engines: $ENGINES_NODE_RAW — ours: $our_version satisfies)"
|
|
||||||
else
|
|
||||||
status="🔸 Drift → upstream=$upstream_major ($upstream_hint)"
|
|
||||||
issue_scripts+=("$slug|$our_version|$upstream_major|$upstream_hint|$repo")
|
|
||||||
drift_count=$((drift_count + 1))
|
|
||||||
fi
|
|
||||||
fi
|
fi
|
||||||
|
|
||||||
report_lines+=("| \`$slug\` | $our_version | $engines_display | $dockerfile_display | [$repo](https://github.com/$repo) | $status |")
|
report_lines+=("| \`$slug\` | $our_version | $engines_display | $dockerfile_display | [$repo](https://github.com/$repo) | $status |")
|
||||||
|
|||||||
119
.github/workflows/close-new-script-prs.yml
generated
vendored
119
.github/workflows/close-new-script-prs.yml
generated
vendored
@@ -1,119 +0,0 @@
|
|||||||
name: Close Unauthorized New Script PRs
|
|
||||||
|
|
||||||
on:
|
|
||||||
pull_request_target:
|
|
||||||
branches: ["main"]
|
|
||||||
types: [opened, labeled]
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
check-new-script:
|
|
||||||
if: github.repository == 'community-scripts/ProxmoxVE'
|
|
||||||
runs-on: coolify-runner
|
|
||||||
permissions:
|
|
||||||
pull-requests: write
|
|
||||||
contents: read
|
|
||||||
steps:
|
|
||||||
- name: Close PR if unauthorized new script submission
|
|
||||||
uses: actions/github-script@v7
|
|
||||||
with:
|
|
||||||
script: |
|
|
||||||
const pr = context.payload.pull_request;
|
|
||||||
const prNumber = pr.number;
|
|
||||||
const author = pr.user.login;
|
|
||||||
const authorType = pr.user.type; // "User" or "Bot"
|
|
||||||
const owner = context.repo.owner;
|
|
||||||
const repo = context.repo.repo;
|
|
||||||
|
|
||||||
// --- Only act on PRs with the "new script" label ---
|
|
||||||
const labels = pr.labels.map(l => l.name);
|
|
||||||
if (!labels.includes("new script")) {
|
|
||||||
core.info(`PR #${prNumber} does not have "new script" label — skipping.`);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
// --- Allow our bots ---
|
|
||||||
const allowedBots = [
|
|
||||||
"push-app-to-main[bot]",
|
|
||||||
"push-app-to-main",
|
|
||||||
];
|
|
||||||
|
|
||||||
if (allowedBots.includes(author)) {
|
|
||||||
core.info(`PR #${prNumber} by allowed bot "${author}" — skipping.`);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
// --- Check if author is a member of the contributor team ---
|
|
||||||
const teamSlug = "contributor";
|
|
||||||
let isMember = false;
|
|
||||||
|
|
||||||
try {
|
|
||||||
const { status } = await github.rest.teams.getMembershipForUserInOrg({
|
|
||||||
org: owner,
|
|
||||||
team_slug: teamSlug,
|
|
||||||
username: author,
|
|
||||||
});
|
|
||||||
// status 200 means the user is a member (active or pending)
|
|
||||||
isMember = true;
|
|
||||||
} catch (error) {
|
|
||||||
if (error.status === 404) {
|
|
||||||
isMember = false;
|
|
||||||
} else {
|
|
||||||
core.warning(`Could not check team membership for ${author}: ${error.message}`);
|
|
||||||
// Fallback: check org membership
|
|
||||||
try {
|
|
||||||
await github.rest.orgs.checkMembershipForUser({
|
|
||||||
org: owner,
|
|
||||||
username: author,
|
|
||||||
});
|
|
||||||
isMember = true;
|
|
||||||
} catch {
|
|
||||||
isMember = false;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (isMember) {
|
|
||||||
core.info(`PR #${prNumber} by contributor "${author}" — skipping.`);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
// --- Unauthorized: close the PR with a comment ---
|
|
||||||
core.info(`Closing PR #${prNumber} by "${author}" — not a contributor or allowed bot.`);
|
|
||||||
|
|
||||||
const comment = [
|
|
||||||
`👋 Hi @${author},`,
|
|
||||||
``,
|
|
||||||
`Thank you for your interest in contributing a new script!`,
|
|
||||||
``,
|
|
||||||
`However, **new scripts must first be submitted to our development repository** for testing and review before they can be merged here.`,
|
|
||||||
``,
|
|
||||||
`> 🛑 New scripts must be submitted to [**ProxmoxVED**](https://github.com/community-scripts/ProxmoxVED) for testing.`,
|
|
||||||
`> PRs without prior testing will be closed.`,
|
|
||||||
``,
|
|
||||||
`Please open your PR at **https://github.com/community-scripts/ProxmoxVED** instead.`,
|
|
||||||
`Once your script has been tested and approved there, it will be pushed to this repository automatically.`,
|
|
||||||
``,
|
|
||||||
`This PR will now be closed. Thank you for understanding! 🙏`,
|
|
||||||
].join("\n");
|
|
||||||
|
|
||||||
await github.rest.issues.createComment({
|
|
||||||
owner,
|
|
||||||
repo,
|
|
||||||
issue_number: prNumber,
|
|
||||||
body: comment,
|
|
||||||
});
|
|
||||||
|
|
||||||
await github.rest.pulls.update({
|
|
||||||
owner,
|
|
||||||
repo,
|
|
||||||
pull_number: prNumber,
|
|
||||||
state: "closed",
|
|
||||||
});
|
|
||||||
|
|
||||||
// Add a label to indicate why it was closed
|
|
||||||
await github.rest.issues.addLabels({
|
|
||||||
owner,
|
|
||||||
repo,
|
|
||||||
issue_number: prNumber,
|
|
||||||
labels: ["not a script issue"],
|
|
||||||
});
|
|
||||||
63
CHANGELOG.md
63
CHANGELOG.md
@@ -407,61 +407,10 @@ Exercise vigilance regarding copycat or coat-tailing sites that seek to exploit
|
|||||||
|
|
||||||
</details>
|
</details>
|
||||||
|
|
||||||
## 2026-02-27
|
|
||||||
|
|
||||||
## 2026-02-26
|
|
||||||
|
|
||||||
### 🆕 New Scripts
|
|
||||||
|
|
||||||
- Kima-Hub ([#12319](https://github.com/community-scripts/ProxmoxVE/pull/12319))
|
|
||||||
|
|
||||||
### 🚀 Updated Scripts
|
|
||||||
|
|
||||||
- #### 🐞 Bug Fixes
|
|
||||||
|
|
||||||
- tools.func: update glx alternatives / nvidia alternative if nvidia glx are missing [@MickLesk](https://github.com/MickLesk) ([#12372](https://github.com/community-scripts/ProxmoxVE/pull/12372))
|
|
||||||
- hotfix: overseer version [@CrazyWolf13](https://github.com/CrazyWolf13) ([#12366](https://github.com/community-scripts/ProxmoxVE/pull/12366))
|
|
||||||
|
|
||||||
- #### ✨ New Features
|
|
||||||
|
|
||||||
- Add ffmpeg for booklore (ffprobe) [@MickLesk](https://github.com/MickLesk) ([#12371](https://github.com/community-scripts/ProxmoxVE/pull/12371))
|
|
||||||
- [QOL] Immich: add warning regarding library compilation time [@vhsdream](https://github.com/vhsdream) ([#12345](https://github.com/community-scripts/ProxmoxVE/pull/12345))
|
|
||||||
|
|
||||||
### 🧰 Tools
|
|
||||||
|
|
||||||
- #### 🐞 Bug Fixes
|
|
||||||
|
|
||||||
- Improves adguardhome-sync addon when running on alpine LXCs [@Darkangeel-hd](https://github.com/Darkangeel-hd) ([#12362](https://github.com/community-scripts/ProxmoxVE/pull/12362))
|
|
||||||
|
|
||||||
- #### ✨ New Features
|
|
||||||
|
|
||||||
- Add Alpine support and improve Tailscale install [@MickLesk](https://github.com/MickLesk) ([#12370](https://github.com/community-scripts/ProxmoxVE/pull/12370))
|
|
||||||
|
|
||||||
### 📚 Documentation
|
|
||||||
|
|
||||||
- fix wrong link on contributions README.md [@Darkangeel-hd](https://github.com/Darkangeel-hd) ([#12363](https://github.com/community-scripts/ProxmoxVE/pull/12363))
|
|
||||||
|
|
||||||
### 📂 Github
|
|
||||||
|
|
||||||
- github: add workflow to autom. close unauthorized new-script PRs [@MickLesk](https://github.com/MickLesk) ([#12356](https://github.com/community-scripts/ProxmoxVE/pull/12356))
|
|
||||||
|
|
||||||
## 2026-02-25
|
## 2026-02-25
|
||||||
|
|
||||||
### 🆕 New Scripts
|
|
||||||
|
|
||||||
- Zerobyte ([#12321](https://github.com/community-scripts/ProxmoxVE/pull/12321))
|
|
||||||
|
|
||||||
### 🚀 Updated Scripts
|
### 🚀 Updated Scripts
|
||||||
|
|
||||||
- #### 🐞 Bug Fixes
|
|
||||||
|
|
||||||
- fix: overseer migration [@CrazyWolf13](https://github.com/CrazyWolf13) ([#12340](https://github.com/community-scripts/ProxmoxVE/pull/12340))
|
|
||||||
- add: vikunja: daemon reload [@CrazyWolf13](https://github.com/CrazyWolf13) ([#12323](https://github.com/community-scripts/ProxmoxVE/pull/12323))
|
|
||||||
- opnsense-VM: Use ip link to verify bridge existence [@MickLesk](https://github.com/MickLesk) ([#12329](https://github.com/community-scripts/ProxmoxVE/pull/12329))
|
|
||||||
- wger: Use $http_host for proxy Host header [@MickLesk](https://github.com/MickLesk) ([#12327](https://github.com/community-scripts/ProxmoxVE/pull/12327))
|
|
||||||
- Passbolt: Update Nginx config `client_max_body_size` [@tremor021](https://github.com/tremor021) ([#12313](https://github.com/community-scripts/ProxmoxVE/pull/12313))
|
|
||||||
- Zammad: configure Elasticsearch before zammad start [@MickLesk](https://github.com/MickLesk) ([#12308](https://github.com/community-scripts/ProxmoxVE/pull/12308))
|
|
||||||
|
|
||||||
- #### 🔧 Refactor
|
- #### 🔧 Refactor
|
||||||
|
|
||||||
- OpenProject: Various fixes [@tremor021](https://github.com/tremor021) ([#12246](https://github.com/community-scripts/ProxmoxVE/pull/12246))
|
- OpenProject: Various fixes [@tremor021](https://github.com/tremor021) ([#12246](https://github.com/community-scripts/ProxmoxVE/pull/12246))
|
||||||
@@ -472,18 +421,6 @@ Exercise vigilance regarding copycat or coat-tailing sites that seek to exploit
|
|||||||
|
|
||||||
- Fix detection of ssh keys [@1-tempest](https://github.com/1-tempest) ([#12230](https://github.com/community-scripts/ProxmoxVE/pull/12230))
|
- Fix detection of ssh keys [@1-tempest](https://github.com/1-tempest) ([#12230](https://github.com/community-scripts/ProxmoxVE/pull/12230))
|
||||||
|
|
||||||
- #### ✨ New Features
|
|
||||||
|
|
||||||
- tools.func: Improve GitHub/Codeberg API error handling and error output [@MickLesk](https://github.com/MickLesk) ([#12330](https://github.com/community-scripts/ProxmoxVE/pull/12330))
|
|
||||||
|
|
||||||
- #### 🔧 Refactor
|
|
||||||
|
|
||||||
- core: remove duplicate traps, consolidate error handling and harden signal traps [@MickLesk](https://github.com/MickLesk) ([#12316](https://github.com/community-scripts/ProxmoxVE/pull/12316))
|
|
||||||
|
|
||||||
### 📂 Github
|
|
||||||
|
|
||||||
- github: improvements for node drift wf [@MickLesk](https://github.com/MickLesk) ([#12309](https://github.com/community-scripts/ProxmoxVE/pull/12309))
|
|
||||||
|
|
||||||
## 2026-02-24
|
## 2026-02-24
|
||||||
|
|
||||||
### 🚀 Updated Scripts
|
### 🚀 Updated Scripts
|
||||||
|
|||||||
@@ -34,7 +34,6 @@ function update_script() {
|
|||||||
NODE_VERSION="22" setup_nodejs
|
NODE_VERSION="22" setup_nodejs
|
||||||
setup_mariadb
|
setup_mariadb
|
||||||
setup_yq
|
setup_yq
|
||||||
ensure_dependencies ffmpeg
|
|
||||||
|
|
||||||
msg_info "Stopping Service"
|
msg_info "Stopping Service"
|
||||||
systemctl stop booklore
|
systemctl stop booklore
|
||||||
|
|||||||
@@ -1,6 +0,0 @@
|
|||||||
__ __ _ __ __ __
|
|
||||||
/ //_/(_)___ ___ ____ _ / / / /_ __/ /_
|
|
||||||
/ ,< / / __ `__ \/ __ `/_____/ /_/ / / / / __ \
|
|
||||||
/ /| |/ / / / / / / /_/ /_____/ __ / /_/ / /_/ /
|
|
||||||
/_/ |_/_/_/ /_/ /_/\__,_/ /_/ /_/\__,_/_.___/
|
|
||||||
|
|
||||||
@@ -1,6 +0,0 @@
|
|||||||
_____ __ __
|
|
||||||
/__ / ___ _________ / /_ __ __/ /____
|
|
||||||
/ / / _ \/ ___/ __ \/ __ \/ / / / __/ _ \
|
|
||||||
/ /__/ __/ / / /_/ / /_/ / /_/ / /_/ __/
|
|
||||||
/____/\___/_/ \____/_.___/\__, /\__/\___/
|
|
||||||
/____/
|
|
||||||
@@ -97,7 +97,7 @@ EOF
|
|||||||
if [[ -f ~/.immich_library_revisions ]]; then
|
if [[ -f ~/.immich_library_revisions ]]; then
|
||||||
libraries=("libjxl" "libheif" "libraw" "imagemagick" "libvips")
|
libraries=("libjxl" "libheif" "libraw" "imagemagick" "libvips")
|
||||||
cd "$BASE_DIR"
|
cd "$BASE_DIR"
|
||||||
msg_warn "Checking for updates to custom image-processing libraries (recompile time: 2-15min per library)"
|
msg_info "Checking for updates to custom image-processing libraries"
|
||||||
$STD git pull
|
$STD git pull
|
||||||
for library in "${libraries[@]}"; do
|
for library in "${libraries[@]}"; do
|
||||||
compile_"$library"
|
compile_"$library"
|
||||||
|
|||||||
@@ -28,7 +28,7 @@ function update_script() {
|
|||||||
exit
|
exit
|
||||||
fi
|
fi
|
||||||
|
|
||||||
NODE_VERSION="24" NODE_MODULE="yarn,npm,pm2" setup_nodejs
|
NODE_VERSION=24 NODE_MODULE="yarn,npm,pm2" setup_nodejs
|
||||||
|
|
||||||
if check_for_gh_release "joplin-server" "laurent22/joplin"; then
|
if check_for_gh_release "joplin-server" "laurent22/joplin"; then
|
||||||
msg_info "Stopping Services"
|
msg_info "Stopping Services"
|
||||||
|
|||||||
@@ -1,79 +0,0 @@
|
|||||||
#!/usr/bin/env bash
|
|
||||||
source <(curl -fsSL https://raw.githubusercontent.com/community-scripts/ProxmoxVE/main/misc/build.func)
|
|
||||||
# Copyright (c) 2021-2026 community-scripts ORG
|
|
||||||
# Author: MickLesk (CanbiZ)
|
|
||||||
# License: MIT | https://github.com/community-scripts/ProxmoxVE/raw/main/LICENSE
|
|
||||||
# Source: https://github.com/Chevron7Locked/kima-hub
|
|
||||||
|
|
||||||
APP="Kima-Hub"
|
|
||||||
var_tags="${var_tags:-music;streaming;media}"
|
|
||||||
var_cpu="${var_cpu:-4}"
|
|
||||||
var_ram="${var_ram:-8192}"
|
|
||||||
var_disk="${var_disk:-20}"
|
|
||||||
var_os="${var_os:-debian}"
|
|
||||||
var_version="${var_version:-13}"
|
|
||||||
var_unprivileged="${var_unprivileged:-1}"
|
|
||||||
|
|
||||||
header_info "$APP"
|
|
||||||
variables
|
|
||||||
color
|
|
||||||
catch_errors
|
|
||||||
|
|
||||||
function update_script() {
|
|
||||||
header_info
|
|
||||||
check_container_storage
|
|
||||||
check_container_resources
|
|
||||||
|
|
||||||
if [[ ! -d /opt/kima-hub ]]; then
|
|
||||||
msg_error "No ${APP} Installation Found!"
|
|
||||||
exit
|
|
||||||
fi
|
|
||||||
|
|
||||||
if check_for_gh_release "kima-hub" "Chevron7Locked/kima-hub"; then
|
|
||||||
msg_info "Stopping Services"
|
|
||||||
systemctl stop kima-frontend kima-backend kima-analyzer kima-analyzer-clap
|
|
||||||
msg_ok "Stopped Services"
|
|
||||||
|
|
||||||
msg_info "Backing up Data"
|
|
||||||
cp /opt/kima-hub/backend/.env /opt/kima-hub-backend-env.bak
|
|
||||||
cp /opt/kima-hub/frontend/.env /opt/kima-hub-frontend-env.bak
|
|
||||||
msg_ok "Backed up Data"
|
|
||||||
|
|
||||||
CLEAN_INSTALL=1 fetch_and_deploy_gh_release "kima-hub" "Chevron7Locked/kima-hub" "tarball"
|
|
||||||
|
|
||||||
msg_info "Restoring Data"
|
|
||||||
cp /opt/kima-hub-backend-env.bak /opt/kima-hub/backend/.env
|
|
||||||
cp /opt/kima-hub-frontend-env.bak /opt/kima-hub/frontend/.env
|
|
||||||
rm -f /opt/kima-hub-backend-env.bak /opt/kima-hub-frontend-env.bak
|
|
||||||
msg_ok "Restored Data"
|
|
||||||
|
|
||||||
msg_info "Rebuilding Backend"
|
|
||||||
cd /opt/kima-hub/backend
|
|
||||||
$STD npm install
|
|
||||||
$STD npm run build
|
|
||||||
$STD npx prisma generate
|
|
||||||
$STD npx prisma migrate deploy
|
|
||||||
msg_ok "Rebuilt Backend"
|
|
||||||
|
|
||||||
msg_info "Rebuilding Frontend"
|
|
||||||
cd /opt/kima-hub/frontend
|
|
||||||
$STD npm install
|
|
||||||
$STD npm run build
|
|
||||||
msg_ok "Rebuilt Frontend"
|
|
||||||
|
|
||||||
msg_info "Starting Services"
|
|
||||||
systemctl start kima-backend kima-frontend kima-analyzer kima-analyzer-clap
|
|
||||||
msg_ok "Started Services"
|
|
||||||
msg_ok "Updated successfully!"
|
|
||||||
fi
|
|
||||||
exit
|
|
||||||
}
|
|
||||||
|
|
||||||
start
|
|
||||||
build_container
|
|
||||||
description
|
|
||||||
|
|
||||||
msg_ok "Completed successfully!\n"
|
|
||||||
echo -e "${CREATING}${GN}${APP} setup has been successfully initialized!${CL}"
|
|
||||||
echo -e "${INFO}${YW} Access it using the following URL:${CL}"
|
|
||||||
echo -e "${TAB}${GATEWAY}${BGN}http://${IP}:3030${CL}"
|
|
||||||
@@ -28,7 +28,7 @@ function update_script() {
|
|||||||
exit
|
exit
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [[ -f "$HOME/.overseerr" ]] && [[ "$(printf '%s\n' "1.35.0" "$(cat "$HOME/.overseerr")" | sort -V | head -n1)" == "1.35.0" ]]; then
|
if [[ -f "$HOME/.overseerr" ]] && [[ "$(cat "$HOME/.overseerr")" == "1.34.0" ]]; then
|
||||||
echo
|
echo
|
||||||
echo "━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━"
|
echo "━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━"
|
||||||
echo "Overseerr v1.34.0 detected."
|
echo "Overseerr v1.34.0 detected."
|
||||||
|
|||||||
@@ -65,7 +65,6 @@ function update_script() {
|
|||||||
msg_ok "Stopped Service"
|
msg_ok "Stopped Service"
|
||||||
|
|
||||||
fetch_and_deploy_gh_release "vikunja" "go-vikunja/vikunja" "binary"
|
fetch_and_deploy_gh_release "vikunja" "go-vikunja/vikunja" "binary"
|
||||||
$STD systemctl daemon-reload
|
|
||||||
|
|
||||||
msg_info "Starting Service"
|
msg_info "Starting Service"
|
||||||
systemctl start vikunja
|
systemctl start vikunja
|
||||||
|
|||||||
@@ -1,71 +0,0 @@
|
|||||||
#!/usr/bin/env bash
|
|
||||||
source <(curl -fsSL https://raw.githubusercontent.com/community-scripts/ProxmoxVE/main/misc/build.func)
|
|
||||||
# Copyright (c) 2021-2026 community-scripts ORG
|
|
||||||
# Author: community-scripts
|
|
||||||
# License: MIT | https://github.com/community-scripts/ProxmoxVE/raw/main/LICENSE
|
|
||||||
# Source: https://github.com/nicotsx/zerobyte
|
|
||||||
|
|
||||||
APP="Zerobyte"
|
|
||||||
var_tags="${var_tags:-backup;encryption;restic}"
|
|
||||||
var_cpu="${var_cpu:-2}"
|
|
||||||
var_ram="${var_ram:-6144}"
|
|
||||||
var_disk="${var_disk:-10}"
|
|
||||||
var_os="${var_os:-debian}"
|
|
||||||
var_version="${var_version:-13}"
|
|
||||||
var_unprivileged="${var_unprivileged:-1}"
|
|
||||||
|
|
||||||
header_info "$APP"
|
|
||||||
variables
|
|
||||||
color
|
|
||||||
catch_errors
|
|
||||||
|
|
||||||
function update_script() {
|
|
||||||
header_info
|
|
||||||
check_container_storage
|
|
||||||
check_container_resources
|
|
||||||
|
|
||||||
if [[ ! -d /opt/zerobyte ]]; then
|
|
||||||
msg_error "No ${APP} Installation Found!"
|
|
||||||
exit
|
|
||||||
fi
|
|
||||||
|
|
||||||
if check_for_gh_release "zerobyte" "nicotsx/zerobyte"; then
|
|
||||||
msg_info "Stopping Service"
|
|
||||||
systemctl stop zerobyte
|
|
||||||
msg_ok "Stopped Service"
|
|
||||||
|
|
||||||
msg_info "Backing up Configuration"
|
|
||||||
cp /opt/zerobyte/.env /opt/zerobyte.env.bak
|
|
||||||
msg_ok "Backed up Configuration"
|
|
||||||
|
|
||||||
NODE_VERSION="24" setup_nodejs
|
|
||||||
CLEAN_INSTALL=1 fetch_and_deploy_gh_release "zerobyte" "nicotsx/zerobyte" "tarball"
|
|
||||||
|
|
||||||
msg_info "Building Zerobyte"
|
|
||||||
export NODE_OPTIONS="--max-old-space-size=3072"
|
|
||||||
cd /opt/zerobyte
|
|
||||||
$STD bun install
|
|
||||||
$STD node ./node_modules/vite/bin/vite.js build
|
|
||||||
msg_ok "Built Zerobyte"
|
|
||||||
|
|
||||||
msg_info "Restoring Configuration"
|
|
||||||
cp /opt/zerobyte.env.bak /opt/zerobyte/.env
|
|
||||||
rm -f /opt/zerobyte.env.bak
|
|
||||||
msg_ok "Restored Configuration"
|
|
||||||
|
|
||||||
msg_info "Starting Service"
|
|
||||||
systemctl start zerobyte
|
|
||||||
msg_ok "Started Service"
|
|
||||||
msg_ok "Updated successfully!"
|
|
||||||
fi
|
|
||||||
exit
|
|
||||||
}
|
|
||||||
|
|
||||||
start
|
|
||||||
build_container
|
|
||||||
description
|
|
||||||
|
|
||||||
msg_ok "Completed Successfully!\n"
|
|
||||||
echo -e "${CREATING}${GN}${APP} setup has been successfully initialized!${CL}"
|
|
||||||
echo -e "${INFO}${YW} Access it using the following URL:${CL}"
|
|
||||||
echo -e "${TAB}${GATEWAY}${BGN}http://${IP}:4096${CL}"
|
|
||||||
@@ -29,7 +29,7 @@ function update_script() {
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
if check_for_gh_release "Zigbee2MQTT" "Koenkk/zigbee2mqtt"; then
|
if check_for_gh_release "Zigbee2MQTT" "Koenkk/zigbee2mqtt"; then
|
||||||
NODE_VERSION="24" NODE_MODULE="pnpm@$(curl -fsSL https://raw.githubusercontent.com/Koenkk/zigbee2mqtt/master/package.json | jq -r '.packageManager | split("@")[1]')" setup_nodejs
|
NODE_VERSION=24 NODE_MODULE="pnpm@$(curl -fsSL https://raw.githubusercontent.com/Koenkk/zigbee2mqtt/master/package.json | jq -r '.packageManager | split("@")[1]')" setup_nodejs
|
||||||
msg_info "Stopping Service"
|
msg_info "Stopping Service"
|
||||||
systemctl stop zigbee2mqtt
|
systemctl stop zigbee2mqtt
|
||||||
msg_ok "Stopped Service"
|
msg_ok "Stopped Service"
|
||||||
|
|||||||
@@ -175,7 +175,7 @@ All scripts and configurations must follow our coding standards to ensure consis
|
|||||||
### Available Guides
|
### Available Guides
|
||||||
|
|
||||||
- **[CONTRIBUTING.md](CONTRIBUTING.md)** - Essential coding standards and best practices
|
- **[CONTRIBUTING.md](CONTRIBUTING.md)** - Essential coding standards and best practices
|
||||||
- **[CODE-AUDIT.md](CODE-AUDIT.md)** - Code review checklist and audit procedures
|
- **[CODE_AUDIT.md](CODE_AUDIT.md)** - Code review checklist and audit procedures
|
||||||
- **[GUIDE.md](GUIDE.md)** - Comprehensive contribution guide
|
- **[GUIDE.md](GUIDE.md)** - Comprehensive contribution guide
|
||||||
- **[HELPER_FUNCTIONS.md](HELPER_FUNCTIONS.md)** - Reference for all tools.func helper functions
|
- **[HELPER_FUNCTIONS.md](HELPER_FUNCTIONS.md)** - Reference for all tools.func helper functions
|
||||||
- **Container Scripts** - `/ct/` templates and guidelines
|
- **Container Scripts** - `/ct/` templates and guidelines
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
{
|
{
|
||||||
"generated": "2026-02-27T00:22:57Z",
|
"generated": "2026-02-25T06:25:10Z",
|
||||||
"versions": [
|
"versions": [
|
||||||
{
|
{
|
||||||
"slug": "2fauth",
|
"slug": "2fauth",
|
||||||
@@ -39,9 +39,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "ampache",
|
"slug": "ampache",
|
||||||
"repo": "ampache/ampache",
|
"repo": "ampache/ampache",
|
||||||
"version": "7.9.1",
|
"version": "7.9.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-25T08:52:58Z"
|
"date": "2026-02-19T07:01:25Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "argus",
|
"slug": "argus",
|
||||||
@@ -109,9 +109,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "bazarr",
|
"slug": "bazarr",
|
||||||
"repo": "morpheus65535/bazarr",
|
"repo": "morpheus65535/bazarr",
|
||||||
"version": "v1.5.6",
|
"version": "v1.5.5",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T11:33:11Z"
|
"date": "2026-02-01T18:00:34Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "bentopdf",
|
"slug": "bentopdf",
|
||||||
@@ -151,9 +151,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "booklore",
|
"slug": "booklore",
|
||||||
"repo": "booklore-app/BookLore",
|
"repo": "booklore-app/BookLore",
|
||||||
"version": "v2.0.3",
|
"version": "v2.0.1",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T19:10:59Z"
|
"date": "2026-02-24T04:15:33Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "bookstack",
|
"slug": "bookstack",
|
||||||
@@ -214,9 +214,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "comfyui",
|
"slug": "comfyui",
|
||||||
"repo": "comfyanonymous/ComfyUI",
|
"repo": "comfyanonymous/ComfyUI",
|
||||||
"version": "v0.15.1",
|
"version": "v0.15.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T22:01:35Z"
|
"date": "2026-02-24T20:56:09Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "commafeed",
|
"slug": "commafeed",
|
||||||
@@ -242,9 +242,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "cosmos",
|
"slug": "cosmos",
|
||||||
"repo": "azukaar/Cosmos-Server",
|
"repo": "azukaar/Cosmos-Server",
|
||||||
"version": "v0.21.3",
|
"version": "v0.20.2",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T20:04:35Z"
|
"date": "2026-01-24T00:12:39Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "cronicle",
|
"slug": "cronicle",
|
||||||
@@ -270,16 +270,16 @@
|
|||||||
{
|
{
|
||||||
"slug": "databasus",
|
"slug": "databasus",
|
||||||
"repo": "databasus/databasus",
|
"repo": "databasus/databasus",
|
||||||
"version": "v3.16.3",
|
"version": "v3.16.2",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-25T19:57:26Z"
|
"date": "2026-02-22T21:10:12Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "dawarich",
|
"slug": "dawarich",
|
||||||
"repo": "Freika/dawarich",
|
"repo": "Freika/dawarich",
|
||||||
"version": "1.3.0",
|
"version": "1.2.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-25T19:30:25Z"
|
"date": "2026-02-15T22:33:56Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "discopanel",
|
"slug": "discopanel",
|
||||||
@@ -291,9 +291,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "dispatcharr",
|
"slug": "dispatcharr",
|
||||||
"repo": "Dispatcharr/Dispatcharr",
|
"repo": "Dispatcharr/Dispatcharr",
|
||||||
"version": "v0.20.1",
|
"version": "v0.19.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T21:38:19Z"
|
"date": "2026-02-10T21:18:10Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "docmost",
|
"slug": "docmost",
|
||||||
@@ -452,9 +452,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "gitea-mirror",
|
"slug": "gitea-mirror",
|
||||||
"repo": "RayLabsHQ/gitea-mirror",
|
"repo": "RayLabsHQ/gitea-mirror",
|
||||||
"version": "v3.9.5",
|
"version": "v3.9.4",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T05:32:12Z"
|
"date": "2026-02-24T06:17:56Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "glance",
|
"slug": "glance",
|
||||||
@@ -606,16 +606,16 @@
|
|||||||
{
|
{
|
||||||
"slug": "invoiceninja",
|
"slug": "invoiceninja",
|
||||||
"repo": "invoiceninja/invoiceninja",
|
"repo": "invoiceninja/invoiceninja",
|
||||||
"version": "v5.12.69",
|
"version": "v5.12.66",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T22:23:32Z"
|
"date": "2026-02-24T09:12:50Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "jackett",
|
"slug": "jackett",
|
||||||
"repo": "Jackett/Jackett",
|
"repo": "Jackett/Jackett",
|
||||||
"version": "v0.24.1218",
|
"version": "v0.24.1205",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T05:55:11Z"
|
"date": "2026-02-25T05:49:14Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "jellystat",
|
"slug": "jellystat",
|
||||||
@@ -627,9 +627,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "joplin-server",
|
"slug": "joplin-server",
|
||||||
"repo": "laurent22/joplin",
|
"repo": "laurent22/joplin",
|
||||||
"version": "v3.5.13",
|
"version": "v3.5.12",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-25T21:19:11Z"
|
"date": "2026-01-17T14:20:33Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "jotty",
|
"slug": "jotty",
|
||||||
@@ -666,19 +666,12 @@
|
|||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-20T09:19:45Z"
|
"date": "2026-02-20T09:19:45Z"
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"slug": "kima-hub",
|
|
||||||
"repo": "Chevron7Locked/kima-hub",
|
|
||||||
"version": "v1.5.7",
|
|
||||||
"pinned": false,
|
|
||||||
"date": "2026-02-23T23:58:59Z"
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"slug": "kimai",
|
"slug": "kimai",
|
||||||
"repo": "kimai/kimai",
|
"repo": "kimai/kimai",
|
||||||
"version": "2.50.0",
|
"version": "2.49.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-25T20:13:51Z"
|
"date": "2026-02-15T20:40:19Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "kitchenowl",
|
"slug": "kitchenowl",
|
||||||
@@ -718,9 +711,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "kubo",
|
"slug": "kubo",
|
||||||
"repo": "ipfs/kubo",
|
"repo": "ipfs/kubo",
|
||||||
"version": "v0.40.0",
|
"version": "v0.39.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-25T23:16:17Z"
|
"date": "2025-11-27T03:47:38Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "kutt",
|
"slug": "kutt",
|
||||||
@@ -795,9 +788,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "lubelogger",
|
"slug": "lubelogger",
|
||||||
"repo": "hargata/lubelog",
|
"repo": "hargata/lubelog",
|
||||||
"version": "v1.6.1",
|
"version": "v1.6.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T20:01:24Z"
|
"date": "2026-02-10T20:16:32Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "mafl",
|
"slug": "mafl",
|
||||||
@@ -816,9 +809,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "mail-archiver",
|
"slug": "mail-archiver",
|
||||||
"repo": "s1t5/mail-archiver",
|
"repo": "s1t5/mail-archiver",
|
||||||
"version": "2602.4",
|
"version": "2602.3",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T08:43:01Z"
|
"date": "2026-02-22T20:24:18Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "managemydamnlife",
|
"slug": "managemydamnlife",
|
||||||
@@ -830,9 +823,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "manyfold",
|
"slug": "manyfold",
|
||||||
"repo": "manyfold3d/manyfold",
|
"repo": "manyfold3d/manyfold",
|
||||||
"version": "v0.133.1",
|
"version": "v0.132.1",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T15:50:34Z"
|
"date": "2026-02-09T22:02:28Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "mealie",
|
"slug": "mealie",
|
||||||
@@ -970,9 +963,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "oauth2-proxy",
|
"slug": "oauth2-proxy",
|
||||||
"repo": "oauth2-proxy/oauth2-proxy",
|
"repo": "oauth2-proxy/oauth2-proxy",
|
||||||
"version": "v7.14.3",
|
"version": "v7.14.2",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T14:10:21Z"
|
"date": "2026-01-18T00:26:09Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "ombi",
|
"slug": "ombi",
|
||||||
@@ -1002,13 +995,6 @@
|
|||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-03T09:00:43Z"
|
"date": "2026-02-03T09:00:43Z"
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"slug": "openproject",
|
|
||||||
"repo": "jemalloc/jemalloc",
|
|
||||||
"version": "5.3.0",
|
|
||||||
"pinned": false,
|
|
||||||
"date": "2022-05-06T19:14:21Z"
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"slug": "ots",
|
"slug": "ots",
|
||||||
"repo": "Luzifer/ots",
|
"repo": "Luzifer/ots",
|
||||||
@@ -1061,9 +1047,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "paperless-gpt",
|
"slug": "paperless-gpt",
|
||||||
"repo": "icereed/paperless-gpt",
|
"repo": "icereed/paperless-gpt",
|
||||||
"version": "v0.25.1",
|
"version": "v0.25.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T14:50:11Z"
|
"date": "2026-02-16T08:31:48Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "paperless-ngx",
|
"slug": "paperless-ngx",
|
||||||
@@ -1173,9 +1159,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "prometheus",
|
"slug": "prometheus",
|
||||||
"repo": "prometheus/prometheus",
|
"repo": "prometheus/prometheus",
|
||||||
"version": "v3.10.0",
|
"version": "v3.9.1",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T01:19:51Z"
|
"date": "2026-01-07T17:05:53Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "prometheus-alertmanager",
|
"slug": "prometheus-alertmanager",
|
||||||
@@ -1271,9 +1257,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "radicale",
|
"slug": "radicale",
|
||||||
"repo": "Kozea/Radicale",
|
"repo": "Kozea/Radicale",
|
||||||
"version": "v3.6.1",
|
"version": "v3.6.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-24T06:36:23Z"
|
"date": "2026-01-10T06:56:46Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "rclone",
|
"slug": "rclone",
|
||||||
@@ -1299,9 +1285,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "recyclarr",
|
"slug": "recyclarr",
|
||||||
"repo": "recyclarr/recyclarr",
|
"repo": "recyclarr/recyclarr",
|
||||||
"version": "v8.3.2",
|
"version": "v8.3.1",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-25T22:39:51Z"
|
"date": "2026-02-25T01:01:31Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "reitti",
|
"slug": "reitti",
|
||||||
@@ -1397,9 +1383,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "signoz",
|
"slug": "signoz",
|
||||||
"repo": "SigNoz/signoz-otel-collector",
|
"repo": "SigNoz/signoz-otel-collector",
|
||||||
"version": "v0.144.2",
|
"version": "v0.144.1",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T05:57:26Z"
|
"date": "2026-02-25T05:57:17Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "silverbullet",
|
"slug": "silverbullet",
|
||||||
@@ -1607,9 +1593,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "tunarr",
|
"slug": "tunarr",
|
||||||
"repo": "chrisbenincasa/tunarr",
|
"repo": "chrisbenincasa/tunarr",
|
||||||
"version": "v1.1.18",
|
"version": "v1.1.16",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-26T22:09:44Z"
|
"date": "2026-02-23T21:24:47Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "uhf",
|
"slug": "uhf",
|
||||||
@@ -1670,9 +1656,9 @@
|
|||||||
{
|
{
|
||||||
"slug": "vikunja",
|
"slug": "vikunja",
|
||||||
"repo": "go-vikunja/vikunja",
|
"repo": "go-vikunja/vikunja",
|
||||||
"version": "v2.0.0",
|
"version": "v1.1.0",
|
||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-25T13:58:47Z"
|
"date": "2026-02-09T10:34:29Z"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"slug": "wallabag",
|
"slug": "wallabag",
|
||||||
@@ -1786,13 +1772,6 @@
|
|||||||
"pinned": false,
|
"pinned": false,
|
||||||
"date": "2026-02-24T15:15:46Z"
|
"date": "2026-02-24T15:15:46Z"
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"slug": "zerobyte",
|
|
||||||
"repo": "restic/restic",
|
|
||||||
"version": "v0.18.1",
|
|
||||||
"pinned": false,
|
|
||||||
"date": "2025-09-21T18:24:38Z"
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"slug": "zigbee2mqtt",
|
"slug": "zigbee2mqtt",
|
||||||
"repo": "Koenkk/zigbee2mqtt",
|
"repo": "Koenkk/zigbee2mqtt",
|
||||||
|
|||||||
@@ -51,10 +51,6 @@
|
|||||||
{
|
{
|
||||||
"text": "Logs: `/var/log/immich`",
|
"text": "Logs: `/var/log/immich`",
|
||||||
"type": "info"
|
"type": "info"
|
||||||
},
|
|
||||||
{
|
|
||||||
"text": "During first install, 5 custom libraries need to be compiled from source. Depending on your CPU, this can take anywhere between 15 minutes and 2 hours. Please be patient. Touch grass or something.",
|
|
||||||
"type": "warning"
|
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,48 +0,0 @@
|
|||||||
{
|
|
||||||
"name": "Kima-Hub",
|
|
||||||
"slug": "kima-hub",
|
|
||||||
"categories": [
|
|
||||||
13
|
|
||||||
],
|
|
||||||
"date_created": "2026-02-26",
|
|
||||||
"type": "ct",
|
|
||||||
"updateable": true,
|
|
||||||
"privileged": false,
|
|
||||||
"interface_port": 3030,
|
|
||||||
"documentation": "https://github.com/Chevron7Locked/kima-hub#readme",
|
|
||||||
"website": "https://github.com/Chevron7Locked/kima-hub",
|
|
||||||
"logo": "https://cdn.jsdelivr.net/gh/selfhst/icons@main/webp/kima-hub.webp",
|
|
||||||
"config_path": "/opt/kima-hub/backend/.env",
|
|
||||||
"description": "Self-hosted, on-demand audio streaming platform with AI-powered vibe matching, mood detection, smart playlists, and Lidarr/Audiobookshelf integration.",
|
|
||||||
"install_methods": [
|
|
||||||
{
|
|
||||||
"type": "default",
|
|
||||||
"script": "ct/kima-hub.sh",
|
|
||||||
"resources": {
|
|
||||||
"cpu": 4,
|
|
||||||
"ram": 8192,
|
|
||||||
"hdd": 20,
|
|
||||||
"os": "Debian",
|
|
||||||
"version": "13"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"default_credentials": {
|
|
||||||
"username": null,
|
|
||||||
"password": null
|
|
||||||
},
|
|
||||||
"notes": [
|
|
||||||
{
|
|
||||||
"text": "First user to register becomes the administrator.",
|
|
||||||
"type": "info"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"text": "Mount your music library to /music in the container.",
|
|
||||||
"type": "warning"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"text": "Audio analysis (mood/vibe detection) requires significant RAM (2-4GB per worker).",
|
|
||||||
"type": "info"
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
@@ -1,40 +0,0 @@
|
|||||||
{
|
|
||||||
"name": "Zerobyte",
|
|
||||||
"slug": "zerobyte",
|
|
||||||
"categories": [
|
|
||||||
7
|
|
||||||
],
|
|
||||||
"date_created": "2026-02-25",
|
|
||||||
"type": "ct",
|
|
||||||
"updateable": true,
|
|
||||||
"privileged": false,
|
|
||||||
"interface_port": 4096,
|
|
||||||
"documentation": "https://github.com/nicotsx/zerobyte#readme",
|
|
||||||
"website": "https://github.com/nicotsx/zerobyte",
|
|
||||||
"logo": "https://cdn.jsdelivr.net/gh/selfhst/icons@main/webp/zerobyte.webp",
|
|
||||||
"config_path": "/opt/zerobyte/.env",
|
|
||||||
"description": "Zerobyte is a backup automation tool built on top of Restic that provides a modern web interface to schedule, manage, and monitor encrypted backups across multiple storage backends including NFS, SMB, WebDAV, SFTP, S3, and local directories.",
|
|
||||||
"install_methods": [
|
|
||||||
{
|
|
||||||
"type": "default",
|
|
||||||
"script": "ct/zerobyte.sh",
|
|
||||||
"resources": {
|
|
||||||
"cpu": 2,
|
|
||||||
"ram": 6144,
|
|
||||||
"hdd": 10,
|
|
||||||
"os": "Debian",
|
|
||||||
"version": "13"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"default_credentials": {
|
|
||||||
"username": null,
|
|
||||||
"password": null
|
|
||||||
},
|
|
||||||
"notes": [
|
|
||||||
{
|
|
||||||
"text": "For remote mount support (NFS, SMB, WebDAV, SFTP), enable FUSE device passthrough on the LXC container. (FUSE is pre-configured)",
|
|
||||||
"type": "info"
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
@@ -13,10 +13,6 @@ setting_up_container
|
|||||||
network_check
|
network_check
|
||||||
update_os
|
update_os
|
||||||
|
|
||||||
msg_info "Installing Dependencies"
|
|
||||||
$STD apt install -y ffmpeg
|
|
||||||
msg_ok "Installed Dependencies"
|
|
||||||
|
|
||||||
JAVA_VERSION="25" setup_java
|
JAVA_VERSION="25" setup_java
|
||||||
NODE_VERSION="22" setup_nodejs
|
NODE_VERSION="22" setup_nodejs
|
||||||
setup_mariadb
|
setup_mariadb
|
||||||
|
|||||||
@@ -39,7 +39,7 @@ $STD apt install -y \
|
|||||||
texlive-xetex
|
texlive-xetex
|
||||||
msg_ok "Installed Dependencies"
|
msg_ok "Installed Dependencies"
|
||||||
|
|
||||||
NODE_VERSION="22" NODE_MODULE="bun" setup_nodejs
|
NODE_VERSION=22 NODE_MODULE="bun" setup_nodejs
|
||||||
fetch_and_deploy_gh_release "ConvertX" "C4illin/ConvertX" "tarball" "latest" "/opt/convertx"
|
fetch_and_deploy_gh_release "ConvertX" "C4illin/ConvertX" "tarball" "latest" "/opt/convertx"
|
||||||
|
|
||||||
msg_info "Installing ConvertX"
|
msg_info "Installing ConvertX"
|
||||||
|
|||||||
@@ -154,7 +154,7 @@ sed -i -e "/^#shared_preload/s/^#//;/^shared_preload/s/''/'vchord.so'/" /etc/pos
|
|||||||
systemctl restart postgresql.service
|
systemctl restart postgresql.service
|
||||||
PG_DB_NAME="immich" PG_DB_USER="immich" PG_DB_GRANT_SUPERUSER="true" PG_DB_SKIP_ALTER_ROLE="true" setup_postgresql_db
|
PG_DB_NAME="immich" PG_DB_USER="immich" PG_DB_GRANT_SUPERUSER="true" PG_DB_SKIP_ALTER_ROLE="true" setup_postgresql_db
|
||||||
|
|
||||||
msg_warn "Compiling Custom Photo-processing Libraries (can take anywhere from 15min to 2h)"
|
msg_info "Compiling Custom Photo-processing Library (extreme patience)"
|
||||||
LD_LIBRARY_PATH=/usr/local/lib
|
LD_LIBRARY_PATH=/usr/local/lib
|
||||||
export LD_RUN_PATH=/usr/local/lib
|
export LD_RUN_PATH=/usr/local/lib
|
||||||
STAGING_DIR=/opt/staging
|
STAGING_DIR=/opt/staging
|
||||||
|
|||||||
@@ -21,7 +21,7 @@ msg_ok "Installed Dependencies"
|
|||||||
|
|
||||||
PG_VERSION="17" setup_postgresql
|
PG_VERSION="17" setup_postgresql
|
||||||
PG_DB_NAME="joplin" PG_DB_USER="joplin" setup_postgresql_db
|
PG_DB_NAME="joplin" PG_DB_USER="joplin" setup_postgresql_db
|
||||||
NODE_VERSION="24" NODE_MODULE="yarn,npm,pm2" setup_nodejs
|
NODE_VERSION=24 NODE_MODULE="yarn,npm,pm2" setup_nodejs
|
||||||
mkdir -p /opt/pm2
|
mkdir -p /opt/pm2
|
||||||
export PM2_HOME=/opt/pm2
|
export PM2_HOME=/opt/pm2
|
||||||
$STD pm2 install pm2-logrotate
|
$STD pm2 install pm2-logrotate
|
||||||
|
|||||||
@@ -1,212 +0,0 @@
|
|||||||
#!/usr/bin/env bash
|
|
||||||
|
|
||||||
# Copyright (c) 2021-2026 community-scripts ORG
|
|
||||||
# Author: MickLesk (CanbiZ)
|
|
||||||
# License: MIT | https://github.com/community-scripts/ProxmoxVE/raw/main/LICENSE
|
|
||||||
# Source: https://github.com/Chevron7Locked/kima-hub
|
|
||||||
|
|
||||||
source /dev/stdin <<<"$FUNCTIONS_FILE_PATH"
|
|
||||||
color
|
|
||||||
verb_ip6
|
|
||||||
catch_errors
|
|
||||||
setting_up_container
|
|
||||||
network_check
|
|
||||||
update_os
|
|
||||||
|
|
||||||
msg_info "Installing Dependencies"
|
|
||||||
$STD apt-get install -y \
|
|
||||||
build-essential \
|
|
||||||
git \
|
|
||||||
openssl \
|
|
||||||
ffmpeg \
|
|
||||||
python3 \
|
|
||||||
python3-pip \
|
|
||||||
python3-dev \
|
|
||||||
python3-numpy \
|
|
||||||
redis-server
|
|
||||||
msg_ok "Installed Dependencies"
|
|
||||||
|
|
||||||
PG_VERSION="16" PG_MODULES="pgvector" setup_postgresql
|
|
||||||
PG_DB_NAME="kima" PG_DB_USER="kima" PG_DB_GRANT_SUPERUSER="true" setup_postgresql_db
|
|
||||||
NODE_VERSION="20" setup_nodejs
|
|
||||||
|
|
||||||
msg_info "Configuring Redis"
|
|
||||||
systemctl enable -q --now redis-server
|
|
||||||
msg_ok "Configured Redis"
|
|
||||||
|
|
||||||
fetch_and_deploy_gh_release "kima-hub" "Chevron7Locked/kima-hub" "tarball"
|
|
||||||
|
|
||||||
msg_info "Installing Python Dependencies"
|
|
||||||
export PIP_BREAK_SYSTEM_PACKAGES=1
|
|
||||||
$STD pip3 install --no-cache-dir \
|
|
||||||
tensorflow \
|
|
||||||
essentia-tensorflow \
|
|
||||||
redis \
|
|
||||||
psycopg2-binary \
|
|
||||||
laion-clap \
|
|
||||||
torch \
|
|
||||||
torchaudio \
|
|
||||||
librosa \
|
|
||||||
transformers \
|
|
||||||
pgvector \
|
|
||||||
python-dotenv \
|
|
||||||
requests
|
|
||||||
msg_ok "Installed Python Dependencies"
|
|
||||||
|
|
||||||
msg_info "Downloading Essentia ML Models"
|
|
||||||
mkdir -p /opt/kima-hub/models
|
|
||||||
cd /opt/kima-hub/models
|
|
||||||
curl -fsSL -o msd-musicnn-1.pb "https://essentia.upf.edu/models/autotagging/msd/msd-musicnn-1.pb"
|
|
||||||
curl -fsSL -o mood_happy-msd-musicnn-1.pb "https://essentia.upf.edu/models/classification-heads/mood_happy/mood_happy-msd-musicnn-1.pb"
|
|
||||||
curl -fsSL -o mood_sad-msd-musicnn-1.pb "https://essentia.upf.edu/models/classification-heads/mood_sad/mood_sad-msd-musicnn-1.pb"
|
|
||||||
curl -fsSL -o mood_relaxed-msd-musicnn-1.pb "https://essentia.upf.edu/models/classification-heads/mood_relaxed/mood_relaxed-msd-musicnn-1.pb"
|
|
||||||
curl -fsSL -o mood_aggressive-msd-musicnn-1.pb "https://essentia.upf.edu/models/classification-heads/mood_aggressive/mood_aggressive-msd-musicnn-1.pb"
|
|
||||||
curl -fsSL -o mood_party-msd-musicnn-1.pb "https://essentia.upf.edu/models/classification-heads/mood_party/mood_party-msd-musicnn-1.pb"
|
|
||||||
curl -fsSL -o mood_acoustic-msd-musicnn-1.pb "https://essentia.upf.edu/models/classification-heads/mood_acoustic/mood_acoustic-msd-musicnn-1.pb"
|
|
||||||
curl -fsSL -o mood_electronic-msd-musicnn-1.pb "https://essentia.upf.edu/models/classification-heads/mood_electronic/mood_electronic-msd-musicnn-1.pb"
|
|
||||||
curl -fsSL -o danceability-msd-musicnn-1.pb "https://essentia.upf.edu/models/classification-heads/danceability/danceability-msd-musicnn-1.pb"
|
|
||||||
curl -fsSL -o voice_instrumental-msd-musicnn-1.pb "https://essentia.upf.edu/models/classification-heads/voice_instrumental/voice_instrumental-msd-musicnn-1.pb"
|
|
||||||
msg_ok "Downloaded Essentia ML Models"
|
|
||||||
|
|
||||||
msg_info "Downloading CLAP Model"
|
|
||||||
curl -fsSL -o /opt/kima-hub/models/music_audioset_epoch_15_esc_90.14.pt "https://huggingface.co/lukewys/laion_clap/resolve/main/music_audioset_epoch_15_esc_90.14.pt"
|
|
||||||
msg_ok "Downloaded CLAP Model"
|
|
||||||
|
|
||||||
msg_info "Building Backend"
|
|
||||||
cd /opt/kima-hub/backend
|
|
||||||
$STD npm ci
|
|
||||||
$STD npm run build
|
|
||||||
msg_ok "Built Backend"
|
|
||||||
|
|
||||||
msg_info "Configuring Backend"
|
|
||||||
SESSION_SECRET=$(openssl rand -hex 32)
|
|
||||||
ENCRYPTION_KEY=$(openssl rand -hex 32)
|
|
||||||
cat <<EOF >/opt/kima-hub/backend/.env
|
|
||||||
NODE_ENV=production
|
|
||||||
DATABASE_URL=postgresql://${PG_DB_USER}:${PG_DB_PASS}@localhost:5432/${PG_DB_NAME}
|
|
||||||
REDIS_URL=redis://localhost:6379
|
|
||||||
PORT=3006
|
|
||||||
MUSIC_PATH=/music
|
|
||||||
TRANSCODE_CACHE_PATH=/opt/kima-hub/cache/transcodes
|
|
||||||
SESSION_SECRET=${SESSION_SECRET}
|
|
||||||
SETTINGS_ENCRYPTION_KEY=${ENCRYPTION_KEY}
|
|
||||||
INTERNAL_API_SECRET=$(openssl rand -hex 16)
|
|
||||||
EOF
|
|
||||||
msg_ok "Configured Backend"
|
|
||||||
|
|
||||||
msg_info "Running Database Migrations"
|
|
||||||
cd /opt/kima-hub/backend
|
|
||||||
$STD npx prisma generate
|
|
||||||
$STD npx prisma migrate deploy
|
|
||||||
msg_ok "Ran Database Migrations"
|
|
||||||
|
|
||||||
msg_info "Building Frontend"
|
|
||||||
cd /opt/kima-hub/frontend
|
|
||||||
$STD npm ci
|
|
||||||
export NEXT_PUBLIC_BACKEND_URL=http://127.0.0.1:3006
|
|
||||||
$STD npm run build
|
|
||||||
msg_ok "Built Frontend"
|
|
||||||
|
|
||||||
msg_info "Configuring Frontend"
|
|
||||||
cat <<EOF >/opt/kima-hub/frontend/.env
|
|
||||||
NODE_ENV=production
|
|
||||||
BACKEND_URL=http://localhost:3006
|
|
||||||
PORT=3030
|
|
||||||
EOF
|
|
||||||
msg_ok "Configured Frontend"
|
|
||||||
|
|
||||||
msg_info "Creating Directories"
|
|
||||||
mkdir -p /opt/kima-hub/cache/transcodes
|
|
||||||
mkdir -p /music
|
|
||||||
msg_ok "Created Directories"
|
|
||||||
|
|
||||||
msg_info "Creating Services"
|
|
||||||
cat <<EOF >/etc/systemd/system/kima-backend.service
|
|
||||||
[Unit]
|
|
||||||
Description=Kima Hub Backend
|
|
||||||
After=network.target postgresql.service redis-server.service
|
|
||||||
Wants=postgresql.service redis-server.service
|
|
||||||
|
|
||||||
[Service]
|
|
||||||
Type=simple
|
|
||||||
User=root
|
|
||||||
WorkingDirectory=/opt/kima-hub/backend
|
|
||||||
EnvironmentFile=/opt/kima-hub/backend/.env
|
|
||||||
ExecStart=/usr/bin/node /opt/kima-hub/backend/dist/index.js
|
|
||||||
Restart=on-failure
|
|
||||||
RestartSec=5
|
|
||||||
|
|
||||||
[Install]
|
|
||||||
WantedBy=multi-user.target
|
|
||||||
EOF
|
|
||||||
|
|
||||||
cat <<EOF >/etc/systemd/system/kima-frontend.service
|
|
||||||
[Unit]
|
|
||||||
Description=Kima Hub Frontend
|
|
||||||
After=network.target kima-backend.service
|
|
||||||
|
|
||||||
[Service]
|
|
||||||
Type=simple
|
|
||||||
User=root
|
|
||||||
WorkingDirectory=/opt/kima-hub/frontend
|
|
||||||
EnvironmentFile=/opt/kima-hub/frontend/.env
|
|
||||||
ExecStart=/usr/bin/npm start
|
|
||||||
Restart=on-failure
|
|
||||||
RestartSec=5
|
|
||||||
|
|
||||||
[Install]
|
|
||||||
WantedBy=multi-user.target
|
|
||||||
EOF
|
|
||||||
|
|
||||||
cat <<EOF >/etc/systemd/system/kima-analyzer.service
|
|
||||||
[Unit]
|
|
||||||
Description=Kima Hub Audio Analyzer (Essentia)
|
|
||||||
After=network.target postgresql.service redis-server.service kima-backend.service
|
|
||||||
|
|
||||||
[Service]
|
|
||||||
Type=simple
|
|
||||||
User=root
|
|
||||||
WorkingDirectory=/opt/kima-hub/services/audio-analyzer
|
|
||||||
Environment=DATABASE_URL=postgresql://${PG_DB_USER}:${PG_DB_PASS}@localhost:5432/${PG_DB_NAME}
|
|
||||||
Environment=REDIS_URL=redis://localhost:6379
|
|
||||||
Environment=MUSIC_PATH=/music
|
|
||||||
Environment=BATCH_SIZE=10
|
|
||||||
Environment=SLEEP_INTERVAL=5
|
|
||||||
Environment=NUM_WORKERS=2
|
|
||||||
Environment=THREADS_PER_WORKER=1
|
|
||||||
ExecStart=/usr/bin/python3 /opt/kima-hub/services/audio-analyzer/analyzer.py
|
|
||||||
Restart=on-failure
|
|
||||||
RestartSec=10
|
|
||||||
|
|
||||||
[Install]
|
|
||||||
WantedBy=multi-user.target
|
|
||||||
EOF
|
|
||||||
|
|
||||||
cat <<EOF >/etc/systemd/system/kima-analyzer-clap.service
|
|
||||||
[Unit]
|
|
||||||
Description=Kima Hub CLAP Audio Analyzer
|
|
||||||
After=network.target postgresql.service redis-server.service kima-backend.service kima-analyzer.service
|
|
||||||
|
|
||||||
[Service]
|
|
||||||
Type=simple
|
|
||||||
User=root
|
|
||||||
WorkingDirectory=/opt/kima-hub/services/audio-analyzer-clap
|
|
||||||
Environment=DATABASE_URL=postgresql://${PG_DB_USER}:${PG_DB_PASS}@localhost:5432/${PG_DB_NAME}
|
|
||||||
Environment=REDIS_URL=redis://localhost:6379
|
|
||||||
Environment=BACKEND_URL=http://localhost:3006
|
|
||||||
Environment=MUSIC_PATH=/music
|
|
||||||
Environment=SLEEP_INTERVAL=5
|
|
||||||
Environment=NUM_WORKERS=1
|
|
||||||
ExecStart=/usr/bin/python3 /opt/kima-hub/services/audio-analyzer-clap/analyzer.py
|
|
||||||
Restart=on-failure
|
|
||||||
RestartSec=10
|
|
||||||
|
|
||||||
[Install]
|
|
||||||
WantedBy=multi-user.target
|
|
||||||
EOF
|
|
||||||
systemctl enable -q --now kima-backend kima-frontend kima-analyzer kima-analyzer-clap
|
|
||||||
msg_ok "Created Services"
|
|
||||||
|
|
||||||
motd_ssh
|
|
||||||
customize
|
|
||||||
cleanup_lxc
|
|
||||||
@@ -44,8 +44,6 @@ echo passbolt-ce-server passbolt/nginx-domain string $LOCAL_IP | debconf-set-sel
|
|||||||
echo passbolt-ce-server passbolt/nginx-certificate-file string /etc/ssl/passbolt/passbolt.crt | debconf-set-selections
|
echo passbolt-ce-server passbolt/nginx-certificate-file string /etc/ssl/passbolt/passbolt.crt | debconf-set-selections
|
||||||
echo passbolt-ce-server passbolt/nginx-certificate-key-file string /etc/ssl/passbolt/passbolt.key | debconf-set-selections
|
echo passbolt-ce-server passbolt/nginx-certificate-key-file string /etc/ssl/passbolt/passbolt.key | debconf-set-selections
|
||||||
$STD apt install -y --no-install-recommends passbolt-ce-server
|
$STD apt install -y --no-install-recommends passbolt-ce-server
|
||||||
sed -i 's/client_max_body_size[[:space:]]\+[0-9]\+M;/client_max_body_size 15M;/' /etc/nginx/sites-enabled/nginx-passbolt.conf
|
|
||||||
systemctl reload nginx
|
|
||||||
msg_ok "Setup Passbolt"
|
msg_ok "Setup Passbolt"
|
||||||
|
|
||||||
motd_ssh
|
motd_ssh
|
||||||
|
|||||||
@@ -21,7 +21,7 @@ $STD apt install -y \
|
|||||||
expect
|
expect
|
||||||
msg_ok "Dependencies installed."
|
msg_ok "Dependencies installed."
|
||||||
|
|
||||||
NODE_VERSION="24" setup_nodejs
|
NODE_VERSION=24 setup_nodejs
|
||||||
fetch_and_deploy_gh_release "ProxmoxVE-Local" "community-scripts/ProxmoxVE-Local" "tarball"
|
fetch_and_deploy_gh_release "ProxmoxVE-Local" "community-scripts/ProxmoxVE-Local" "tarball"
|
||||||
|
|
||||||
msg_info "Installing PVE Scripts local"
|
msg_info "Installing PVE Scripts local"
|
||||||
|
|||||||
@@ -164,7 +164,7 @@ server {
|
|||||||
|
|
||||||
location / {
|
location / {
|
||||||
proxy_pass http://127.0.0.1:8000;
|
proxy_pass http://127.0.0.1:8000;
|
||||||
proxy_set_header Host $http_host;
|
proxy_set_header Host $host;
|
||||||
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
||||||
proxy_set_header X-Forwarded-Proto $scheme;
|
proxy_set_header X-Forwarded-Proto $scheme;
|
||||||
proxy_redirect off;
|
proxy_redirect off;
|
||||||
|
|||||||
@@ -28,23 +28,12 @@ setup_deb822_repo \
|
|||||||
"stable" \
|
"stable" \
|
||||||
"main"
|
"main"
|
||||||
$STD apt install -y elasticsearch
|
$STD apt install -y elasticsearch
|
||||||
sed -i 's/^#\{0,2\} *-Xms[0-9]*g.*/-Xms2g/' /etc/elasticsearch/jvm.options
|
sed -i 's/^-Xms.*/-Xms2g/' /etc/elasticsearch/jvm.options
|
||||||
sed -i 's/^#\{0,2\} *-Xmx[0-9]*g.*/-Xmx2g/' /etc/elasticsearch/jvm.options
|
sed -i 's/^-Xmx.*/-Xmx2g/' /etc/elasticsearch/jvm.options
|
||||||
cat <<EOF >>/etc/elasticsearch/elasticsearch.yml
|
|
||||||
discovery.type: single-node
|
|
||||||
xpack.security.enabled: false
|
|
||||||
bootstrap.memory_lock: false
|
|
||||||
EOF
|
|
||||||
$STD /usr/share/elasticsearch/bin/elasticsearch-plugin install ingest-attachment -b
|
$STD /usr/share/elasticsearch/bin/elasticsearch-plugin install ingest-attachment -b
|
||||||
systemctl daemon-reload
|
systemctl daemon-reload
|
||||||
systemctl enable -q elasticsearch
|
systemctl enable -q elasticsearch
|
||||||
systemctl restart -q elasticsearch
|
systemctl restart -q elasticsearch
|
||||||
for i in $(seq 1 30); do
|
|
||||||
if curl -s http://localhost:9200 >/dev/null 2>&1; then
|
|
||||||
break
|
|
||||||
fi
|
|
||||||
sleep 2
|
|
||||||
done
|
|
||||||
msg_ok "Setup Elasticsearch"
|
msg_ok "Setup Elasticsearch"
|
||||||
|
|
||||||
msg_info "Installing Zammad"
|
msg_info "Installing Zammad"
|
||||||
|
|||||||
@@ -1,96 +0,0 @@
|
|||||||
#!/usr/bin/env bash
|
|
||||||
|
|
||||||
# Copyright (c) 2021-2026 community-scripts ORG
|
|
||||||
# Author: community-scripts
|
|
||||||
# License: MIT | https://github.com/community-scripts/ProxmoxVE/raw/main/LICENSE
|
|
||||||
# Source: https://github.com/nicotsx/zerobyte
|
|
||||||
|
|
||||||
source /dev/stdin <<<"$FUNCTIONS_FILE_PATH"
|
|
||||||
color
|
|
||||||
verb_ip6
|
|
||||||
catch_errors
|
|
||||||
setting_up_container
|
|
||||||
network_check
|
|
||||||
update_os
|
|
||||||
|
|
||||||
msg_info "Installing Dependencies"
|
|
||||||
echo "davfs2 davfs2/suid_file boolean false" | debconf-set-selections
|
|
||||||
$STD apt-get install -y \
|
|
||||||
bzip2 \
|
|
||||||
fuse3 \
|
|
||||||
sshfs \
|
|
||||||
davfs2 \
|
|
||||||
openssh-client
|
|
||||||
msg_ok "Installed Dependencies"
|
|
||||||
|
|
||||||
fetch_and_deploy_gh_release "restic" "restic/restic" "singlefile" "latest" "/usr/local/bin" "restic_*_linux_amd64.bz2"
|
|
||||||
mv /usr/local/bin/restic /usr/local/bin/restic.bz2
|
|
||||||
bzip2 -d /usr/local/bin/restic.bz2
|
|
||||||
chmod +x /usr/local/bin/restic
|
|
||||||
|
|
||||||
fetch_and_deploy_gh_release "rclone" "rclone/rclone" "prebuild" "latest" "/opt/rclone" "rclone-*-linux-amd64.zip"
|
|
||||||
ln -sf /opt/rclone/rclone /usr/local/bin/rclone
|
|
||||||
|
|
||||||
fetch_and_deploy_gh_release "shoutrrr" "nicholas-fedor/shoutrrr" "prebuild" "latest" "/opt/shoutrrr" "shoutrrr_linux_amd64_*.tar.gz"
|
|
||||||
ln -sf /opt/shoutrrr/shoutrrr /usr/local/bin/shoutrrr
|
|
||||||
|
|
||||||
msg_info "Installing Bun"
|
|
||||||
export BUN_INSTALL="/root/.bun"
|
|
||||||
curl -fsSL https://bun.sh/install | $STD bash
|
|
||||||
ln -sf /root/.bun/bin/bun /usr/local/bin/bun
|
|
||||||
ln -sf /root/.bun/bin/bunx /usr/local/bin/bunx
|
|
||||||
msg_ok "Installed Bun"
|
|
||||||
|
|
||||||
NODE_VERSION="24" setup_nodejs
|
|
||||||
fetch_and_deploy_gh_release "zerobyte" "nicotsx/zerobyte" "tarball"
|
|
||||||
|
|
||||||
msg_info "Building Zerobyte (Patience)"
|
|
||||||
cd /opt/zerobyte
|
|
||||||
export VITE_RESTIC_VERSION=$(cat ~/.restic)
|
|
||||||
export VITE_RCLONE_VERSION=$(cat ~/.rclone)
|
|
||||||
export VITE_SHOUTRRR_VERSION=$(cat ~/.shoutrrr)
|
|
||||||
export NODE_OPTIONS="--max-old-space-size=3072"
|
|
||||||
$STD bun install
|
|
||||||
$STD node ./node_modules/vite/bin/vite.js build
|
|
||||||
msg_ok "Built Zerobyte"
|
|
||||||
|
|
||||||
msg_info "Configuring Zerobyte"
|
|
||||||
mkdir -p /var/lib/zerobyte/{data,restic/cache,repositories,volumes}
|
|
||||||
APP_SECRET=$(openssl rand -hex 32)
|
|
||||||
cat <<EOF >/opt/zerobyte/.env
|
|
||||||
BASE_URL=http://${LOCAL_IP}:4096
|
|
||||||
APP_SECRET=${APP_SECRET}
|
|
||||||
PORT=4096
|
|
||||||
ZEROBYTE_DATABASE_URL=/var/lib/zerobyte/data/zerobyte.db
|
|
||||||
RESTIC_CACHE_DIR=/var/lib/zerobyte/restic/cache
|
|
||||||
ZEROBYTE_REPOSITORIES_DIR=/var/lib/zerobyte/repositories
|
|
||||||
ZEROBYTE_VOLUMES_DIR=/var/lib/zerobyte/volumes
|
|
||||||
MIGRATIONS_PATH=/opt/zerobyte/app/drizzle
|
|
||||||
NODE_ENV=production
|
|
||||||
EOF
|
|
||||||
msg_ok "Configured Zerobyte"
|
|
||||||
|
|
||||||
msg_info "Creating Service"
|
|
||||||
cat <<EOF >/etc/systemd/system/zerobyte.service
|
|
||||||
[Unit]
|
|
||||||
Description=Zerobyte Backup Automation
|
|
||||||
After=network.target
|
|
||||||
|
|
||||||
[Service]
|
|
||||||
Type=simple
|
|
||||||
User=root
|
|
||||||
WorkingDirectory=/opt/zerobyte
|
|
||||||
EnvironmentFile=/opt/zerobyte/.env
|
|
||||||
ExecStart=/usr/local/bin/bun .output/server/index.mjs
|
|
||||||
Restart=on-failure
|
|
||||||
RestartSec=5
|
|
||||||
|
|
||||||
[Install]
|
|
||||||
WantedBy=multi-user.target
|
|
||||||
EOF
|
|
||||||
systemctl enable -q --now zerobyte
|
|
||||||
msg_ok "Created Service"
|
|
||||||
|
|
||||||
motd_ssh
|
|
||||||
customize
|
|
||||||
cleanup_lxc
|
|
||||||
@@ -4098,11 +4098,10 @@ EOF'
|
|||||||
|
|
||||||
# Installation failed?
|
# Installation failed?
|
||||||
if [[ $install_exit_code -ne 0 ]]; then
|
if [[ $install_exit_code -ne 0 ]]; then
|
||||||
# Prevent job-control signals from suspending the script during recovery.
|
# Prevent SIGTSTP (Ctrl+Z) from suspending the script during recovery.
|
||||||
# In non-interactive shells (bash -c), background processes (spinner) can
|
# In non-interactive shells (bash -c), background processes (spinner) can
|
||||||
# trigger terminal-related signals that stop the entire process group.
|
# trigger terminal-related signals that stop the entire process group.
|
||||||
# TSTP = Ctrl+Z, TTIN = bg read from tty, TTOU = bg write to tty (tostop)
|
trap '' TSTP
|
||||||
trap '' TSTP TTIN TTOU
|
|
||||||
|
|
||||||
msg_error "Installation failed in container ${CTID} (exit code: ${install_exit_code})"
|
msg_error "Installation failed in container ${CTID} (exit code: ${install_exit_code})"
|
||||||
|
|
||||||
@@ -4552,8 +4551,8 @@ EOF'
|
|||||||
post_update_to_api "failed" "$install_exit_code" "force"
|
post_update_to_api "failed" "$install_exit_code" "force"
|
||||||
$STD echo -e "${TAB}${CM:-✔} Telemetry finalized"
|
$STD echo -e "${TAB}${CM:-✔} Telemetry finalized"
|
||||||
|
|
||||||
# Restore default job-control signal handling before exit
|
# Restore default SIGTSTP handling before exit
|
||||||
trap - TSTP TTIN TTOU
|
trap - TSTP
|
||||||
exit $install_exit_code
|
exit $install_exit_code
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
|||||||
@@ -607,7 +607,6 @@ stop_spinner() {
|
|||||||
|
|
||||||
unset SPINNER_PID SPINNER_MSG
|
unset SPINNER_PID SPINNER_MSG
|
||||||
stty sane 2>/dev/null || true
|
stty sane 2>/dev/null || true
|
||||||
stty -tostop 2>/dev/null || true
|
|
||||||
}
|
}
|
||||||
|
|
||||||
# ==============================================================================
|
# ==============================================================================
|
||||||
|
|||||||
135
misc/tools.func
135
misc/tools.func
@@ -783,25 +783,16 @@ github_api_call() {
|
|||||||
|
|
||||||
for attempt in $(seq 1 $max_retries); do
|
for attempt in $(seq 1 $max_retries); do
|
||||||
local http_code
|
local http_code
|
||||||
http_code=$(curl -sSL -w "%{http_code}" -o "$output_file" \
|
http_code=$(curl -fsSL -w "%{http_code}" -o "$output_file" \
|
||||||
-H "Accept: application/vnd.github+json" \
|
-H "Accept: application/vnd.github+json" \
|
||||||
-H "X-GitHub-Api-Version: 2022-11-28" \
|
-H "X-GitHub-Api-Version: 2022-11-28" \
|
||||||
"${header_args[@]}" \
|
"${header_args[@]}" \
|
||||||
"$url" 2>/dev/null) || true
|
"$url" 2>/dev/null || echo "000")
|
||||||
|
|
||||||
case "$http_code" in
|
case "$http_code" in
|
||||||
200)
|
200)
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
401)
|
|
||||||
msg_error "GitHub API authentication failed (HTTP 401)."
|
|
||||||
if [[ -n "${GITHUB_TOKEN:-}" ]]; then
|
|
||||||
msg_error "Your GITHUB_TOKEN appears to be invalid or expired."
|
|
||||||
else
|
|
||||||
msg_error "The repository may require authentication. Try: export GITHUB_TOKEN=\"ghp_your_token\""
|
|
||||||
fi
|
|
||||||
return 1
|
|
||||||
;;
|
|
||||||
403)
|
403)
|
||||||
# Rate limit - check if we can retry
|
# Rate limit - check if we can retry
|
||||||
if [[ $attempt -lt $max_retries ]]; then
|
if [[ $attempt -lt $max_retries ]]; then
|
||||||
@@ -810,22 +801,11 @@ github_api_call() {
|
|||||||
retry_delay=$((retry_delay * 2))
|
retry_delay=$((retry_delay * 2))
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
msg_error "GitHub API rate limit exceeded. Set GITHUB_TOKEN to increase limits."
|
||||||
msg_error "To increase the limit, export a GitHub token before running the script:"
|
|
||||||
msg_error " export GITHUB_TOKEN=\"ghp_your_token_here\""
|
|
||||||
return 1
|
return 1
|
||||||
;;
|
;;
|
||||||
404)
|
404)
|
||||||
msg_error "GitHub repository or release not found (HTTP 404): $url"
|
msg_error "GitHub API endpoint not found: $url"
|
||||||
return 1
|
|
||||||
;;
|
|
||||||
000 | "")
|
|
||||||
if [[ $attempt -lt $max_retries ]]; then
|
|
||||||
sleep "$retry_delay"
|
|
||||||
continue
|
|
||||||
fi
|
|
||||||
msg_error "GitHub API connection failed (no response)."
|
|
||||||
msg_error "Check your network/DNS: curl -sSL https://api.github.com/rate_limit"
|
|
||||||
return 1
|
return 1
|
||||||
;;
|
;;
|
||||||
*)
|
*)
|
||||||
@@ -833,7 +813,7 @@ github_api_call() {
|
|||||||
sleep "$retry_delay"
|
sleep "$retry_delay"
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
msg_error "GitHub API call failed (HTTP $http_code)."
|
msg_error "GitHub API call failed with HTTP $http_code"
|
||||||
return 1
|
return 1
|
||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
@@ -853,18 +833,14 @@ codeberg_api_call() {
|
|||||||
|
|
||||||
for attempt in $(seq 1 $max_retries); do
|
for attempt in $(seq 1 $max_retries); do
|
||||||
local http_code
|
local http_code
|
||||||
http_code=$(curl -sSL -w "%{http_code}" -o "$output_file" \
|
http_code=$(curl -fsSL -w "%{http_code}" -o "$output_file" \
|
||||||
-H "Accept: application/json" \
|
-H "Accept: application/json" \
|
||||||
"$url" 2>/dev/null) || true
|
"$url" 2>/dev/null || echo "000")
|
||||||
|
|
||||||
case "$http_code" in
|
case "$http_code" in
|
||||||
200)
|
200)
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
401)
|
|
||||||
msg_error "Codeberg API authentication failed (HTTP 401)."
|
|
||||||
return 1
|
|
||||||
;;
|
|
||||||
403)
|
403)
|
||||||
# Rate limit - retry
|
# Rate limit - retry
|
||||||
if [[ $attempt -lt $max_retries ]]; then
|
if [[ $attempt -lt $max_retries ]]; then
|
||||||
@@ -873,20 +849,11 @@ codeberg_api_call() {
|
|||||||
retry_delay=$((retry_delay * 2))
|
retry_delay=$((retry_delay * 2))
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
msg_error "Codeberg API rate limit exceeded (HTTP 403)."
|
msg_error "Codeberg API rate limit exceeded."
|
||||||
return 1
|
return 1
|
||||||
;;
|
;;
|
||||||
404)
|
404)
|
||||||
msg_error "Codeberg repository or release not found (HTTP 404): $url"
|
msg_error "Codeberg API endpoint not found: $url"
|
||||||
return 1
|
|
||||||
;;
|
|
||||||
000 | "")
|
|
||||||
if [[ $attempt -lt $max_retries ]]; then
|
|
||||||
sleep "$retry_delay"
|
|
||||||
continue
|
|
||||||
fi
|
|
||||||
msg_error "Codeberg API connection failed (no response)."
|
|
||||||
msg_error "Check your network/DNS: curl -sSL https://codeberg.org"
|
|
||||||
return 1
|
return 1
|
||||||
;;
|
;;
|
||||||
*)
|
*)
|
||||||
@@ -894,7 +861,7 @@ codeberg_api_call() {
|
|||||||
sleep "$retry_delay"
|
sleep "$retry_delay"
|
||||||
continue
|
continue
|
||||||
fi
|
fi
|
||||||
msg_error "Codeberg API call failed (HTTP $http_code)."
|
msg_error "Codeberg API call failed with HTTP $http_code"
|
||||||
return 1
|
return 1
|
||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
@@ -1474,7 +1441,7 @@ get_latest_github_release() {
|
|||||||
|
|
||||||
if ! github_api_call "https://api.github.com/repos/${repo}/releases/latest" "$temp_file"; then
|
if ! github_api_call "https://api.github.com/repos/${repo}/releases/latest" "$temp_file"; then
|
||||||
rm -f "$temp_file"
|
rm -f "$temp_file"
|
||||||
return 0
|
return 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
local version
|
local version
|
||||||
@@ -1482,8 +1449,7 @@ get_latest_github_release() {
|
|||||||
rm -f "$temp_file"
|
rm -f "$temp_file"
|
||||||
|
|
||||||
if [[ -z "$version" ]]; then
|
if [[ -z "$version" ]]; then
|
||||||
msg_error "Could not determine latest version for ${repo}"
|
return 1
|
||||||
return 0
|
|
||||||
fi
|
fi
|
||||||
|
|
||||||
echo "$version"
|
echo "$version"
|
||||||
@@ -1500,7 +1466,7 @@ get_latest_codeberg_release() {
|
|||||||
# Codeberg API: get all releases and pick the first non-draft/non-prerelease
|
# Codeberg API: get all releases and pick the first non-draft/non-prerelease
|
||||||
if ! codeberg_api_call "https://codeberg.org/api/v1/repos/${repo}/releases" "$temp_file"; then
|
if ! codeberg_api_call "https://codeberg.org/api/v1/repos/${repo}/releases" "$temp_file"; then
|
||||||
rm -f "$temp_file"
|
rm -f "$temp_file"
|
||||||
return 0
|
return 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
local version
|
local version
|
||||||
@@ -1514,8 +1480,7 @@ get_latest_codeberg_release() {
|
|||||||
rm -f "$temp_file"
|
rm -f "$temp_file"
|
||||||
|
|
||||||
if [[ -z "$version" ]]; then
|
if [[ -z "$version" ]]; then
|
||||||
msg_error "Could not determine latest version for ${repo}"
|
return 1
|
||||||
return 0
|
|
||||||
fi
|
fi
|
||||||
|
|
||||||
echo "$version"
|
echo "$version"
|
||||||
@@ -1602,34 +1567,13 @@ get_latest_gh_tag() {
|
|||||||
"${header_args[@]}" \
|
"${header_args[@]}" \
|
||||||
"https://api.github.com/repos/${repo}/tags?per_page=100" 2>/dev/null) || true
|
"https://api.github.com/repos/${repo}/tags?per_page=100" 2>/dev/null) || true
|
||||||
|
|
||||||
if [[ "$http_code" == "401" ]]; then
|
|
||||||
msg_error "GitHub API authentication failed (HTTP 401)."
|
|
||||||
if [[ -n "${GITHUB_TOKEN:-}" ]]; then
|
|
||||||
msg_error "Your GITHUB_TOKEN appears to be invalid or expired."
|
|
||||||
else
|
|
||||||
msg_error "The repository may require authentication. Try: export GITHUB_TOKEN=\"ghp_your_token\""
|
|
||||||
fi
|
|
||||||
rm -f /tmp/gh_tags.json
|
|
||||||
return 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [[ "$http_code" == "403" ]]; then
|
if [[ "$http_code" == "403" ]]; then
|
||||||
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
msg_warn "GitHub API rate limit exceeded while fetching tags for ${repo}"
|
||||||
msg_error "To increase the limit, export a GitHub token before running the script:"
|
|
||||||
msg_error " export GITHUB_TOKEN=\"ghp_your_token_here\""
|
|
||||||
rm -f /tmp/gh_tags.json
|
|
||||||
return 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [[ "$http_code" == "000" || -z "$http_code" ]]; then
|
|
||||||
msg_error "GitHub API connection failed (no response)."
|
|
||||||
msg_error "Check your network/DNS: curl -sSL https://api.github.com/rate_limit"
|
|
||||||
rm -f /tmp/gh_tags.json
|
rm -f /tmp/gh_tags.json
|
||||||
return 1
|
return 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [[ "$http_code" != "200" ]] || [[ ! -s /tmp/gh_tags.json ]]; then
|
if [[ "$http_code" != "200" ]] || [[ ! -s /tmp/gh_tags.json ]]; then
|
||||||
msg_error "Unable to fetch tags for ${repo} (HTTP ${http_code})"
|
|
||||||
rm -f /tmp/gh_tags.json
|
rm -f /tmp/gh_tags.json
|
||||||
return 1
|
return 1
|
||||||
fi
|
fi
|
||||||
@@ -1715,15 +1659,6 @@ check_for_gh_release() {
|
|||||||
|
|
||||||
if [[ "$http_code" == "200" ]] && [[ -s /tmp/gh_check.json ]]; then
|
if [[ "$http_code" == "200" ]] && [[ -s /tmp/gh_check.json ]]; then
|
||||||
releases_json="[$(</tmp/gh_check.json)]"
|
releases_json="[$(</tmp/gh_check.json)]"
|
||||||
elif [[ "$http_code" == "401" ]]; then
|
|
||||||
msg_error "GitHub API authentication failed (HTTP 401)."
|
|
||||||
if [[ -n "${GITHUB_TOKEN:-}" ]]; then
|
|
||||||
msg_error "Your GITHUB_TOKEN appears to be invalid or expired."
|
|
||||||
else
|
|
||||||
msg_error "The repository may require authentication. Try: export GITHUB_TOKEN=\"ghp_your_token\""
|
|
||||||
fi
|
|
||||||
rm -f /tmp/gh_check.json
|
|
||||||
return 1
|
|
||||||
elif [[ "$http_code" == "403" ]]; then
|
elif [[ "$http_code" == "403" ]]; then
|
||||||
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
||||||
msg_error "To increase the limit, export a GitHub token before running the script:"
|
msg_error "To increase the limit, export a GitHub token before running the script:"
|
||||||
@@ -1744,26 +1679,12 @@ check_for_gh_release() {
|
|||||||
|
|
||||||
if [[ "$http_code" == "200" ]] && [[ -s /tmp/gh_check.json ]]; then
|
if [[ "$http_code" == "200" ]] && [[ -s /tmp/gh_check.json ]]; then
|
||||||
releases_json=$(</tmp/gh_check.json)
|
releases_json=$(</tmp/gh_check.json)
|
||||||
elif [[ "$http_code" == "401" ]]; then
|
|
||||||
msg_error "GitHub API authentication failed (HTTP 401)."
|
|
||||||
if [[ -n "${GITHUB_TOKEN:-}" ]]; then
|
|
||||||
msg_error "Your GITHUB_TOKEN appears to be invalid or expired."
|
|
||||||
else
|
|
||||||
msg_error "The repository may require authentication. Try: export GITHUB_TOKEN=\"ghp_your_token\""
|
|
||||||
fi
|
|
||||||
rm -f /tmp/gh_check.json
|
|
||||||
return 1
|
|
||||||
elif [[ "$http_code" == "403" ]]; then
|
elif [[ "$http_code" == "403" ]]; then
|
||||||
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
||||||
msg_error "To increase the limit, export a GitHub token before running the script:"
|
msg_error "To increase the limit, export a GitHub token before running the script:"
|
||||||
msg_error " export GITHUB_TOKEN=\"ghp_your_token_here\""
|
msg_error " export GITHUB_TOKEN=\"ghp_your_token_here\""
|
||||||
rm -f /tmp/gh_check.json
|
rm -f /tmp/gh_check.json
|
||||||
return 1
|
return 1
|
||||||
elif [[ "$http_code" == "000" || -z "$http_code" ]]; then
|
|
||||||
msg_error "GitHub API connection failed (no response)."
|
|
||||||
msg_error "Check your network/DNS: curl -sSL https://api.github.com/rate_limit"
|
|
||||||
rm -f /tmp/gh_check.json
|
|
||||||
return 1
|
|
||||||
else
|
else
|
||||||
msg_error "Unable to fetch releases for ${app} (HTTP ${http_code})"
|
msg_error "Unable to fetch releases for ${app} (HTTP ${http_code})"
|
||||||
rm -f /tmp/gh_check.json
|
rm -f /tmp/gh_check.json
|
||||||
@@ -2687,22 +2608,12 @@ function fetch_and_deploy_gh_release() {
|
|||||||
done
|
done
|
||||||
|
|
||||||
if ! $success; then
|
if ! $success; then
|
||||||
if [[ "$http_code" == "401" ]]; then
|
if [[ "$http_code" == "403" ]]; then
|
||||||
msg_error "GitHub API authentication failed (HTTP 401)."
|
|
||||||
if [[ -n "${GITHUB_TOKEN:-}" ]]; then
|
|
||||||
msg_error "Your GITHUB_TOKEN appears to be invalid or expired."
|
|
||||||
else
|
|
||||||
msg_error "The repository may require authentication. Try: export GITHUB_TOKEN=\"ghp_your_token\""
|
|
||||||
fi
|
|
||||||
elif [[ "$http_code" == "403" ]]; then
|
|
||||||
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
msg_error "GitHub API rate limit exceeded (HTTP 403)."
|
||||||
msg_error "To increase the limit, export a GitHub token before running the script:"
|
msg_error "To increase the limit, export a GitHub token before running the script:"
|
||||||
msg_error " export GITHUB_TOKEN=\"ghp_your_token_here\""
|
msg_error " export GITHUB_TOKEN=\"ghp_your_token_here\""
|
||||||
elif [[ "$http_code" == "000" || -z "$http_code" ]]; then
|
|
||||||
msg_error "GitHub API connection failed (no response)."
|
|
||||||
msg_error "Check your network/DNS: curl -sSL https://api.github.com/rate_limit"
|
|
||||||
else
|
else
|
||||||
msg_error "Failed to fetch release metadata (HTTP $http_code)"
|
msg_error "Failed to fetch release metadata from $api_url after $max_retries attempts (HTTP $http_code)"
|
||||||
fi
|
fi
|
||||||
return 1
|
return 1
|
||||||
fi
|
fi
|
||||||
@@ -4237,18 +4148,6 @@ NVIDIA_PIN
|
|||||||
# VA-API for hybrid setups (Intel + NVIDIA)
|
# VA-API for hybrid setups (Intel + NVIDIA)
|
||||||
$STD apt-get -y install va-driver-all vainfo 2>/dev/null || true
|
$STD apt-get -y install va-driver-all vainfo 2>/dev/null || true
|
||||||
|
|
||||||
# Fix GLX alternatives: nvidia-alternative diverts mesa libs but in LXC
|
|
||||||
# containers the nvidia GLX libs are typically missing, leaving libGL.so.1
|
|
||||||
# pointing nowhere. Fall back to mesa if nvidia GLX dir is empty/missing.
|
|
||||||
if command -v update-glx &>/dev/null; then
|
|
||||||
local nvidia_glx_dir="/usr/lib/nvidia"
|
|
||||||
if [[ ! -f "${nvidia_glx_dir}/libGL.so.1" ]] && [[ -d /usr/lib/mesa-diverted ]]; then
|
|
||||||
msg_info "NVIDIA GLX libs missing in container - falling back to mesa"
|
|
||||||
$STD update-glx --set glx /usr/lib/mesa-diverted 2>/dev/null || true
|
|
||||||
ldconfig 2>/dev/null || true
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
|
|
||||||
msg_ok "NVIDIA GPU configured"
|
msg_ok "NVIDIA GPU configured"
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -76,90 +76,70 @@ grep -q "lxc.mount.entry: /dev/net/tun" "$CTID_CONFIG_PATH" || echo "lxc.mount.e
|
|||||||
header_info
|
header_info
|
||||||
msg_info "Installing Tailscale in CT $CTID"
|
msg_info "Installing Tailscale in CT $CTID"
|
||||||
|
|
||||||
pct exec "$CTID" -- sh -c '
|
pct exec "$CTID" -- bash -c '
|
||||||
set -e
|
set -e
|
||||||
|
export DEBIAN_FRONTEND=noninteractive
|
||||||
|
|
||||||
# Detect OS inside container
|
# Source os-release properly (handles quoted values)
|
||||||
if [ -f /etc/alpine-release ]; then
|
source /etc/os-release
|
||||||
# ── Alpine Linux ──
|
|
||||||
echo "[INFO] Alpine Linux detected, installing Tailscale via apk..."
|
|
||||||
|
|
||||||
# Enable community repo if not already enabled
|
# Fallback if DNS is poisoned or blocked
|
||||||
if ! grep -q "^[^#].*community" /etc/apk/repositories 2>/dev/null; then
|
ORIG_RESOLV="/etc/resolv.conf"
|
||||||
ALPINE_VERSION=$(cat /etc/alpine-release | cut -d. -f1,2)
|
BACKUP_RESOLV="/tmp/resolv.conf.backup"
|
||||||
echo "https://dl-cdn.alpinelinux.org/alpine/v${ALPINE_VERSION}/community" >> /etc/apk/repositories
|
|
||||||
fi
|
|
||||||
|
|
||||||
apk update
|
# Check DNS resolution using multiple methods (dig may not be installed)
|
||||||
apk add --no-cache tailscale
|
dns_check_failed=true
|
||||||
|
if command -v dig &>/dev/null; then
|
||||||
# Enable and start Tailscale service
|
if dig +short pkgs.tailscale.com 2>/dev/null | grep -qvE "^127\.|^0\.0\.0\.0$|^$"; then
|
||||||
rc-update add tailscale default 2>/dev/null || true
|
|
||||||
rc-service tailscale start 2>/dev/null || true
|
|
||||||
|
|
||||||
else
|
|
||||||
# ── Debian / Ubuntu ──
|
|
||||||
export DEBIAN_FRONTEND=noninteractive
|
|
||||||
|
|
||||||
# Source os-release properly (handles quoted values)
|
|
||||||
. /etc/os-release
|
|
||||||
|
|
||||||
# Fallback if DNS is poisoned or blocked
|
|
||||||
ORIG_RESOLV="/etc/resolv.conf"
|
|
||||||
BACKUP_RESOLV="/tmp/resolv.conf.backup"
|
|
||||||
|
|
||||||
# Check DNS resolution using multiple methods (dig may not be installed)
|
|
||||||
dns_check_failed=true
|
|
||||||
if command -v dig >/dev/null 2>&1; then
|
|
||||||
if dig +short pkgs.tailscale.com 2>/dev/null | grep -qvE "^127\.|^0\.0\.0\.0$|^$"; then
|
|
||||||
dns_check_failed=false
|
|
||||||
fi
|
|
||||||
elif command -v host >/dev/null 2>&1; then
|
|
||||||
if host pkgs.tailscale.com 2>/dev/null | grep -q "has address"; then
|
|
||||||
dns_check_failed=false
|
|
||||||
fi
|
|
||||||
elif command -v nslookup >/dev/null 2>&1; then
|
|
||||||
if nslookup pkgs.tailscale.com 2>/dev/null | grep -q "Address:"; then
|
|
||||||
dns_check_failed=false
|
|
||||||
fi
|
|
||||||
elif command -v getent >/dev/null 2>&1; then
|
|
||||||
if getent hosts pkgs.tailscale.com >/dev/null 2>&1; then
|
|
||||||
dns_check_failed=false
|
|
||||||
fi
|
|
||||||
else
|
|
||||||
# No DNS tools available, try curl directly and assume DNS works
|
|
||||||
dns_check_failed=false
|
dns_check_failed=false
|
||||||
fi
|
fi
|
||||||
|
elif command -v host &>/dev/null; then
|
||||||
if $dns_check_failed; then
|
if host pkgs.tailscale.com 2>/dev/null | grep -q "has address"; then
|
||||||
echo "[INFO] DNS resolution for pkgs.tailscale.com failed (blocked or redirected)."
|
dns_check_failed=false
|
||||||
echo "[INFO] Temporarily overriding /etc/resolv.conf with Cloudflare DNS (1.1.1.1)"
|
|
||||||
cp "$ORIG_RESOLV" "$BACKUP_RESOLV"
|
|
||||||
echo "nameserver 1.1.1.1" >"$ORIG_RESOLV"
|
|
||||||
fi
|
fi
|
||||||
|
elif command -v nslookup &>/dev/null; then
|
||||||
if ! command -v curl >/dev/null 2>&1; then
|
if nslookup pkgs.tailscale.com 2>/dev/null | grep -q "Address:"; then
|
||||||
echo "[INFO] curl not found, installing..."
|
dns_check_failed=false
|
||||||
apt-get update -qq
|
|
||||||
apt-get install -y curl >/dev/null
|
|
||||||
fi
|
fi
|
||||||
|
elif command -v getent &>/dev/null; then
|
||||||
|
if getent hosts pkgs.tailscale.com &>/dev/null; then
|
||||||
|
dns_check_failed=false
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
# No DNS tools available, try curl directly and assume DNS works
|
||||||
|
dns_check_failed=false
|
||||||
|
fi
|
||||||
|
|
||||||
# Ensure keyrings directory exists
|
if $dns_check_failed; then
|
||||||
mkdir -p /usr/share/keyrings
|
echo "[INFO] DNS resolution for pkgs.tailscale.com failed (blocked or redirected)."
|
||||||
|
echo "[INFO] Temporarily overriding /etc/resolv.conf with Cloudflare DNS (1.1.1.1)"
|
||||||
curl -fsSL "https://pkgs.tailscale.com/stable/${ID}/${VERSION_CODENAME}.noarmor.gpg" \
|
cp "$ORIG_RESOLV" "$BACKUP_RESOLV"
|
||||||
| tee /usr/share/keyrings/tailscale-archive-keyring.gpg >/dev/null
|
echo "nameserver 1.1.1.1" >"$ORIG_RESOLV"
|
||||||
|
fi
|
||||||
echo "deb [signed-by=/usr/share/keyrings/tailscale-archive-keyring.gpg] https://pkgs.tailscale.com/stable/${ID} ${VERSION_CODENAME} main" \
|
|
||||||
>/etc/apt/sources.list.d/tailscale.list
|
|
||||||
|
|
||||||
|
if ! command -v curl &>/dev/null; then
|
||||||
|
echo "[INFO] curl not found, installing..."
|
||||||
apt-get update -qq
|
apt-get update -qq
|
||||||
apt-get install -y tailscale >/dev/null
|
apt update -qq
|
||||||
|
apt install -y curl >/dev/null
|
||||||
|
fi
|
||||||
|
|
||||||
if [ -f /tmp/resolv.conf.backup ]; then
|
# Ensure keyrings directory exists
|
||||||
echo "[INFO] Restoring original /etc/resolv.conf"
|
mkdir -p /usr/share/keyrings
|
||||||
mv /tmp/resolv.conf.backup /etc/resolv.conf
|
|
||||||
fi
|
curl -fsSL "https://pkgs.tailscale.com/stable/${ID}/${VERSION_CODENAME}.noarmor.gpg" \
|
||||||
|
| tee /usr/share/keyrings/tailscale-archive-keyring.gpg >/dev/null
|
||||||
|
|
||||||
|
echo "deb [signed-by=/usr/share/keyrings/tailscale-archive-keyring.gpg] https://pkgs.tailscale.com/stable/${ID} ${VERSION_CODENAME} main" \
|
||||||
|
>/etc/apt/sources.list.d/tailscale.list
|
||||||
|
|
||||||
|
apt-get update -qq
|
||||||
|
apt update -qq
|
||||||
|
apt install -y tailscale >/dev/null
|
||||||
|
|
||||||
|
if [[ -f /tmp/resolv.conf.backup ]]; then
|
||||||
|
echo "[INFO] Restoring original /etc/resolv.conf"
|
||||||
|
mv /tmp/resolv.conf.backup /etc/resolv.conf
|
||||||
fi
|
fi
|
||||||
'
|
'
|
||||||
|
|
||||||
|
|||||||
@@ -7,12 +7,8 @@
|
|||||||
|
|
||||||
if ! command -v curl &>/dev/null; then
|
if ! command -v curl &>/dev/null; then
|
||||||
printf "\r\e[2K%b" '\033[93m Setup Source \033[m' >&2
|
printf "\r\e[2K%b" '\033[93m Setup Source \033[m' >&2
|
||||||
if [[ -f "/etc/alpine-release" ]]; then
|
apt-get update >/dev/null 2>&1
|
||||||
apk -U add curl >/dev/null 2>&1
|
apt-get install -y curl >/dev/null 2>&1
|
||||||
else
|
|
||||||
apt-get update >/dev/null 2>&1
|
|
||||||
apt-get install -y curl >/dev/null 2>&1
|
|
||||||
fi
|
|
||||||
fi
|
fi
|
||||||
source <(curl -fsSL https://raw.githubusercontent.com/community-scripts/ProxmoxVE/main/misc/core.func)
|
source <(curl -fsSL https://raw.githubusercontent.com/community-scripts/ProxmoxVE/main/misc/core.func)
|
||||||
source <(curl -fsSL https://raw.githubusercontent.com/community-scripts/ProxmoxVE/main/misc/tools.func)
|
source <(curl -fsSL https://raw.githubusercontent.com/community-scripts/ProxmoxVE/main/misc/tools.func)
|
||||||
@@ -55,7 +51,7 @@ EOF
|
|||||||
# HELPER FUNCTIONS
|
# HELPER FUNCTIONS
|
||||||
# ==============================================================================
|
# ==============================================================================
|
||||||
get_ip() {
|
get_ip() {
|
||||||
ifconfig | grep -v '127.0.0.1' | grep -Eo 'inet (addr:)?([0-9]*\.){3}[0-9]*' | grep -m1 -Eo '([0-9]*\.){3}[0-9]*' || echo "127.0.0.1"
|
hostname -I 2>/dev/null | awk '{print $1}' || echo "127.0.0.1"
|
||||||
}
|
}
|
||||||
|
|
||||||
# ==============================================================================
|
# ==============================================================================
|
||||||
@@ -72,16 +68,6 @@ else
|
|||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
# ==============================================================================
|
|
||||||
# DEPENDENCY CHECK
|
|
||||||
# ==============================================================================
|
|
||||||
if ! command -v jq &>/dev/null; then
|
|
||||||
printf "\r\e[2K%b" '\033[93m Installing jq \033[m' >&2
|
|
||||||
if [[ "$OS" == "Alpine" ]]; then
|
|
||||||
apk -U add jq >/dev/null 2>&1
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
|
|
||||||
# ==============================================================================
|
# ==============================================================================
|
||||||
# UNINSTALL
|
# UNINSTALL
|
||||||
# ==============================================================================
|
# ==============================================================================
|
||||||
|
|||||||
@@ -28,7 +28,7 @@ INSTALL_PATH="/opt/immich-proxy"
|
|||||||
CONFIG_PATH="/opt/immich-proxy/app"
|
CONFIG_PATH="/opt/immich-proxy/app"
|
||||||
DEFAULT_PORT=3000
|
DEFAULT_PORT=3000
|
||||||
|
|
||||||
# Initialize all core functions (colors, formatting, icons, $STD mode)
|
# Initialize all core functions (colors, formatting, icons, STD mode)
|
||||||
load_functions
|
load_functions
|
||||||
init_tool_telemetry "" "addon"
|
init_tool_telemetry "" "addon"
|
||||||
|
|
||||||
|
|||||||
@@ -288,8 +288,8 @@ function default_settings() {
|
|||||||
echo -e "${DGN}Using Hostname: ${BGN}${HN}${CL}"
|
echo -e "${DGN}Using Hostname: ${BGN}${HN}${CL}"
|
||||||
echo -e "${DGN}Allocated Cores: ${BGN}${CORE_COUNT}${CL}"
|
echo -e "${DGN}Allocated Cores: ${BGN}${CORE_COUNT}${CL}"
|
||||||
echo -e "${DGN}Allocated RAM: ${BGN}${RAM_SIZE}${CL}"
|
echo -e "${DGN}Allocated RAM: ${BGN}${RAM_SIZE}${CL}"
|
||||||
if ! ip link show "${BRG}" &>/dev/null; then
|
if ! grep -q "^iface ${BRG}" /etc/network/interfaces; then
|
||||||
msg_error "Bridge '${BRG}' does not exist"
|
msg_error "Bridge '${BRG}' does not exist in /etc/network/interfaces"
|
||||||
exit
|
exit
|
||||||
else
|
else
|
||||||
echo -e "${DGN}Using LAN Bridge: ${BGN}${BRG}${CL}"
|
echo -e "${DGN}Using LAN Bridge: ${BGN}${BRG}${CL}"
|
||||||
@@ -305,8 +305,8 @@ function default_settings() {
|
|||||||
if [ "$NETWORK_MODE" = "dual" ]; then
|
if [ "$NETWORK_MODE" = "dual" ]; then
|
||||||
echo -e "${DGN}Network Mode: ${BGN}Dual Interface (Firewall)${CL}"
|
echo -e "${DGN}Network Mode: ${BGN}Dual Interface (Firewall)${CL}"
|
||||||
echo -e "${DGN}Using WAN MAC Address: ${BGN}${WAN_MAC}${CL}"
|
echo -e "${DGN}Using WAN MAC Address: ${BGN}${WAN_MAC}${CL}"
|
||||||
if ! ip link show "${WAN_BRG}" &>/dev/null; then
|
if ! grep -q "^iface ${WAN_BRG}" /etc/network/interfaces; then
|
||||||
msg_error "Bridge '${WAN_BRG}' does not exist"
|
msg_error "Bridge '${WAN_BRG}' does not exist in /etc/network/interfaces"
|
||||||
exit
|
exit
|
||||||
else
|
else
|
||||||
echo -e "${DGN}Using WAN Bridge: ${BGN}${WAN_BRG}${CL}"
|
echo -e "${DGN}Using WAN Bridge: ${BGN}${WAN_BRG}${CL}"
|
||||||
@@ -424,8 +424,8 @@ function advanced_settings() {
|
|||||||
if [ -z $BRG ]; then
|
if [ -z $BRG ]; then
|
||||||
BRG="vmbr0"
|
BRG="vmbr0"
|
||||||
fi
|
fi
|
||||||
if ! ip link show "${BRG}" &>/dev/null; then
|
if ! grep -q "^iface ${BRG}" /etc/network/interfaces; then
|
||||||
msg_error "Bridge '${BRG}' does not exist"
|
msg_error "Bridge '${BRG}' does not exist in /etc/network/interfaces"
|
||||||
exit
|
exit
|
||||||
fi
|
fi
|
||||||
echo -e "${DGN}Using LAN Bridge: ${BGN}$BRG${CL}"
|
echo -e "${DGN}Using LAN Bridge: ${BGN}$BRG${CL}"
|
||||||
@@ -474,8 +474,8 @@ function advanced_settings() {
|
|||||||
if [ -z $WAN_BRG ]; then
|
if [ -z $WAN_BRG ]; then
|
||||||
WAN_BRG="vmbr1"
|
WAN_BRG="vmbr1"
|
||||||
fi
|
fi
|
||||||
if ! ip link show "${WAN_BRG}" &>/dev/null; then
|
if ! grep -q "^iface ${WAN_BRG}" /etc/network/interfaces; then
|
||||||
msg_error "WAN Bridge '${WAN_BRG}' does not exist"
|
msg_error "WAN Bridge '${WAN_BRG}' does not exist in /etc/network/interfaces"
|
||||||
exit
|
exit
|
||||||
fi
|
fi
|
||||||
echo -e "${DGN}Using WAN Bridge: ${BGN}$WAN_BRG${CL}"
|
echo -e "${DGN}Using WAN Bridge: ${BGN}$WAN_BRG${CL}"
|
||||||
|
|||||||
Reference in New Issue
Block a user