Compare commits
6 Commits
0.1.2
...
e0242bc823
| Author | SHA1 | Date | |
|---|---|---|---|
| e0242bc823 | |||
| 75f8106ca5 | |||
| 4e4f8ad5d4 | |||
| 5c5d51350f | |||
| ba1559e790 | |||
| ab9d03be8a |
@@ -58,7 +58,3 @@ INIT_ADMIN_PASSWORD=ChangeMe123!
|
||||
# ------------------------------
|
||||
# Host port mapped to frontend container port 80.
|
||||
FRONTEND_PORT=5173
|
||||
# Base API URL used at frontend build time.
|
||||
# For reverse proxy + SSL, keep this relative to avoid mixed-content issues.
|
||||
# Example direct mode: VITE_API_URL=http://localhost:8000/api/v1
|
||||
VITE_API_URL=/api/v1
|
||||
|
||||
91
.github/workflows/docker-release.yml
vendored
Normal file
91
.github/workflows/docker-release.yml
vendored
Normal file
@@ -0,0 +1,91 @@
|
||||
name: Docker Publish (Release)
|
||||
|
||||
on:
|
||||
release:
|
||||
types: [published]
|
||||
workflow_dispatch:
|
||||
inputs:
|
||||
version:
|
||||
description: "Version tag to publish (e.g. 0.1.2 or v0.1.2)"
|
||||
required: false
|
||||
type: string
|
||||
|
||||
jobs:
|
||||
publish:
|
||||
name: Build and Push Docker Images
|
||||
runs-on: ubuntu-latest
|
||||
permissions:
|
||||
contents: read
|
||||
|
||||
env:
|
||||
# Optional repo variable. If unset, DOCKERHUB_USERNAME is used.
|
||||
IMAGE_NAMESPACE: ${{ vars.DOCKERHUB_NAMESPACE }}
|
||||
|
||||
steps:
|
||||
- name: Checkout
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Resolve version/tag
|
||||
id: ver
|
||||
shell: bash
|
||||
run: |
|
||||
RAW_TAG="${{ github.event.release.tag_name }}"
|
||||
if [ -z "$RAW_TAG" ]; then
|
||||
RAW_TAG="${{ inputs.version }}"
|
||||
fi
|
||||
if [ -z "$RAW_TAG" ]; then
|
||||
RAW_TAG="${GITHUB_REF_NAME}"
|
||||
fi
|
||||
|
||||
CLEAN_TAG="${RAW_TAG#v}"
|
||||
echo "raw=$RAW_TAG" >> "$GITHUB_OUTPUT"
|
||||
echo "clean=$CLEAN_TAG" >> "$GITHUB_OUTPUT"
|
||||
|
||||
- name: Set image namespace
|
||||
id: ns
|
||||
shell: bash
|
||||
run: |
|
||||
NS="${IMAGE_NAMESPACE}"
|
||||
if [ -z "$NS" ]; then
|
||||
NS="${{ secrets.DOCKERHUB_USERNAME }}"
|
||||
fi
|
||||
if [ -z "$NS" ]; then
|
||||
echo "Missing Docker Hub namespace. Set repo var DOCKERHUB_NAMESPACE or secret DOCKERHUB_USERNAME."
|
||||
exit 1
|
||||
fi
|
||||
echo "value=$NS" >> "$GITHUB_OUTPUT"
|
||||
|
||||
- name: Set up Docker Buildx
|
||||
uses: docker/setup-buildx-action@v3
|
||||
|
||||
- name: Login to Docker Hub
|
||||
uses: docker/login-action@v3
|
||||
with:
|
||||
username: ${{ secrets.DOCKERHUB_USERNAME }}
|
||||
password: ${{ secrets.DOCKERHUB_TOKEN }}
|
||||
|
||||
- name: Build and push backend image
|
||||
uses: docker/build-push-action@v6
|
||||
with:
|
||||
context: ./backend
|
||||
file: ./backend/Dockerfile
|
||||
push: true
|
||||
tags: |
|
||||
${{ steps.ns.outputs.value }}/nexapg-backend:${{ steps.ver.outputs.clean }}
|
||||
${{ steps.ns.outputs.value }}/nexapg-backend:latest
|
||||
cache-from: type=registry,ref=${{ steps.ns.outputs.value }}/nexapg-backend:buildcache
|
||||
cache-to: type=registry,ref=${{ steps.ns.outputs.value }}/nexapg-backend:buildcache,mode=max
|
||||
|
||||
- name: Build and push frontend image
|
||||
uses: docker/build-push-action@v6
|
||||
with:
|
||||
context: ./frontend
|
||||
file: ./frontend/Dockerfile
|
||||
push: true
|
||||
build-args: |
|
||||
VITE_API_URL=/api/v1
|
||||
tags: |
|
||||
${{ steps.ns.outputs.value }}/nexapg-frontend:${{ steps.ver.outputs.clean }}
|
||||
${{ steps.ns.outputs.value }}/nexapg-frontend:latest
|
||||
cache-from: type=registry,ref=${{ steps.ns.outputs.value }}/nexapg-frontend:buildcache
|
||||
cache-to: type=registry,ref=${{ steps.ns.outputs.value }}/nexapg-frontend:buildcache,mode=max
|
||||
3
Makefile
3
Makefile
@@ -1,7 +1,8 @@
|
||||
.PHONY: up down logs migrate
|
||||
|
||||
up:
|
||||
docker compose up -d --build
|
||||
docker compose pull
|
||||
docker compose up -d
|
||||
|
||||
down:
|
||||
docker compose down
|
||||
|
||||
58
README.md
58
README.md
@@ -9,7 +9,7 @@ It combines FastAPI, React, and PostgreSQL in a Docker Compose stack with RBAC,
|
||||
|
||||
## Table of Contents
|
||||
|
||||
- [Quick Start](#quick-start)
|
||||
- [Quick Deploy (Prebuilt Images)](#quick-deploy-prebuilt-images)
|
||||
- [Prerequisites](#prerequisites)
|
||||
- [Make Commands](#make-commands)
|
||||
- [Configuration Reference (`.env`)](#configuration-reference-env)
|
||||
@@ -93,27 +93,50 @@ Optional:
|
||||
|
||||
- `psql` for manual DB checks
|
||||
|
||||
## Quick Start
|
||||
## Quick Deploy (Prebuilt Images)
|
||||
|
||||
1. Copy environment template:
|
||||
If you only want to run NexaPG from published Docker Hub images, use the bootstrap script:
|
||||
|
||||
```bash
|
||||
cp .env.example .env
|
||||
mkdir -p /opt/NexaPG
|
||||
cd /opt/NexaPG
|
||||
wget -O bootstrap-compose.sh https://git.nesterovic.cc/nessi/NexaPG/raw/branch/main/ops/scripts/bootstrap-compose.sh
|
||||
chmod +x bootstrap-compose.sh
|
||||
./bootstrap-compose.sh
|
||||
```
|
||||
|
||||
2. Generate a Fernet key and set `ENCRYPTION_KEY` in `.env`:
|
||||
This downloads:
|
||||
|
||||
- `docker-compose.yml`
|
||||
- `.env.example`
|
||||
- `Makefile`
|
||||
|
||||
Then:
|
||||
|
||||
```bash
|
||||
# generate JWT secret
|
||||
python -c "import secrets; print(secrets.token_urlsafe(64))"
|
||||
# generate Fernet encryption key
|
||||
python -c "from cryptography.fernet import Fernet; print(Fernet.generate_key().decode())"
|
||||
```
|
||||
|
||||
3. Start the stack:
|
||||
|
||||
```bash
|
||||
# put both values into .env (JWT_SECRET_KEY / ENCRYPTION_KEY)
|
||||
# note: .env is auto-created by bootstrap if it does not exist
|
||||
make up
|
||||
```
|
||||
|
||||
4. Open the application:
|
||||
Manual download alternative:
|
||||
|
||||
```bash
|
||||
mkdir -p /opt/NexaPG
|
||||
cd /opt/NexaPG
|
||||
wget https://git.nesterovic.cc/nessi/NexaPG/raw/branch/main/docker-compose.yml
|
||||
wget https://git.nesterovic.cc/nessi/NexaPG/raw/branch/main/.env.example
|
||||
wget https://git.nesterovic.cc/nessi/NexaPG/raw/branch/main/Makefile
|
||||
cp .env.example .env
|
||||
```
|
||||
|
||||
`make up` pulls `nesterovicit/nexapg-backend:latest` and `nesterovicit/nexapg-frontend:latest`, then starts the stack.
|
||||
|
||||
Open the application:
|
||||
|
||||
- Frontend: `http://<SERVER_IP>:<FRONTEND_PORT>`
|
||||
- API base: `http://<SERVER_IP>:<BACKEND_PORT>/api/v1`
|
||||
@@ -127,7 +150,7 @@ Initial admin bootstrap user (created from `.env` if missing):
|
||||
## Make Commands
|
||||
|
||||
```bash
|
||||
make up # build and start all services
|
||||
make up # pull latest images and start all services
|
||||
make down # stop all services
|
||||
make logs # follow compose logs
|
||||
make migrate # optional/manual: run alembic upgrade head in backend container
|
||||
@@ -183,12 +206,6 @@ Note: Migrations run automatically when the backend container starts (`entrypoin
|
||||
| Variable | Description |
|
||||
|---|---|
|
||||
| `FRONTEND_PORT` | Host port mapped to frontend container port `80` |
|
||||
| `VITE_API_URL` | Frontend API base URL (build-time) |
|
||||
|
||||
Recommended values for `VITE_API_URL`:
|
||||
|
||||
- Reverse proxy setup: `/api/v1`
|
||||
- Direct backend access: `http://<SERVER_IP>:<BACKEND_PORT>/api/v1`
|
||||
|
||||
## Core Functional Areas
|
||||
|
||||
@@ -318,7 +335,7 @@ For production, serve frontend and API under the same public origin via reverse
|
||||
|
||||
- Frontend URL example: `https://monitor.example.com`
|
||||
- Proxy API path `/api/` to backend service
|
||||
- Use `VITE_API_URL=/api/v1`
|
||||
- Route `/api/v1` to the backend service
|
||||
|
||||
This prevents mixed-content and CORS issues.
|
||||
|
||||
@@ -351,8 +368,7 @@ docker compose logs --tail=200 db
|
||||
|
||||
### CORS or mixed-content issues behind SSL proxy
|
||||
|
||||
- Set `VITE_API_URL=/api/v1`
|
||||
- Ensure proxy forwards `/api/` to backend
|
||||
- Ensure proxy forwards `/api/` (or `/api/v1`) to backend
|
||||
- Set correct frontend origin(s) in `CORS_ORIGINS`
|
||||
|
||||
### `rejected SSL upgrade` for a target
|
||||
|
||||
@@ -2,7 +2,7 @@ from functools import lru_cache
|
||||
from pydantic import field_validator
|
||||
from pydantic_settings import BaseSettings, SettingsConfigDict
|
||||
|
||||
NEXAPG_VERSION = "0.1.2"
|
||||
NEXAPG_VERSION = "0.1.3"
|
||||
|
||||
|
||||
class Settings(BaseSettings):
|
||||
|
||||
@@ -17,10 +17,10 @@ class DiskSpaceProvider:
|
||||
class NullDiskSpaceProvider(DiskSpaceProvider):
|
||||
async def get_free_bytes(self, target_host: str) -> DiskSpaceProbeResult:
|
||||
return DiskSpaceProbeResult(
|
||||
source="none",
|
||||
source="agentless",
|
||||
status="unavailable",
|
||||
free_bytes=None,
|
||||
message=f"No infra probe configured for host {target_host}. Add SSH/Agent provider later.",
|
||||
message=f"Agentless mode: host-level free disk is not available for {target_host}.",
|
||||
)
|
||||
|
||||
|
||||
|
||||
@@ -18,8 +18,8 @@ services:
|
||||
retries: 10
|
||||
|
||||
backend:
|
||||
build:
|
||||
context: ./backend
|
||||
image: nesterovicit/nexapg-backend:latest
|
||||
pull_policy: always
|
||||
container_name: nexapg-backend
|
||||
restart: unless-stopped
|
||||
environment:
|
||||
@@ -47,10 +47,8 @@ services:
|
||||
- "${BACKEND_PORT}:8000"
|
||||
|
||||
frontend:
|
||||
build:
|
||||
context: ./frontend
|
||||
args:
|
||||
VITE_API_URL: ${VITE_API_URL}
|
||||
image: nesterovicit/nexapg-frontend:latest
|
||||
pull_policy: always
|
||||
container_name: nexapg-frontend
|
||||
restart: unless-stopped
|
||||
depends_on:
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
import React, { useEffect, useState } from "react";
|
||||
import React, { useEffect, useRef, useState } from "react";
|
||||
import { apiFetch } from "../api";
|
||||
import { useAuth } from "../state";
|
||||
|
||||
@@ -62,6 +62,7 @@ function buildQueryTips(row) {
|
||||
|
||||
export function QueryInsightsPage() {
|
||||
const { tokens, refresh } = useAuth();
|
||||
const refreshRef = useRef(refresh);
|
||||
const [targets, setTargets] = useState([]);
|
||||
const [targetId, setTargetId] = useState("");
|
||||
const [rows, setRows] = useState([]);
|
||||
@@ -71,6 +72,10 @@ export function QueryInsightsPage() {
|
||||
const [error, setError] = useState("");
|
||||
const [loading, setLoading] = useState(true);
|
||||
|
||||
useEffect(() => {
|
||||
refreshRef.current = refresh;
|
||||
}, [refresh]);
|
||||
|
||||
useEffect(() => {
|
||||
(async () => {
|
||||
try {
|
||||
@@ -89,17 +94,26 @@ export function QueryInsightsPage() {
|
||||
|
||||
useEffect(() => {
|
||||
if (!targetId) return;
|
||||
let active = true;
|
||||
(async () => {
|
||||
try {
|
||||
const data = await apiFetch(`/targets/${targetId}/top-queries`, {}, tokens, refresh);
|
||||
const data = await apiFetch(`/targets/${targetId}/top-queries`, {}, tokens, refreshRef.current);
|
||||
if (!active) return;
|
||||
setRows(data);
|
||||
setSelectedQuery(data[0] || null);
|
||||
setPage(1);
|
||||
setSelectedQuery((prev) => {
|
||||
if (!prev) return data[0] || null;
|
||||
const keep = data.find((row) => row.queryid === prev.queryid);
|
||||
return keep || data[0] || null;
|
||||
});
|
||||
setPage((prev) => (prev === 1 ? prev : 1));
|
||||
} catch (e) {
|
||||
setError(String(e.message || e));
|
||||
if (active) setError(String(e.message || e));
|
||||
}
|
||||
})();
|
||||
}, [targetId, tokens, refresh]);
|
||||
return () => {
|
||||
active = false;
|
||||
};
|
||||
}, [targetId, tokens?.accessToken, tokens?.refreshToken]);
|
||||
|
||||
const dedupedByQueryId = [...rows].reduce((acc, row) => {
|
||||
if (!row?.queryid) return acc;
|
||||
|
||||
@@ -41,6 +41,19 @@ function formatNumber(value, digits = 2) {
|
||||
return Number(value).toFixed(digits);
|
||||
}
|
||||
|
||||
function formatHostMetricUnavailable() {
|
||||
return "N/A (agentless)";
|
||||
}
|
||||
|
||||
function formatDiskSpaceAgentless(diskSpace) {
|
||||
if (!diskSpace) return formatHostMetricUnavailable();
|
||||
if (diskSpace.free_bytes !== null && diskSpace.free_bytes !== undefined) {
|
||||
return formatBytes(diskSpace.free_bytes);
|
||||
}
|
||||
if (diskSpace.status === "unavailable") return formatHostMetricUnavailable();
|
||||
return "-";
|
||||
}
|
||||
|
||||
function MetricsTooltip({ active, payload, label }) {
|
||||
if (!active || !payload || payload.length === 0) return null;
|
||||
const row = payload[0]?.payload || {};
|
||||
@@ -346,6 +359,9 @@ export function TargetDetailPage() {
|
||||
{uiMode === "dba" && overview && (
|
||||
<div className="card">
|
||||
<h3>Database Overview</h3>
|
||||
<p className="muted" style={{ marginTop: 2 }}>
|
||||
Agentless mode: host-level CPU, RAM, and free-disk metrics are not available.
|
||||
</p>
|
||||
<div className="grid three overview-kv">
|
||||
<div><span>PostgreSQL Version</span><strong>{overview.instance.server_version || "-"}</strong></div>
|
||||
<div>
|
||||
@@ -366,8 +382,8 @@ export function TargetDetailPage() {
|
||||
<div title="Total WAL directory size (when available)">
|
||||
<span>WAL Size</span><strong>{formatBytes(overview.storage.wal_directory_size_bytes)}</strong>
|
||||
</div>
|
||||
<div title="Optional metric via future Agent/SSH provider">
|
||||
<span>Free Disk</span><strong>{formatBytes(overview.storage.disk_space.free_bytes)}</strong>
|
||||
<div title={overview.storage.disk_space?.message || "Agentless mode: host-level free disk is unavailable."}>
|
||||
<span>Free Disk</span><strong>{formatDiskSpaceAgentless(overview.storage.disk_space)}</strong>
|
||||
</div>
|
||||
<div title="Replication replay delay on standby">
|
||||
<span>Replay Lag</span>
|
||||
@@ -378,6 +394,12 @@ export function TargetDetailPage() {
|
||||
<div><span>Replication Slots</span><strong>{overview.replication.replication_slots_count ?? "-"}</strong></div>
|
||||
<div><span>Repl Clients</span><strong>{overview.replication.active_replication_clients ?? "-"}</strong></div>
|
||||
<div><span>Autovacuum Workers</span><strong>{overview.performance.autovacuum_workers ?? "-"}</strong></div>
|
||||
<div title="Host CPU requires OS-level telemetry">
|
||||
<span>Host CPU</span><strong>{formatHostMetricUnavailable()}</strong>
|
||||
</div>
|
||||
<div title="Host RAM requires OS-level telemetry">
|
||||
<span>Host RAM</span><strong>{formatHostMetricUnavailable()}</strong>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div className="grid two">
|
||||
|
||||
41
ops/scripts/bootstrap-compose.sh
Normal file
41
ops/scripts/bootstrap-compose.sh
Normal file
@@ -0,0 +1,41 @@
|
||||
#!/usr/bin/env bash
|
||||
set -euo pipefail
|
||||
|
||||
# Usage:
|
||||
# bash bootstrap-compose.sh
|
||||
# BASE_URL="https://git.nesterovic.cc/nessi/NexaPG/raw/branch/main" bash bootstrap-compose.sh
|
||||
|
||||
BASE_URL="${BASE_URL:-https://git.nesterovic.cc/nessi/NexaPG/raw/branch/main}"
|
||||
|
||||
echo "[bootstrap] Using base URL: ${BASE_URL}"
|
||||
|
||||
fetch_file() {
|
||||
local path="$1"
|
||||
local out="$2"
|
||||
|
||||
if command -v wget >/dev/null 2>&1; then
|
||||
wget -q -O "${out}" "${BASE_URL}/${path}"
|
||||
elif command -v curl >/dev/null 2>&1; then
|
||||
curl -fsSL "${BASE_URL}/${path}" -o "${out}"
|
||||
else
|
||||
echo "[bootstrap] ERROR: wget or curl is required"
|
||||
exit 1
|
||||
fi
|
||||
}
|
||||
|
||||
fetch_file "docker-compose.yml" "docker-compose.yml"
|
||||
fetch_file ".env.example" ".env.example"
|
||||
fetch_file "Makefile" "Makefile"
|
||||
|
||||
if [[ ! -f ".env" ]]; then
|
||||
cp .env.example .env
|
||||
echo "[bootstrap] Created .env from .env.example"
|
||||
else
|
||||
echo "[bootstrap] .env already exists, keeping it"
|
||||
fi
|
||||
|
||||
echo
|
||||
echo "[bootstrap] Next steps:"
|
||||
echo " 1) Edit .env (set JWT_SECRET_KEY and ENCRYPTION_KEY at minimum)"
|
||||
echo " 2) Run: make up"
|
||||
echo
|
||||
Reference in New Issue
Block a user