Compare commits
2 Commits
2e719fc029
...
1eb0aa3658
| Author | SHA1 | Date | |
|---|---|---|---|
| 1eb0aa3658 | |||
| 4e5aff1c02 |
45
.dockerignore
Normal file
45
.dockerignore
Normal file
@@ -0,0 +1,45 @@
|
||||
# Git
|
||||
.git
|
||||
.gitignore
|
||||
|
||||
# Documentation
|
||||
README.md
|
||||
*.md
|
||||
|
||||
# Python
|
||||
__pycache__/
|
||||
*.py[cod]
|
||||
*$py.class
|
||||
*.so
|
||||
.Python
|
||||
env/
|
||||
venv/
|
||||
ENV/
|
||||
env.bak/
|
||||
venv.bak/
|
||||
|
||||
# IDEs
|
||||
.vscode/
|
||||
.idea/
|
||||
*.swp
|
||||
*.swo
|
||||
*~
|
||||
|
||||
# OS
|
||||
.DS_Store
|
||||
Thumbs.db
|
||||
|
||||
# Data folders (these will be mounted as volumes)
|
||||
data/
|
||||
|
||||
# Logs
|
||||
*.log
|
||||
|
||||
# Environment files
|
||||
.env
|
||||
.env.local
|
||||
.env.example
|
||||
|
||||
# Temporary files
|
||||
tmp/
|
||||
temp/
|
||||
44
.env.example
44
.env.example
@@ -1,21 +1,41 @@
|
||||
# .env - Flask environment variables
|
||||
# DigiServer Environment Configuration
|
||||
# Copy this file to .env and modify the values as needed
|
||||
|
||||
# Flask secret key (change this to something secure in production)
|
||||
SECRET_KEY=Ana_Are_Multe_Mere-Si_Nu_Are_Pere
|
||||
# Flask Configuration
|
||||
FLASK_APP=app.py
|
||||
FLASK_RUN_HOST=0.0.0.0
|
||||
FLASK_ENV=production
|
||||
|
||||
# Flask environment: development or production
|
||||
FLASK_ENV=development
|
||||
# Security
|
||||
SECRET_KEY=Ma_Duc_Dupa_Merele_Lui_Ana
|
||||
# Change this to a secure random string in production!
|
||||
|
||||
# Database location (optional, defaults to instance/dashboard.db)
|
||||
# Default Admin User
|
||||
ADMIN_USER=admin
|
||||
ADMIN_PASSWORD=Initial01!
|
||||
# Change the default password after first login!
|
||||
|
||||
# Database Configuration
|
||||
# SQLite database file will be created in data/instance/dashboard.db
|
||||
# SQLALCHEMY_DATABASE_URI=sqlite:///instance/dashboard.db
|
||||
|
||||
# Default admin user credentials (used for auto-creation)
|
||||
DEFAULT_USER=admin
|
||||
DEFAULT_PASSWORD=1234
|
||||
# Application Settings
|
||||
MAX_CONTENT_LENGTH=2147483648 # 2GB in bytes
|
||||
UPLOAD_FOLDER=static/uploads
|
||||
UPLOAD_FOLDERLOGO=static/resurse
|
||||
|
||||
# Flask server settings
|
||||
# Server Information
|
||||
SERVER_VERSION=1.1.0
|
||||
BUILD_DATE=2025-06-29
|
||||
|
||||
# Docker Configuration (for docker-compose.yml)
|
||||
DIGISERVER_PORT=8880
|
||||
CONTAINER_NAME=digiserver
|
||||
|
||||
# Flask server settings (for development)
|
||||
HOST=0.0.0.0
|
||||
PORT=5000
|
||||
|
||||
# Maximum upload size (in bytes, 2GB)
|
||||
MAX_CONTENT_LENGTH=2147483648
|
||||
# Optional: External Database (for advanced users)
|
||||
# DATABASE_URL=postgresql://user:password@localhost/digiserver
|
||||
# DATABASE_URL=mysql://user:password@localhost/digiserver
|
||||
37
.gitignore
vendored
37
.gitignore
vendored
@@ -1,4 +1,41 @@
|
||||
# Python
|
||||
__pycache__/
|
||||
*.py[cod]
|
||||
*$py.class
|
||||
*.so
|
||||
.Python
|
||||
env/
|
||||
venv/
|
||||
ENV/
|
||||
env.bak/
|
||||
venv.bak/
|
||||
|
||||
# Environment files
|
||||
.env
|
||||
.env.local
|
||||
venv/
|
||||
# Data directories (persistent storage)
|
||||
data/
|
||||
instance/
|
||||
instance.bak/
|
||||
|
||||
# Legacy directories (can be removed after migration)
|
||||
digiscreen/
|
||||
|
||||
# IDE
|
||||
.vscode/
|
||||
.idea/
|
||||
|
||||
# OS
|
||||
.DS_Store
|
||||
Thumbs.db
|
||||
|
||||
# Logs
|
||||
*.log
|
||||
|
||||
# Backups
|
||||
backups/
|
||||
|
||||
# Temporary files
|
||||
tmp/
|
||||
temp/
|
||||
|
||||
92
DEPLOYMENT.md
Normal file
92
DEPLOYMENT.md
Normal file
@@ -0,0 +1,92 @@
|
||||
# DigiServer v1.1.0 - Production Deployment Guide
|
||||
|
||||
## 🎯 Ready for Deployment
|
||||
|
||||
Your DigiServer application has been cleaned and prepared for Docker deployment.
|
||||
|
||||
### ✅ What's Been Prepared
|
||||
|
||||
1. **Application Cleaned**
|
||||
- Python cache files removed (`__pycache__/`, `*.pyc`)
|
||||
- Development artifacts cleaned
|
||||
- Production-ready structure
|
||||
|
||||
2. **Docker Configuration**
|
||||
- Dockerfile optimized with LibreOffice and poppler-utils
|
||||
- docker-compose.yml configured for production
|
||||
- .dockerignore updated to exclude development files
|
||||
- Data persistence configured via volumes
|
||||
|
||||
3. **Deployment Scripts**
|
||||
- `deploy-docker.sh` - Automated deployment script
|
||||
- `cleanup-docker.sh` - Complete cleanup script
|
||||
- Both scripts use modern `docker compose` syntax
|
||||
|
||||
4. **Data Structure**
|
||||
- `./data/instance/` - Database files
|
||||
- `./data/uploads/` - Media uploads
|
||||
- `./data/resurse/` - System resources
|
||||
- All directories auto-created and volume-mounted
|
||||
|
||||
### 🚀 Quick Deployment
|
||||
|
||||
```bash
|
||||
# Deploy DigiServer
|
||||
./deploy-docker.sh
|
||||
|
||||
# Access at: http://localhost:8880
|
||||
# Username: admin
|
||||
# Password: Initial01!
|
||||
```
|
||||
|
||||
### 📋 Features Ready
|
||||
|
||||
- ✅ **Document Processing**: LibreOffice + poppler-utils integrated
|
||||
- ✅ **File Uploads**: PPTX → PDF → 4K JPG workflow
|
||||
- ✅ **Path Resolution**: Absolute path handling for containerized deployment
|
||||
- ✅ **File Management**: Bulk delete functions with physical file cleanup
|
||||
- ✅ **User Management**: Admin user auto-creation
|
||||
- ✅ **Data Persistence**: Volume-mounted data directories
|
||||
- ✅ **Health Checks**: Container health monitoring
|
||||
- ✅ **Production Logging**: Structured output and error handling
|
||||
|
||||
### 🔧 System Requirements
|
||||
|
||||
- Docker Engine 20.10+
|
||||
- Docker Compose v2 (plugin)
|
||||
- 2GB RAM minimum
|
||||
- 10GB disk space
|
||||
|
||||
### 📁 Deployment Structure
|
||||
|
||||
```
|
||||
digiserver/
|
||||
├── 📁 app/ # Application code
|
||||
├── 📁 data/ # Persistent data (auto-created)
|
||||
│ ├── 📁 instance/ # Database
|
||||
│ ├── 📁 uploads/ # Media files
|
||||
│ └── 📁 resurse/ # Resources
|
||||
├── 🐳 Dockerfile # Production image
|
||||
├── 🔧 docker-compose.yml # Container orchestration
|
||||
├── 🚀 deploy-docker.sh # Deployment script
|
||||
├── 🧹 cleanup-docker.sh # Cleanup script
|
||||
└── 📖 README.md # Documentation
|
||||
```
|
||||
|
||||
### 🔐 Security Notes
|
||||
|
||||
- Change default password after first login
|
||||
- SECRET_KEY configured for session security
|
||||
- File upload restrictions in place
|
||||
- Container runs with proper permissions
|
||||
|
||||
### 📊 Monitoring
|
||||
|
||||
- Health checks configured (30s intervals)
|
||||
- Container auto-restart on failure
|
||||
- Logs available via `docker compose logs -f`
|
||||
- Status monitoring with `docker compose ps`
|
||||
|
||||
---
|
||||
|
||||
**Next Step**: Run `./deploy-docker.sh` to deploy your DigiServer! 🚀
|
||||
56
Dockerfile
56
Dockerfile
@@ -1,31 +1,59 @@
|
||||
# Use Python 3.11 slim image
|
||||
FROM python:3.11-slim
|
||||
|
||||
# Set working directory
|
||||
WORKDIR /app
|
||||
|
||||
# Install system dependencies, including Rust and build tools
|
||||
# Install system dependencies including LibreOffice and poppler-utils
|
||||
RUN apt-get update && apt-get install -y \
|
||||
libreoffice poppler-utils ffmpeg \
|
||||
libpoppler-cpp-dev libmagic1 \
|
||||
libffi-dev libssl-dev g++ curl libjpeg-dev zlib1g-dev \
|
||||
libxml2-dev libxslt-dev build-essential cargo \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
poppler-utils \
|
||||
libreoffice \
|
||||
ffmpeg \
|
||||
libpoppler-cpp-dev \
|
||||
libmagic1 \
|
||||
libffi-dev \
|
||||
libssl-dev \
|
||||
g++ \
|
||||
curl \
|
||||
libjpeg-dev \
|
||||
zlib1g-dev \
|
||||
libxml2-dev \
|
||||
libxslt-dev \
|
||||
build-essential \
|
||||
cargo \
|
||||
fonts-dejavu-core \
|
||||
&& rm -rf /var/lib/apt/lists/* \
|
||||
&& apt-get clean
|
||||
|
||||
# Debug: Verify Rust installation
|
||||
RUN rustc --version && cargo --version
|
||||
|
||||
# Copy application files
|
||||
COPY . /app
|
||||
# Verify LibreOffice and poppler-utils installation
|
||||
RUN libreoffice --version && pdftoppm -v
|
||||
|
||||
# Copy entrypoint script and make it executable
|
||||
COPY entrypoint.sh /entrypoint.sh
|
||||
RUN chmod +x /entrypoint.sh
|
||||
# Copy requirements first for better layer caching
|
||||
COPY app/requirements.txt .
|
||||
|
||||
# Upgrade pip and install Python dependencies (using piwheels for ARM)
|
||||
# Upgrade pip and install Python dependencies
|
||||
RUN python -m pip install --upgrade pip && \
|
||||
pip install --no-cache-dir --extra-index-url https://www.piwheels.org/simple -r requirements.txt
|
||||
pip install --no-cache-dir -r requirements.txt
|
||||
|
||||
# Copy application code
|
||||
COPY app/ .
|
||||
|
||||
# Make entrypoint script executable
|
||||
RUN chmod +x entrypoint.sh
|
||||
|
||||
# Create necessary directories for volumes
|
||||
RUN mkdir -p /app/static/uploads /app/static/resurse /app/instance
|
||||
|
||||
# Expose the application port
|
||||
EXPOSE 5000
|
||||
|
||||
# Set environment variables
|
||||
ENV FLASK_APP=app.py
|
||||
ENV FLASK_RUN_HOST=0.0.0.0
|
||||
ENV PYTHONPATH=/app
|
||||
|
||||
# Use entrypoint script
|
||||
ENTRYPOINT ["/entrypoint.sh"]
|
||||
ENTRYPOINT ["./entrypoint.sh"]
|
||||
258
README.md
Normal file
258
README.md
Normal file
@@ -0,0 +1,258 @@
|
||||
# DigiServer - Digital Signage Management Platform
|
||||
|
||||

|
||||

|
||||

|
||||

|
||||
|
||||
DigiServer is a comprehensive digital signage management platform built with Flask. It allows you to manage digital displays, create playlists, organize content into groups, and control multiple players from a centralized dashboard.
|
||||
|
||||
## 🚀 Features
|
||||
|
||||
- **Multi-Player Management**: Control multiple digital signage players from a single dashboard
|
||||
- **Group Management**: Organize players into groups for synchronized content
|
||||
- **Content Management**: Upload and manage various media types (images, videos, PDFs, PowerPoint presentations)
|
||||
- **Real-time Updates**: Players automatically sync with the latest content
|
||||
- **User Management**: Admin and user role-based access control
|
||||
- **Orientation Support**: Configure display orientation (Landscape/Portrait) per player and group
|
||||
- **API Integration**: RESTful API for player authentication and playlist retrieval
|
||||
- **Docker Support**: Easy deployment with Docker containers
|
||||
|
||||
## 📋 Requirements
|
||||
|
||||
- Docker and Docker Compose
|
||||
- Python 3.11+ (if running without Docker)
|
||||
- FFmpeg (for video processing)
|
||||
- LibreOffice (for document conversion)
|
||||
|
||||
## 📁 Project Structure
|
||||
|
||||
```
|
||||
digiserver/
|
||||
├── app/ # Application code
|
||||
│ ├── models/ # Database models
|
||||
│ ├── templates/ # HTML templates
|
||||
│ ├── utils/ # Utility functions
|
||||
│ ├── app.py # Main Flask application
|
||||
│ ├── extensions.py # Flask extensions
|
||||
│ ├── requirements.txt # Python dependencies
|
||||
│ └── entrypoint.sh # Container entry point
|
||||
├── data/ # Persistent data (created on first run)
|
||||
│ ├── instance/ # Database files
|
||||
│ ├── uploads/ # Media uploads
|
||||
│ └── resurse/ # System resources (logos, etc.)
|
||||
├── docker-compose.yml # Docker Compose configuration
|
||||
├── Dockerfile # Docker image definition
|
||||
└── README.md # This file
|
||||
```
|
||||
|
||||
## 🐳 Quick Start with Docker
|
||||
|
||||
### Automated Deployment (Recommended)
|
||||
|
||||
1. **Clone the repository**
|
||||
```bash
|
||||
git clone <repository-url>
|
||||
cd digiserver
|
||||
```
|
||||
|
||||
2. **Deploy with automated script**
|
||||
```bash
|
||||
./deploy-docker.sh
|
||||
```
|
||||
|
||||
This script will:
|
||||
- Check Docker requirements
|
||||
- Build the DigiServer image
|
||||
- Create necessary data directories
|
||||
- Start the containers
|
||||
- Display access information
|
||||
|
||||
3. **Access the application**
|
||||
- Open your browser and navigate to `http://localhost:8880`
|
||||
- Default admin credentials:
|
||||
- Username: `admin`
|
||||
- Password: `Initial01!`
|
||||
|
||||
### Manual Docker Commands
|
||||
|
||||
Alternatively, you can use Docker commands directly:
|
||||
|
||||
```bash
|
||||
# Build and start
|
||||
docker compose up -d
|
||||
|
||||
# Stop
|
||||
docker compose down
|
||||
|
||||
# View logs
|
||||
docker compose logs -f
|
||||
|
||||
# Check status
|
||||
docker compose ps
|
||||
```
|
||||
|
||||
### Clean Up
|
||||
|
||||
To completely remove DigiServer containers and images:
|
||||
```bash
|
||||
./cleanup-docker.sh
|
||||
```
|
||||
|
||||
## 🔧 Configuration
|
||||
|
||||
### Environment Variables
|
||||
|
||||
You can customize the application by modifying the environment variables in `docker-compose.yml`:
|
||||
|
||||
- `ADMIN_USER`: Default admin username (default: admin)
|
||||
- `ADMIN_PASSWORD`: Default admin password (default: Initial01!)
|
||||
- `SECRET_KEY`: Flask secret key for session security
|
||||
- `FLASK_APP`: Flask application entry point
|
||||
- `FLASK_RUN_HOST`: Host to bind the Flask application
|
||||
|
||||
### Data Persistence
|
||||
|
||||
All persistent data is stored in the `data/` folder:
|
||||
- `data/instance/`: SQLite database files
|
||||
- `data/uploads/`: Uploaded media files
|
||||
- `data/resurse/`: System resources (logo, login images)
|
||||
|
||||
This folder will be created automatically on first run and persists between container restarts.
|
||||
|
||||
## 💻 Manual Installation (Development)
|
||||
|
||||
If you prefer to run without Docker:
|
||||
|
||||
1. **Install system dependencies**
|
||||
```bash
|
||||
# Ubuntu/Debian
|
||||
sudo apt-get update
|
||||
sudo apt-get install python3.11 python3-pip libreoffice ffmpeg
|
||||
|
||||
# CentOS/RHEL
|
||||
sudo yum install python3.11 python3-pip libreoffice ffmpeg
|
||||
```
|
||||
|
||||
2. **Install Python dependencies**
|
||||
```bash
|
||||
cd app/
|
||||
pip install -r requirements.txt
|
||||
```
|
||||
|
||||
3. **Run the application**
|
||||
```bash
|
||||
python app.py
|
||||
```
|
||||
|
||||
## 🎮 Usage
|
||||
|
||||
### Managing Players
|
||||
|
||||
1. **Add a Player**: Navigate to the dashboard and click "Add Player"
|
||||
2. **Configure Player**: Set username, hostname, passwords, and orientation
|
||||
3. **Upload Content**: Upload media files to the player's playlist
|
||||
4. **Player Authentication**: Players can authenticate using hostname and password/quickconnect code
|
||||
|
||||
### Managing Groups
|
||||
|
||||
1. **Create Group**: Group multiple players for synchronized content
|
||||
2. **Assign Players**: Add/remove players from groups
|
||||
3. **Upload Group Content**: Upload content that will be shared across all players in the group
|
||||
4. **Group Display**: View group content in fullscreen mode
|
||||
|
||||
### Content Types Supported
|
||||
|
||||
- **Images**: JPG, PNG, GIF
|
||||
- **Videos**: MP4, AVI, MOV (automatically converted to MP4)
|
||||
- **Documents**: PDF (converted to images)
|
||||
- **Presentations**: PPTX (converted to images)
|
||||
|
||||
## 🔌 API Endpoints
|
||||
|
||||
### Player API
|
||||
|
||||
- `GET /api/playlists?hostname={hostname}&quickconnect_code={code}`: Get player playlist
|
||||
- `GET /api/playlist_version?hostname={hostname}&quickconnect_code={code}`: Get playlist version
|
||||
- `GET /media/{filename}`: Serve media files
|
||||
|
||||
### Authentication
|
||||
|
||||
Players authenticate using:
|
||||
- **Hostname**: Unique identifier for the player
|
||||
- **Password**: Primary authentication method
|
||||
- **Quickconnect Code**: Alternative authentication method
|
||||
|
||||
## 🛠️ Development
|
||||
|
||||
### Building the Docker Image
|
||||
|
||||
```bash
|
||||
docker build -t digiserver:latest .
|
||||
```
|
||||
|
||||
### Running Tests
|
||||
|
||||
```bash
|
||||
# Install test dependencies
|
||||
pip install pytest pytest-flask
|
||||
|
||||
# Run tests
|
||||
pytest
|
||||
```
|
||||
|
||||
### Database Management
|
||||
|
||||
The application uses SQLite with Flask-Migrate for database management:
|
||||
|
||||
```bash
|
||||
# Initialize database
|
||||
flask db init
|
||||
|
||||
# Create migration
|
||||
flask db migrate -m "Description of changes"
|
||||
|
||||
# Apply migration
|
||||
flask db upgrade
|
||||
```
|
||||
|
||||
## 🔒 Security
|
||||
|
||||
- **User Authentication**: Role-based access control (admin/user)
|
||||
- **Player Authentication**: Secure hostname and password-based authentication
|
||||
- **File Upload Security**: Secure filename handling and file type validation
|
||||
- **Session Management**: Secure session handling with configurable secret key
|
||||
|
||||
## 📊 Monitoring
|
||||
|
||||
- **Server Logs**: View recent server activities from the dashboard
|
||||
- **Health Check**: Docker health check endpoint for monitoring
|
||||
- **Content Management**: Track content usage and cleanup unused files
|
||||
|
||||
## 🤝 Contributing
|
||||
|
||||
1. Fork the repository
|
||||
2. Create a feature branch (`git checkout -b feature/amazing-feature`)
|
||||
3. Commit your changes (`git commit -m 'Add some amazing feature'`)
|
||||
4. Push to the branch (`git push origin feature/amazing-feature`)
|
||||
5. Open a Pull Request
|
||||
|
||||
## 📝 License
|
||||
|
||||
This project is licensed under the MIT License - see the LICENSE file for details.
|
||||
|
||||
## 🆘 Support
|
||||
|
||||
For support and questions:
|
||||
- Create an issue in the repository
|
||||
- Check the documentation in the `docs/` folder
|
||||
- Review the application logs for troubleshooting
|
||||
|
||||
## 🔄 Version History
|
||||
|
||||
- **1.1.0** (2025-06-29): Added orientation support, improved group management
|
||||
- **1.0.0**: Initial release with basic digital signage functionality
|
||||
|
||||
---
|
||||
|
||||
**Note**: Make sure to change the default admin password after first login for security purposes.
|
||||
Binary file not shown.
@@ -1,25 +1,21 @@
|
||||
import os
|
||||
import click
|
||||
import time
|
||||
import psutil
|
||||
import threading
|
||||
from flask import Flask, render_template, request, redirect, url_for, session, flash, jsonify, send_from_directory
|
||||
import shutil
|
||||
import zipfile
|
||||
import tempfile
|
||||
from flask import Flask, render_template, request, redirect, url_for, session, flash, jsonify, send_from_directory, send_file
|
||||
from flask_migrate import Migrate
|
||||
import subprocess
|
||||
from werkzeug.utils import secure_filename
|
||||
from functools import wraps, lru_cache
|
||||
from functools import wraps
|
||||
from extensions import db, bcrypt, login_manager
|
||||
from sqlalchemy import text
|
||||
from dotenv import load_dotenv
|
||||
import logging
|
||||
import gc
|
||||
|
||||
# Load environment variables from .env file
|
||||
load_dotenv()
|
||||
|
||||
# Configure logging for better performance monitoring
|
||||
logging.basicConfig(level=logging.WARNING)
|
||||
|
||||
# First import models
|
||||
from models import User, Player, Group, Content, ServerLog, group_player
|
||||
|
||||
@@ -46,7 +42,7 @@ from utils.uploads import (
|
||||
add_image_to_playlist,
|
||||
convert_video_and_update_playlist,
|
||||
process_pdf,
|
||||
process_pptx_improved,
|
||||
process_pptx,
|
||||
process_uploaded_files
|
||||
)
|
||||
|
||||
@@ -54,7 +50,15 @@ from utils.uploads import (
|
||||
SERVER_VERSION = "1.1.0"
|
||||
BUILD_DATE = "2025-06-29"
|
||||
|
||||
app = Flask(__name__, instance_relative_config=True)
|
||||
# Get the absolute path of the app directory
|
||||
app_dir = os.path.dirname(os.path.abspath(__file__))
|
||||
template_dir = os.path.join(app_dir, 'templates')
|
||||
static_dir = os.path.join(app_dir, 'static')
|
||||
|
||||
app = Flask(__name__,
|
||||
instance_relative_config=True,
|
||||
template_folder=template_dir,
|
||||
static_folder=static_dir)
|
||||
|
||||
# Set the secret key from environment variable or use a default value
|
||||
app.config['SECRET_KEY'] = os.getenv('SECRET_KEY', 'Ana_Are_Multe_Mere-Si_Nu_Are_Pere')
|
||||
@@ -65,18 +69,8 @@ db_path = os.path.join(instance_dir, 'dashboard.db')
|
||||
app.config['SQLALCHEMY_DATABASE_URI'] = f'sqlite:///{db_path}'
|
||||
app.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False
|
||||
|
||||
# Performance configuration
|
||||
app.config['SQLALCHEMY_ENGINE_OPTIONS'] = {
|
||||
'pool_pre_ping': True,
|
||||
'pool_recycle': 300,
|
||||
'connect_args': {'timeout': 10}
|
||||
}
|
||||
|
||||
# Set maximum content length to 1GB (reduced from 2GB)
|
||||
app.config['MAX_CONTENT_LENGTH'] = 1024 * 1024 * 1024 # 1GB
|
||||
|
||||
# Add timeout configuration
|
||||
app.config['SEND_FILE_MAX_AGE_DEFAULT'] = 31536000 # Cache static files for 1 year
|
||||
# Set maximum content length to 1GB
|
||||
app.config['MAX_CONTENT_LENGTH'] = 2048 * 2048 * 2048 # 2GB, adjust as needed
|
||||
|
||||
# Ensure the instance folder exists
|
||||
os.makedirs(app.instance_path, exist_ok=True)
|
||||
@@ -113,6 +107,52 @@ def admin_required(f):
|
||||
return f(*args, **kwargs)
|
||||
return decorated_function
|
||||
|
||||
def get_system_info():
|
||||
"""Get system monitoring information"""
|
||||
try:
|
||||
# CPU information
|
||||
cpu_percent = psutil.cpu_percent(interval=1)
|
||||
cpu_count = psutil.cpu_count()
|
||||
|
||||
# Memory information
|
||||
memory = psutil.virtual_memory()
|
||||
memory_percent = memory.percent
|
||||
memory_used = round(memory.used / (1024**3), 2) # GB
|
||||
memory_total = round(memory.total / (1024**3), 2) # GB
|
||||
|
||||
# Disk information
|
||||
disk = psutil.disk_usage('/')
|
||||
disk_percent = round((disk.used / disk.total) * 100, 1)
|
||||
disk_used = round(disk.used / (1024**3), 2) # GB
|
||||
disk_total = round(disk.total / (1024**3), 2) # GB
|
||||
disk_free = round(disk.free / (1024**3), 2) # GB
|
||||
|
||||
# Upload folder size
|
||||
upload_folder_size = 0
|
||||
if os.path.exists(UPLOAD_FOLDER):
|
||||
for dirpath, dirnames, filenames in os.walk(UPLOAD_FOLDER):
|
||||
for filename in filenames:
|
||||
filepath = os.path.join(dirpath, filename)
|
||||
if os.path.exists(filepath):
|
||||
upload_folder_size += os.path.getsize(filepath)
|
||||
upload_folder_size_gb = round(upload_folder_size / (1024**3), 2)
|
||||
|
||||
return {
|
||||
'cpu_percent': cpu_percent,
|
||||
'cpu_count': cpu_count,
|
||||
'memory_percent': memory_percent,
|
||||
'memory_used': memory_used,
|
||||
'memory_total': memory_total,
|
||||
'disk_percent': disk_percent,
|
||||
'disk_used': disk_used,
|
||||
'disk_total': disk_total,
|
||||
'disk_free': disk_free,
|
||||
'upload_folder_size': upload_folder_size_gb
|
||||
}
|
||||
except Exception as e:
|
||||
print(f"Error getting system info: {e}")
|
||||
return None
|
||||
|
||||
@app.route('/')
|
||||
@login_required
|
||||
def dashboard():
|
||||
@@ -185,9 +225,12 @@ def upload_content():
|
||||
|
||||
players = [{'id': player.id, 'username': player.username} for player in Player.query.all()]
|
||||
groups = [{'id': group.id, 'name': group.name} for group in Group.query.all()]
|
||||
|
||||
# Get system information for monitoring
|
||||
system_info = get_system_info()
|
||||
|
||||
return render_template('upload_content.html', target_type=target_type, target_id=target_id,
|
||||
players=players, groups=groups, return_url=return_url)
|
||||
players=players, groups=groups, return_url=return_url, system_info=system_info)
|
||||
|
||||
@app.route('/admin')
|
||||
@login_required
|
||||
@@ -196,13 +239,18 @@ def admin():
|
||||
logo_exists = os.path.exists(os.path.join(app.config['UPLOAD_FOLDERLOGO'], 'logo.png'))
|
||||
login_picture_exists = os.path.exists(os.path.join(app.config['UPLOAD_FOLDERLOGO'], 'login_picture.png'))
|
||||
users = User.query.all()
|
||||
|
||||
# Get system information for monitoring
|
||||
system_info = get_system_info()
|
||||
|
||||
return render_template(
|
||||
'admin.html',
|
||||
users=users,
|
||||
logo_exists=logo_exists,
|
||||
login_picture_exists=login_picture_exists,
|
||||
server_version=SERVER_VERSION,
|
||||
build_date=BUILD_DATE
|
||||
build_date=BUILD_DATE,
|
||||
system_info=system_info
|
||||
)
|
||||
|
||||
@app.route('/admin/change_role/<int:user_id>', methods=['POST'])
|
||||
@@ -284,6 +332,61 @@ def delete_content(content_id):
|
||||
db.session.commit()
|
||||
return redirect(url_for('player_page', player_id=player_id))
|
||||
|
||||
@app.route('/player/<int:player_id>/bulk_delete', methods=['POST'])
|
||||
@login_required
|
||||
def bulk_delete_player_content(player_id):
|
||||
"""Bulk delete selected media files from player"""
|
||||
player = Player.query.get_or_404(player_id)
|
||||
|
||||
# Check if player is in a group (should be managed at group level)
|
||||
if player.groups:
|
||||
flash('Cannot delete media from players that are in groups. Manage media at the group level.', 'warning')
|
||||
return redirect(url_for('player_page', player_id=player_id))
|
||||
|
||||
selected_content_ids = request.form.getlist('selected_content')
|
||||
|
||||
if not selected_content_ids:
|
||||
flash('No media files selected for deletion.', 'warning')
|
||||
return redirect(url_for('player_page', player_id=player_id))
|
||||
|
||||
try:
|
||||
deleted_files = []
|
||||
deleted_count = 0
|
||||
|
||||
for content_id in selected_content_ids:
|
||||
content = Content.query.filter_by(id=content_id, player_id=player_id).first()
|
||||
if content:
|
||||
# Delete file from filesystem using absolute path
|
||||
upload_folder = app.config['UPLOAD_FOLDER']
|
||||
if not os.path.isabs(upload_folder):
|
||||
upload_folder = os.path.abspath(upload_folder)
|
||||
file_path = os.path.join(upload_folder, content.file_name)
|
||||
|
||||
if os.path.exists(file_path):
|
||||
try:
|
||||
os.remove(file_path)
|
||||
deleted_files.append(content.file_name)
|
||||
print(f"Deleted file: {file_path}")
|
||||
except OSError as e:
|
||||
print(f"Error deleting file {file_path}: {e}")
|
||||
|
||||
# Delete from database
|
||||
db.session.delete(content)
|
||||
deleted_count += 1
|
||||
|
||||
# Update playlist version for the player
|
||||
player.playlist_version += 1
|
||||
db.session.commit()
|
||||
|
||||
flash(f'Successfully deleted {deleted_count} media file(s). Playlist updated to version {player.playlist_version}.', 'success')
|
||||
|
||||
except Exception as e:
|
||||
db.session.rollback()
|
||||
print(f"Error in bulk delete: {e}")
|
||||
flash('An error occurred while deleting media files.', 'danger')
|
||||
|
||||
return redirect(url_for('player_page', player_id=player_id))
|
||||
|
||||
@app.route('/player/<int:player_id>/fullscreen', methods=['GET', 'POST'])
|
||||
def player_fullscreen(player_id):
|
||||
player = Player.query.get_or_404(player_id)
|
||||
@@ -332,7 +435,6 @@ def add_player():
|
||||
orientation = request.form.get('orientation', 'Landscape') # <-- Get orientation
|
||||
add_player_util(username, hostname, password, quickconnect_password, orientation) # <-- Pass orientation
|
||||
flash(f'Player "{username}" added successfully.', 'success')
|
||||
clear_player_cache() # Clear cache when player is added
|
||||
return redirect(url_for('dashboard'))
|
||||
return render_template('add_player.html')
|
||||
|
||||
@@ -349,7 +451,6 @@ def edit_player(player_id):
|
||||
orientation = request.form.get('orientation', player.orientation) # <-- Get orientation
|
||||
edit_player_util(player_id, username, hostname, password, quickconnect_password, orientation) # <-- Pass orientation
|
||||
flash(f'Player "{username}" updated successfully.', 'success')
|
||||
clear_player_cache() # Clear cache when player is updated
|
||||
return redirect(url_for('player_page', player_id=player.id))
|
||||
|
||||
return_url = request.args.get('return_url', url_for('player_page', player_id=player.id))
|
||||
@@ -364,103 +465,6 @@ def change_theme():
|
||||
db.session.commit()
|
||||
return redirect(url_for('admin'))
|
||||
|
||||
# Group management routes
|
||||
@app.route('/group/create', methods=['GET', 'POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def create_group():
|
||||
if request.method == 'POST':
|
||||
name = request.form['name']
|
||||
player_ids = request.form.getlist('players')
|
||||
orientation = request.form.get('orientation', 'Landscape')
|
||||
|
||||
try:
|
||||
# Convert player_ids to integers
|
||||
player_ids = [int(pid) for pid in player_ids]
|
||||
group = create_group_util(name, player_ids, orientation)
|
||||
flash(f'Group "{name}" created successfully.', 'success')
|
||||
return redirect(url_for('dashboard'))
|
||||
except ValueError as e:
|
||||
flash(str(e), 'danger')
|
||||
return redirect(url_for('create_group'))
|
||||
|
||||
# GET request - show create group form
|
||||
players = Player.query.filter_by(locked_to_group_id=None).all() # Only available players
|
||||
return render_template('create_group.html', players=players)
|
||||
|
||||
@app.route('/group/<int:group_id>/edit', methods=['GET', 'POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def edit_group(group_id):
|
||||
group = Group.query.get_or_404(group_id)
|
||||
|
||||
if request.method == 'POST':
|
||||
name = request.form['name']
|
||||
player_ids = request.form.getlist('players')
|
||||
orientation = request.form.get('orientation', group.orientation)
|
||||
|
||||
try:
|
||||
# Convert player_ids to integers
|
||||
player_ids = [int(pid) for pid in player_ids]
|
||||
edit_group_util(group_id, name, player_ids, orientation)
|
||||
flash(f'Group "{name}" updated successfully.', 'success')
|
||||
return redirect(url_for('dashboard'))
|
||||
except ValueError as e:
|
||||
flash(str(e), 'danger')
|
||||
return redirect(url_for('edit_group', group_id=group_id))
|
||||
|
||||
# GET request - show edit group form
|
||||
players = Player.query.all()
|
||||
return render_template('edit_group.html', group=group, players=players)
|
||||
|
||||
@app.route('/group/<int:group_id>/delete', methods=['POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def delete_group(group_id):
|
||||
delete_group_util(group_id)
|
||||
flash('Group deleted successfully.', 'success')
|
||||
return redirect(url_for('dashboard'))
|
||||
|
||||
@app.route('/group/<int:group_id>')
|
||||
@login_required
|
||||
def manage_group(group_id):
|
||||
group = Group.query.get_or_404(group_id)
|
||||
content = get_group_content(group_id)
|
||||
return render_template('manage_group.html', group=group, content=content)
|
||||
|
||||
@app.route('/group/<int:group_id>/fullscreen', methods=['GET', 'POST'])
|
||||
def group_fullscreen(group_id):
|
||||
group = Group.query.get_or_404(group_id)
|
||||
content = get_group_content(group_id)
|
||||
return render_template('group_fullscreen.html', group=group, content=content)
|
||||
|
||||
@app.route('/group/<int:group_id>/media/<int:content_id>/edit', methods=['POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def edit_group_media(group_id, content_id):
|
||||
new_duration = int(request.form['duration'])
|
||||
success = edit_group_media(group_id, content_id, new_duration)
|
||||
|
||||
if success:
|
||||
flash('Media duration updated successfully.', 'success')
|
||||
else:
|
||||
flash('Error updating media duration.', 'danger')
|
||||
|
||||
return redirect(url_for('manage_group', group_id=group_id))
|
||||
|
||||
@app.route('/group/<int:group_id>/media/<int:content_id>/delete', methods=['POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def delete_group_media(group_id, content_id):
|
||||
success = delete_group_media(group_id, content_id)
|
||||
|
||||
if success:
|
||||
flash('Media deleted successfully.', 'success')
|
||||
else:
|
||||
flash('Error deleting media.', 'danger')
|
||||
|
||||
return redirect(url_for('manage_group', group_id=group_id))
|
||||
|
||||
@app.route('/upload_logo', methods=['POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
@@ -521,113 +525,259 @@ def clean_unused_files():
|
||||
print("Used files:", used_files)
|
||||
print("Unused files:", unused_files)
|
||||
|
||||
# Delete unused files
|
||||
# Delete unused files using absolute path
|
||||
upload_folder = app.config['UPLOAD_FOLDER']
|
||||
if not os.path.isabs(upload_folder):
|
||||
upload_folder = os.path.abspath(upload_folder)
|
||||
|
||||
for file_name in unused_files:
|
||||
file_path = os.path.join(app.config['UPLOAD_FOLDER'], file_name)
|
||||
file_path = os.path.join(upload_folder, file_name)
|
||||
if os.path.isfile(file_path):
|
||||
print(f"Deleting file: {file_path}") # Debugging: Print the file being deleted
|
||||
print(f"Deleting unused file: {file_path}")
|
||||
os.remove(file_path)
|
||||
|
||||
flash('Unused files have been cleaned.', 'success')
|
||||
return redirect(url_for('admin'))
|
||||
|
||||
# Cache for frequently accessed data
|
||||
@lru_cache(maxsize=128)
|
||||
def get_player_by_hostname(hostname):
|
||||
"""Cached function to get player by hostname"""
|
||||
return Player.query.filter_by(hostname=hostname).first()
|
||||
|
||||
# Clear cache when players are modified
|
||||
def clear_player_cache():
|
||||
get_player_by_hostname.cache_clear()
|
||||
|
||||
# Optimized API endpoint with caching
|
||||
@app.route('/api/playlists', methods=['GET'])
|
||||
def get_playlists():
|
||||
hostname = request.args.get('hostname')
|
||||
quickconnect_code = request.args.get('quickconnect_code')
|
||||
|
||||
# Validate parameters early
|
||||
# Validate the parameters
|
||||
if not hostname or not quickconnect_code:
|
||||
return jsonify({'error': 'Hostname and quick connect code are required'}), 400
|
||||
|
||||
try:
|
||||
# Use cached function for better performance
|
||||
player = get_player_by_hostname(hostname)
|
||||
if not player:
|
||||
return jsonify({'error': 'Player not found'}), 404
|
||||
|
||||
# Verify quickconnect code
|
||||
if not bcrypt.check_password_hash(player.quickconnect_password, quickconnect_code):
|
||||
return jsonify({'error': 'Invalid credentials'}), 401
|
||||
# Find the player by hostname and verify the quickconnect code
|
||||
player = Player.query.filter_by(hostname=hostname).first()
|
||||
if not player or not bcrypt.check_password_hash(player.quickconnect_password, quickconnect_code):
|
||||
return jsonify({'error': 'Invalid hostname or quick connect code'}), 404
|
||||
|
||||
# Optimized content query
|
||||
if player.locked_to_group_id:
|
||||
# More efficient group content query
|
||||
content = db.session.query(Content).join(Player).filter(
|
||||
Player.locked_to_group_id == player.locked_to_group_id
|
||||
).distinct(Content.file_name).order_by(Content.position).all()
|
||||
else:
|
||||
# Get player's individual content with limit
|
||||
content = Content.query.filter_by(player_id=player.id).order_by(Content.position).all()
|
||||
# Check if player is locked to a group
|
||||
if player.locked_to_group_id:
|
||||
# Get content for all players in the group to ensure shared content
|
||||
group_players = player.locked_to_group.players
|
||||
player_ids = [p.id for p in group_players]
|
||||
|
||||
# Build playlist efficiently
|
||||
playlist = []
|
||||
for media in content:
|
||||
playlist.append({
|
||||
'file_name': media.file_name,
|
||||
'url': f"http://{request.host}/media/{media.file_name}",
|
||||
'duration': media.duration
|
||||
})
|
||||
|
||||
# Force garbage collection for memory management
|
||||
gc.collect()
|
||||
|
||||
return jsonify({
|
||||
'playlist': playlist,
|
||||
'playlist_version': player.playlist_version,
|
||||
'hashed_quickconnect': player.quickconnect_password
|
||||
})
|
||||
# Use the first occurrence of each file for the playlist
|
||||
content_query = (
|
||||
db.session.query(
|
||||
Content.file_name,
|
||||
db.func.min(Content.id).label('id'),
|
||||
db.func.min(Content.duration).label('duration')
|
||||
)
|
||||
.filter(Content.player_id.in_(player_ids))
|
||||
.group_by(Content.file_name)
|
||||
)
|
||||
|
||||
except Exception as e:
|
||||
app.logger.error(f"API Error: {str(e)}")
|
||||
return jsonify({'error': 'Internal server error'}), 500
|
||||
content = db.session.query(Content).filter(
|
||||
Content.id.in_([c.id for c in content_query])
|
||||
).all()
|
||||
else:
|
||||
# Get player's individual content
|
||||
content = Content.query.filter_by(player_id=player.id).all()
|
||||
|
||||
playlist = [
|
||||
{
|
||||
'file_name': media.file_name,
|
||||
'url': f"http://{request.host}/media/{media.file_name}",
|
||||
'duration': media.duration
|
||||
}
|
||||
for media in content
|
||||
]
|
||||
|
||||
# Return the playlist, version, and hashed quickconnect code
|
||||
return jsonify({
|
||||
'playlist': playlist,
|
||||
'playlist_version': player.playlist_version,
|
||||
'hashed_quickconnect': player.quickconnect_password
|
||||
})
|
||||
|
||||
# Optimized media serving with proper caching
|
||||
@app.route('/media/<path:filename>')
|
||||
def media(filename):
|
||||
try:
|
||||
response = send_from_directory(app.config['UPLOAD_FOLDER'], filename)
|
||||
# Add caching headers for better performance
|
||||
response.cache_control.max_age = 86400 # Cache for 24 hours
|
||||
response.cache_control.public = True
|
||||
return response
|
||||
except Exception as e:
|
||||
app.logger.error(f"Media serving error: {str(e)}")
|
||||
return jsonify({'error': 'File not found'}), 404
|
||||
return send_from_directory(app.config['UPLOAD_FOLDER'], filename)
|
||||
|
||||
@app.context_processor
|
||||
def inject_theme():
|
||||
if current_user.is_authenticated:
|
||||
theme = current_user.theme
|
||||
else:
|
||||
theme = 'light'
|
||||
return dict(theme=theme)
|
||||
|
||||
@app.route('/group/create', methods=['GET', 'POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def create_group():
|
||||
if request.method == 'POST':
|
||||
group_name = request.form['name']
|
||||
player_ids = request.form.getlist('players')
|
||||
orientation = request.form.get('orientation', 'Landscape')
|
||||
create_group_util(group_name, player_ids, orientation)
|
||||
flash(f'Group "{group_name}" created successfully.', 'success')
|
||||
return redirect(url_for('dashboard'))
|
||||
players = Player.query.all()
|
||||
return render_template('create_group.html', players=players)
|
||||
|
||||
@app.route('/group/<int:group_id>/manage')
|
||||
@login_required
|
||||
@admin_required
|
||||
def manage_group(group_id):
|
||||
group = Group.query.get_or_404(group_id)
|
||||
content = get_group_content(group_id)
|
||||
# Debug content ordering
|
||||
print("Group content positions before sorting:", [(c.id, c.file_name, c.position) for c in content])
|
||||
content = sorted(content, key=lambda c: c.position)
|
||||
print("Group content positions after sorting:", [(c.id, c.file_name, c.position) for c in content])
|
||||
return render_template('manage_group.html', group=group, content=content)
|
||||
|
||||
@app.route('/group/<int:group_id>/edit', methods=['GET', 'POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def edit_group(group_id):
|
||||
group = Group.query.get_or_404(group_id)
|
||||
if request.method == 'POST':
|
||||
name = request.form['name']
|
||||
player_ids = request.form.getlist('players')
|
||||
orientation = request.form.get('orientation', group.orientation)
|
||||
edit_group_util(group_id, name, player_ids, orientation)
|
||||
flash(f'Group "{name}" updated successfully.', 'success')
|
||||
return redirect(url_for('dashboard'))
|
||||
players = Player.query.all()
|
||||
return render_template('edit_group.html', group=group, players=players)
|
||||
|
||||
@app.route('/group/<int:group_id>/delete', methods=['POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def delete_group(group_id):
|
||||
group = Group.query.get_or_404(group_id)
|
||||
group_name = group.name
|
||||
delete_group_util(group_id)
|
||||
flash(f'Group "{group_name}" deleted successfully.', 'success')
|
||||
return redirect(url_for('dashboard'))
|
||||
|
||||
@app.route('/group/<int:group_id>/fullscreen', methods=['GET'])
|
||||
@login_required
|
||||
def group_fullscreen(group_id):
|
||||
group = Group.query.get_or_404(group_id)
|
||||
content = Content.query.filter(Content.player_id.in_([player.id for player in group.players])).order_by(Content.position).all()
|
||||
return render_template('group_fullscreen.html', group=group, content=content)
|
||||
|
||||
@app.route('/group/<int:group_id>/media/<int:content_id>/edit', methods=['POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def edit_group_media_route(group_id, content_id):
|
||||
new_duration = int(request.form['duration'])
|
||||
success = edit_group_media(group_id, content_id, new_duration)
|
||||
|
||||
if success:
|
||||
flash('Media duration updated successfully.', 'success')
|
||||
else:
|
||||
flash('Error updating media duration.', 'danger')
|
||||
|
||||
return redirect(url_for('manage_group', group_id=group_id))
|
||||
|
||||
@app.route('/group/<int:group_id>/media/<int:content_id>/delete', methods=['POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def delete_group_media_route(group_id, content_id):
|
||||
success = delete_group_media(group_id, content_id)
|
||||
|
||||
if success:
|
||||
flash('Media deleted successfully.', 'success')
|
||||
else:
|
||||
flash('Error deleting media.', 'danger')
|
||||
|
||||
return redirect(url_for('manage_group', group_id=group_id))
|
||||
|
||||
@app.route('/group/<int:group_id>/bulk_delete', methods=['POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def bulk_delete_group_content(group_id):
|
||||
"""Bulk delete selected media files from group"""
|
||||
group = Group.query.get_or_404(group_id)
|
||||
selected_content_ids = request.form.getlist('selected_content')
|
||||
|
||||
if not selected_content_ids:
|
||||
flash('No media files selected for deletion.', 'warning')
|
||||
return redirect(url_for('manage_group', group_id=group_id))
|
||||
|
||||
try:
|
||||
deleted_files = []
|
||||
deleted_count = 0
|
||||
player_ids = [player.id for player in group.players]
|
||||
|
||||
for content_id in selected_content_ids:
|
||||
content = Content.query.filter(
|
||||
Content.id == content_id,
|
||||
Content.player_id.in_(player_ids)
|
||||
).first()
|
||||
|
||||
if content:
|
||||
# Delete file from filesystem using absolute path
|
||||
upload_folder = app.config['UPLOAD_FOLDER']
|
||||
if not os.path.isabs(upload_folder):
|
||||
upload_folder = os.path.abspath(upload_folder)
|
||||
file_path = os.path.join(upload_folder, content.file_name)
|
||||
|
||||
if os.path.exists(file_path):
|
||||
try:
|
||||
os.remove(file_path)
|
||||
deleted_files.append(content.file_name)
|
||||
print(f"Deleted file: {file_path}")
|
||||
except OSError as e:
|
||||
print(f"Error deleting file {file_path}: {e}")
|
||||
|
||||
# Delete from database
|
||||
db.session.delete(content)
|
||||
deleted_count += 1
|
||||
|
||||
# Update playlist version for all players in the group
|
||||
for player in group.players:
|
||||
player.playlist_version += 1
|
||||
|
||||
db.session.commit()
|
||||
|
||||
flash(f'Successfully deleted {deleted_count} media file(s) from group. All player playlists updated.', 'success')
|
||||
|
||||
except Exception as e:
|
||||
db.session.rollback()
|
||||
print(f"Error in group bulk delete: {e}")
|
||||
flash('An error occurred while deleting media files.', 'danger')
|
||||
|
||||
return redirect(url_for('manage_group', group_id=group_id))
|
||||
|
||||
# Optimized playlist version check
|
||||
@app.route('/api/playlist_version', methods=['GET'])
|
||||
def get_playlist_version():
|
||||
hostname = request.args.get('hostname')
|
||||
quickconnect_code = request.args.get('quickconnect_code')
|
||||
|
||||
# Validate the parameters
|
||||
if not hostname or not quickconnect_code:
|
||||
return jsonify({'error': 'Hostname and quick connect code are required'}), 400
|
||||
|
||||
try:
|
||||
# Use cached function
|
||||
player = get_player_by_hostname(hostname)
|
||||
if not player or not bcrypt.check_password_hash(player.quickconnect_password, quickconnect_code):
|
||||
return jsonify({'error': 'Invalid credentials'}), 401
|
||||
# Find the player by hostname and verify the quickconnect code
|
||||
player = Player.query.filter_by(hostname=hostname).first()
|
||||
if not player or not bcrypt.check_password_hash(player.quickconnect_password, quickconnect_code):
|
||||
return jsonify({'error': 'Invalid hostname or quick connect code'}), 404
|
||||
|
||||
return jsonify({
|
||||
'playlist_version': player.playlist_version,
|
||||
'hashed_quickconnect': player.quickconnect_password
|
||||
})
|
||||
except Exception as e:
|
||||
app.logger.error(f"Version check error: {str(e)}")
|
||||
return jsonify({'error': 'Internal server error'}), 500
|
||||
# Return the playlist version and hashed quickconnect code
|
||||
return jsonify({
|
||||
'playlist_version': player.playlist_version,
|
||||
'hashed_quickconnect': player.quickconnect_password
|
||||
})
|
||||
|
||||
@app.route('/api/system_info', methods=['GET'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def api_system_info():
|
||||
"""API endpoint to get real-time system information"""
|
||||
system_info = get_system_info()
|
||||
if system_info:
|
||||
return jsonify(system_info)
|
||||
else:
|
||||
return jsonify({'error': 'Could not retrieve system information'}), 500
|
||||
|
||||
@app.route('/player/<int:player_id>/update_order', methods=['POST'])
|
||||
@login_required
|
||||
@@ -703,218 +853,6 @@ if not app.debug or os.environ.get('WERKZEUG_RUN_MAIN') == 'true':
|
||||
db.create_all()
|
||||
create_default_user(db, User, bcrypt)
|
||||
|
||||
# Performance monitoring functions
|
||||
def get_system_stats():
|
||||
"""Get current system performance statistics"""
|
||||
try:
|
||||
cpu_percent = psutil.cpu_percent(interval=1)
|
||||
memory = psutil.virtual_memory()
|
||||
disk = psutil.disk_usage('/')
|
||||
|
||||
return {
|
||||
'cpu_percent': cpu_percent,
|
||||
'memory_percent': memory.percent,
|
||||
'memory_used_mb': memory.used / (1024 * 1024),
|
||||
'memory_total_mb': memory.total / (1024 * 1024),
|
||||
'disk_percent': disk.percent,
|
||||
'disk_used_gb': disk.used / (1024 * 1024 * 1024),
|
||||
'disk_total_gb': disk.total / (1024 * 1024 * 1024),
|
||||
'timestamp': time.time()
|
||||
}
|
||||
except Exception as e:
|
||||
print(f"Error getting system stats: {e}")
|
||||
return None
|
||||
|
||||
# Performance monitoring endpoint
|
||||
@app.route('/api/performance', methods=['GET'])
|
||||
@login_required
|
||||
def get_performance_stats():
|
||||
"""API endpoint to get real-time performance statistics"""
|
||||
stats = get_system_stats()
|
||||
if stats:
|
||||
return jsonify(stats)
|
||||
else:
|
||||
return jsonify({'error': 'Unable to get system stats'}), 500
|
||||
|
||||
# Enhanced upload endpoint with monitoring
|
||||
@app.route('/upload_content_monitored', methods=['POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def upload_content_monitored():
|
||||
"""Enhanced upload endpoint with performance monitoring"""
|
||||
start_time = time.time()
|
||||
start_stats = get_system_stats()
|
||||
|
||||
target_type = request.form.get('target_type')
|
||||
target_id = request.form.get('target_id')
|
||||
files = request.files.getlist('files')
|
||||
duration = int(request.form['duration'])
|
||||
return_url = request.form.get('return_url')
|
||||
media_type = request.form['media_type']
|
||||
|
||||
print(f"=== UPLOAD MONITORING START ===")
|
||||
print(f"Target Type: {target_type}, Target ID: {target_id}, Media Type: {media_type}")
|
||||
print(f"Number of files: {len(files)}")
|
||||
print(f"Start CPU: {start_stats['cpu_percent']}%, Memory: {start_stats['memory_percent']}%")
|
||||
|
||||
if not target_type or not target_id:
|
||||
flash('Please select a target type and target ID.', 'danger')
|
||||
return redirect(url_for('upload_content'))
|
||||
|
||||
# Monitor during file processing
|
||||
def monitor_upload():
|
||||
"""Background monitoring thread"""
|
||||
while True:
|
||||
stats = get_system_stats()
|
||||
if stats:
|
||||
print(f"[MONITOR] CPU: {stats['cpu_percent']}%, Memory: {stats['memory_percent']}%, Time: {time.time() - start_time:.1f}s")
|
||||
time.sleep(2)
|
||||
|
||||
# Start monitoring thread
|
||||
monitor_thread = threading.Thread(target=monitor_upload, daemon=True)
|
||||
monitor_thread.start()
|
||||
|
||||
# Process uploaded files and get results
|
||||
results = process_uploaded_files(app, files, media_type, duration, target_type, target_id)
|
||||
|
||||
end_time = time.time()
|
||||
end_stats = get_system_stats()
|
||||
total_time = end_time - start_time
|
||||
|
||||
print(f"=== UPLOAD MONITORING END ===")
|
||||
print(f"Total processing time: {total_time:.2f} seconds")
|
||||
print(f"End CPU: {end_stats['cpu_percent']}%, Memory: {end_stats['memory_percent']}%")
|
||||
print(f"CPU change: {end_stats['cpu_percent'] - start_stats['cpu_percent']:.1f}%")
|
||||
print(f"Memory change: {end_stats['memory_percent'] - start_stats['memory_percent']:.1f}%")
|
||||
|
||||
# Log performance metrics
|
||||
log_action(f"Upload completed: {len(files)} files, {total_time:.2f}s, CPU: {start_stats['cpu_percent']}% → {end_stats['cpu_percent']}%")
|
||||
|
||||
return redirect(return_url)
|
||||
@app.route('/player/<int:player_id>/bulk_delete_content', methods=['POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def bulk_delete_player_content(player_id):
|
||||
"""Bulk delete content for a specific player"""
|
||||
if not request.is_json:
|
||||
return jsonify({'success': False, 'error': 'Invalid request format'}), 400
|
||||
|
||||
player = Player.query.get_or_404(player_id)
|
||||
content_ids = request.json.get('content_ids', [])
|
||||
|
||||
if not content_ids:
|
||||
return jsonify({'success': False, 'error': 'No content IDs provided'}), 400
|
||||
|
||||
try:
|
||||
# Get all content items to delete
|
||||
content_items = Content.query.filter(
|
||||
Content.id.in_(content_ids),
|
||||
Content.player_id == player_id
|
||||
).all()
|
||||
|
||||
if not content_items:
|
||||
return jsonify({'success': False, 'error': 'No valid content found to delete'}), 404
|
||||
|
||||
# Delete the content items
|
||||
deleted_count = 0
|
||||
for content in content_items:
|
||||
# Delete the actual file from filesystem
|
||||
file_path = os.path.join(app.config['UPLOAD_FOLDER'], content.file_name)
|
||||
if os.path.exists(file_path):
|
||||
try:
|
||||
os.remove(file_path)
|
||||
except OSError as e:
|
||||
app.logger.warning(f"Could not delete file {file_path}: {e}")
|
||||
|
||||
# Delete from database
|
||||
db.session.delete(content)
|
||||
deleted_count += 1
|
||||
|
||||
# Update playlist version for the player
|
||||
player.playlist_version += 1
|
||||
db.session.commit()
|
||||
|
||||
# Clear cache
|
||||
clear_player_cache()
|
||||
|
||||
# Log the action
|
||||
log_action(f"Bulk deleted {deleted_count} content items from player {player.username}")
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'deleted_count': deleted_count,
|
||||
'new_playlist_version': player.playlist_version
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
db.session.rollback()
|
||||
app.logger.error(f"Error in bulk delete: {str(e)}")
|
||||
return jsonify({'success': False, 'error': 'Database error occurred'}), 500
|
||||
|
||||
@app.route('/group/<int:group_id>/bulk_delete_content', methods=['POST'])
|
||||
@login_required
|
||||
@admin_required
|
||||
def bulk_delete_group_content(group_id):
|
||||
"""Bulk delete content for a specific group"""
|
||||
if not request.is_json:
|
||||
return jsonify({'success': False, 'error': 'Invalid request format'}), 400
|
||||
|
||||
group = Group.query.get_or_404(group_id)
|
||||
content_ids = request.json.get('content_ids', [])
|
||||
|
||||
if not content_ids:
|
||||
return jsonify({'success': False, 'error': 'No content IDs provided'}), 400
|
||||
|
||||
try:
|
||||
# Get player IDs in the group
|
||||
player_ids = [p.id for p in group.players]
|
||||
|
||||
# Get all content items to delete that belong to players in this group
|
||||
content_items = Content.query.filter(
|
||||
Content.id.in_(content_ids),
|
||||
Content.player_id.in_(player_ids)
|
||||
).all()
|
||||
|
||||
if not content_items:
|
||||
return jsonify({'success': False, 'error': 'No valid content found to delete'}), 404
|
||||
|
||||
# Delete the content items
|
||||
deleted_count = 0
|
||||
for content in content_items:
|
||||
# Delete the actual file from filesystem
|
||||
file_path = os.path.join(app.config['UPLOAD_FOLDER'], content.file_name)
|
||||
if os.path.exists(file_path):
|
||||
try:
|
||||
os.remove(file_path)
|
||||
except OSError as e:
|
||||
app.logger.warning(f"Could not delete file {file_path}: {e}")
|
||||
|
||||
# Delete from database
|
||||
db.session.delete(content)
|
||||
deleted_count += 1
|
||||
|
||||
# Update playlist version for all players in the group
|
||||
for player in group.players:
|
||||
player.playlist_version += 1
|
||||
|
||||
db.session.commit()
|
||||
|
||||
# Clear cache
|
||||
clear_player_cache()
|
||||
|
||||
# Log the action
|
||||
log_action(f"Bulk deleted {deleted_count} content items from group {group.name}")
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'deleted_count': deleted_count
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
db.session.rollback()
|
||||
app.logger.error(f"Error in group bulk delete: {str(e)}")
|
||||
return jsonify({'success': False, 'error': 'Database error occurred'}), 500
|
||||
|
||||
# Add this at the end of app.py
|
||||
if __name__ == '__main__':
|
||||
app.run(debug=True, host='0.0.0.0', port=5000)
|
||||
@@ -21,7 +21,6 @@ greenlet==3.1.1
|
||||
# File Processing
|
||||
pdf2image==1.17.0
|
||||
PyPDF2==3.0.1
|
||||
python-pptx==0.6.21
|
||||
Pillow==10.0.1
|
||||
cairosvg==2.7.0
|
||||
ffmpeg-python==0.2.0
|
||||
@@ -39,6 +38,7 @@ gevent==23.9.1
|
||||
# Monitoring & Performance
|
||||
prometheus-flask-exporter==0.22.4
|
||||
sentry-sdk[flask]==1.40.0
|
||||
psutil==6.1.0
|
||||
|
||||
# Utilities
|
||||
typing_extensions==4.12.2
|
||||
|
Before Width: | Height: | Size: 153 KiB After Width: | Height: | Size: 153 KiB |
|
Before Width: | Height: | Size: 52 KiB After Width: | Height: | Size: 52 KiB |
@@ -207,59 +207,94 @@
|
||||
<p><strong>Date of Build:</strong> {{ build_date }}</p>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Performance Monitoring Dashboard -->
|
||||
<div class="card mb-4 {{ 'dark-mode' if theme == 'dark' else '' }}">
|
||||
<div class="card-header bg-success text-white">
|
||||
<h2>Performance Monitor</h2>
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<div class="row">
|
||||
<div class="col-md-4">
|
||||
<div class="text-center">
|
||||
<h5>CPU Usage</h5>
|
||||
<div id="cpu-gauge" class="progress mb-2">
|
||||
<div class="progress-bar bg-info" role="progressbar" style="width: 0%"></div>
|
||||
</div>
|
||||
<small id="cpu-text">0%</small>
|
||||
</div>
|
||||
<!-- System Monitoring Card -->
|
||||
{% if system_info %}
|
||||
<div class="card mb-4 {{ 'dark-mode' if theme == 'dark' else '' }}">
|
||||
<div class="card-header">
|
||||
<h2>📊 System Monitoring</h2>
|
||||
</div>
|
||||
<div class="col-md-4">
|
||||
<div class="text-center">
|
||||
<h5>Memory Usage</h5>
|
||||
<div id="memory-gauge" class="progress mb-2">
|
||||
<div class="progress-bar bg-warning" role="progressbar" style="width: 0%"></div>
|
||||
<div class="card-body">
|
||||
<div class="row">
|
||||
<!-- CPU Information -->
|
||||
<div class="col-md-3 col-6 text-center mb-3">
|
||||
<div class="h6">CPU Usage</div>
|
||||
<div class="progress mb-2" style="height: 25px;">
|
||||
<div class="progress-bar
|
||||
{% if system_info.cpu_percent < 50 %}bg-success
|
||||
{% elif system_info.cpu_percent < 80 %}bg-warning
|
||||
{% else %}bg-danger{% endif %}"
|
||||
role="progressbar"
|
||||
style="width: {{ system_info.cpu_percent }}%;">
|
||||
{{ system_info.cpu_percent }}%
|
||||
</div>
|
||||
</div>
|
||||
<small class="text-muted">{{ system_info.cpu_count }} cores available</small>
|
||||
</div>
|
||||
|
||||
<!-- Memory Information -->
|
||||
<div class="col-md-3 col-6 text-center mb-3">
|
||||
<div class="h6">Memory Usage</div>
|
||||
<div class="progress mb-2" style="height: 25px;">
|
||||
<div class="progress-bar
|
||||
{% if system_info.memory_percent < 60 %}bg-success
|
||||
{% elif system_info.memory_percent < 85 %}bg-warning
|
||||
{% else %}bg-danger{% endif %}"
|
||||
role="progressbar"
|
||||
style="width: {{ system_info.memory_percent }}%;">
|
||||
{{ system_info.memory_percent }}%
|
||||
</div>
|
||||
</div>
|
||||
<small class="text-muted">{{ system_info.memory_used }}GB / {{ system_info.memory_total }}GB</small>
|
||||
</div>
|
||||
|
||||
<!-- Disk Information -->
|
||||
<div class="col-md-3 col-6 text-center mb-3">
|
||||
<div class="h6">Disk Usage</div>
|
||||
<div class="progress mb-2" style="height: 25px;">
|
||||
<div class="progress-bar
|
||||
{% if system_info.disk_percent < 70 %}bg-success
|
||||
{% elif system_info.disk_percent < 90 %}bg-warning
|
||||
{% else %}bg-danger{% endif %}"
|
||||
role="progressbar"
|
||||
style="width: {{ system_info.disk_percent }}%;">
|
||||
{{ system_info.disk_percent }}%
|
||||
</div>
|
||||
</div>
|
||||
<small class="text-muted">{{ system_info.disk_used }}GB / {{ system_info.disk_total }}GB</small>
|
||||
</div>
|
||||
|
||||
<!-- Upload Folder Size -->
|
||||
<div class="col-md-3 col-6 text-center mb-3">
|
||||
<div class="h6">Media Storage</div>
|
||||
<div class="text-primary display-6">{{ system_info.upload_folder_size }}GB</div>
|
||||
<small class="text-muted">Total media files</small>
|
||||
</div>
|
||||
<small id="memory-text">0%</small>
|
||||
</div>
|
||||
</div>
|
||||
<div class="col-md-4">
|
||||
<div class="text-center">
|
||||
<h5>Disk Usage</h5>
|
||||
<div id="disk-gauge" class="progress mb-2">
|
||||
<div class="progress-bar bg-danger" role="progressbar" style="width: 0%"></div>
|
||||
|
||||
<!-- System Details -->
|
||||
<div class="row mt-3">
|
||||
<div class="col-12">
|
||||
<hr>
|
||||
<div class="row text-center">
|
||||
<div class="col-md-4 col-12 mb-2">
|
||||
<strong>Available Disk Space:</strong><br>
|
||||
<span class="text-success">{{ system_info.disk_free }}GB free</span>
|
||||
</div>
|
||||
<div class="col-md-4 col-12 mb-2">
|
||||
<strong>Total Disk Space:</strong><br>
|
||||
<span class="text-info">{{ system_info.disk_total }}GB total</span>
|
||||
</div>
|
||||
<div class="col-md-4 col-12 mb-2">
|
||||
<strong>Last Updated:</strong><br>
|
||||
<span class="text-muted" id="last-update-admin">Just now</span>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
<small id="disk-text">0%</small>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
<div class="row mt-3">
|
||||
<div class="col-12">
|
||||
<button id="toggle-monitor" class="btn btn-primary">Start Monitoring</button>
|
||||
<button id="reset-stats" class="btn btn-secondary">Reset Stats</button>
|
||||
<span id="monitor-status" class="ms-3 text-muted">Monitoring stopped</span>
|
||||
</div>
|
||||
</div>
|
||||
<div class="row mt-3">
|
||||
<div class="col-12">
|
||||
<h6>Performance Log:</h6>
|
||||
<div id="perf-log" style="height: 100px; overflow-y: scroll; background-color: #f8f9fa; padding: 10px; border-radius: 5px; font-family: monospace; font-size: 12px;">
|
||||
<div class="text-muted">Performance monitoring ready...</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
{% endif %}
|
||||
</div>
|
||||
</div>
|
||||
|
||||
@@ -281,115 +316,76 @@
|
||||
}, 5000);
|
||||
}
|
||||
|
||||
// Performance monitoring functionality
|
||||
let monitoringInterval = null;
|
||||
let isMonitoring = false;
|
||||
let maxCpu = 0, maxMemory = 0;
|
||||
|
||||
function updateGauge(elementId, textId, value, color) {
|
||||
const gauge = document.querySelector(`#${elementId} .progress-bar`);
|
||||
const text = document.getElementById(textId);
|
||||
gauge.style.width = `${value}%`;
|
||||
gauge.className = `progress-bar ${color}`;
|
||||
text.textContent = `${value.toFixed(1)}%`;
|
||||
}
|
||||
|
||||
function logPerformance(message) {
|
||||
const log = document.getElementById('perf-log');
|
||||
const timestamp = new Date().toLocaleTimeString();
|
||||
const logEntry = document.createElement('div');
|
||||
logEntry.innerHTML = `<span class="text-muted">[${timestamp}]</span> ${message}`;
|
||||
log.appendChild(logEntry);
|
||||
log.scrollTop = log.scrollHeight;
|
||||
|
||||
// Keep only last 50 entries
|
||||
if (log.children.length > 50) {
|
||||
log.removeChild(log.firstChild);
|
||||
}
|
||||
}
|
||||
|
||||
function updatePerformanceStats() {
|
||||
fetch('/api/performance')
|
||||
// Auto-refresh system monitoring every 15 seconds
|
||||
{% if system_info %}
|
||||
function updateAdminSystemInfo() {
|
||||
fetch('/api/system_info')
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
if (data.error) {
|
||||
logPerformance(`Error: ${data.error}`);
|
||||
console.warn('Could not fetch system info:', data.error);
|
||||
return;
|
||||
}
|
||||
|
||||
// Update gauges
|
||||
updateGauge('cpu-gauge', 'cpu-text', data.cpu_percent, 'bg-info');
|
||||
updateGauge('memory-gauge', 'memory-text', data.memory_percent, 'bg-warning');
|
||||
updateGauge('disk-gauge', 'disk-text', data.disk_percent, 'bg-danger');
|
||||
// Update progress bars and their colors
|
||||
const progressBars = document.querySelectorAll('.progress-bar');
|
||||
|
||||
// Track maximum values
|
||||
if (data.cpu_percent > maxCpu) {
|
||||
maxCpu = data.cpu_percent;
|
||||
logPerformance(`New CPU peak: ${maxCpu.toFixed(1)}%`);
|
||||
}
|
||||
if (data.memory_percent > maxMemory) {
|
||||
maxMemory = data.memory_percent;
|
||||
logPerformance(`New Memory peak: ${maxMemory.toFixed(1)}%`);
|
||||
if (progressBars.length >= 3) {
|
||||
// CPU Bar
|
||||
progressBars[0].style.width = data.cpu_percent + '%';
|
||||
progressBars[0].textContent = data.cpu_percent + '%';
|
||||
progressBars[0].className = 'progress-bar ' +
|
||||
(data.cpu_percent < 50 ? 'bg-success' :
|
||||
data.cpu_percent < 80 ? 'bg-warning' : 'bg-danger');
|
||||
|
||||
// Memory Bar
|
||||
progressBars[1].style.width = data.memory_percent + '%';
|
||||
progressBars[1].textContent = data.memory_percent + '%';
|
||||
progressBars[1].className = 'progress-bar ' +
|
||||
(data.memory_percent < 60 ? 'bg-success' :
|
||||
data.memory_percent < 85 ? 'bg-warning' : 'bg-danger');
|
||||
|
||||
// Disk Bar
|
||||
progressBars[2].style.width = data.disk_percent + '%';
|
||||
progressBars[2].textContent = data.disk_percent + '%';
|
||||
progressBars[2].className = 'progress-bar ' +
|
||||
(data.disk_percent < 70 ? 'bg-success' :
|
||||
data.disk_percent < 90 ? 'bg-warning' : 'bg-danger');
|
||||
}
|
||||
|
||||
// Log significant changes
|
||||
if (data.cpu_percent > 80) {
|
||||
logPerformance(`<span class="text-danger">High CPU usage: ${data.cpu_percent.toFixed(1)}%</span>`);
|
||||
// Update text values
|
||||
const smallTexts = document.querySelectorAll('.text-muted');
|
||||
smallTexts.forEach((text, index) => {
|
||||
if (index === 1) text.textContent = data.memory_used + 'GB / ' + data.memory_total + 'GB';
|
||||
if (index === 2) text.textContent = data.disk_used + 'GB / ' + data.disk_total + 'GB';
|
||||
});
|
||||
|
||||
// Update storage size
|
||||
const storageDisplay = document.querySelector('.display-6');
|
||||
if (storageDisplay) {
|
||||
storageDisplay.textContent = data.upload_folder_size + 'GB';
|
||||
}
|
||||
if (data.memory_percent > 80) {
|
||||
logPerformance(`<span class="text-danger">High Memory usage: ${data.memory_percent.toFixed(1)}%</span>`);
|
||||
|
||||
// Update disk space info
|
||||
const diskFree = document.querySelector('.text-success');
|
||||
const diskTotal = document.querySelector('.text-info');
|
||||
if (diskFree) diskFree.textContent = data.disk_free + 'GB free';
|
||||
if (diskTotal) diskTotal.textContent = data.disk_total + 'GB total';
|
||||
|
||||
// Update timestamp
|
||||
const lastUpdate = document.getElementById('last-update-admin');
|
||||
if (lastUpdate) {
|
||||
lastUpdate.textContent = new Date().toLocaleTimeString();
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
logPerformance(`Fetch error: ${error.message}`);
|
||||
console.warn('Admin system monitoring update failed:', error);
|
||||
});
|
||||
}
|
||||
|
||||
function toggleMonitoring() {
|
||||
const toggleButton = document.getElementById('toggle-monitor');
|
||||
const statusSpan = document.getElementById('monitor-status');
|
||||
|
||||
if (isMonitoring) {
|
||||
// Stop monitoring
|
||||
clearInterval(monitoringInterval);
|
||||
isMonitoring = false;
|
||||
toggleButton.textContent = 'Start Monitoring';
|
||||
toggleButton.className = 'btn btn-primary';
|
||||
statusSpan.textContent = 'Monitoring stopped';
|
||||
statusSpan.className = 'ms-3 text-muted';
|
||||
logPerformance('Monitoring stopped');
|
||||
} else {
|
||||
// Start monitoring
|
||||
isMonitoring = true;
|
||||
toggleButton.textContent = 'Stop Monitoring';
|
||||
toggleButton.className = 'btn btn-danger';
|
||||
statusSpan.textContent = 'Monitoring active';
|
||||
statusSpan.className = 'ms-3 text-success';
|
||||
logPerformance('Monitoring started');
|
||||
|
||||
// Update immediately and then every 2 seconds
|
||||
updatePerformanceStats();
|
||||
monitoringInterval = setInterval(updatePerformanceStats, 2000);
|
||||
}
|
||||
}
|
||||
|
||||
function resetStats() {
|
||||
maxCpu = 0;
|
||||
maxMemory = 0;
|
||||
const log = document.getElementById('perf-log');
|
||||
log.innerHTML = '<div class="text-muted">Performance log reset...</div>';
|
||||
logPerformance('Stats reset');
|
||||
}
|
||||
|
||||
// Event listeners
|
||||
document.getElementById('toggle-monitor').addEventListener('click', toggleMonitoring);
|
||||
document.getElementById('reset-stats').addEventListener('click', resetStats);
|
||||
|
||||
// Auto-start monitoring when page loads
|
||||
document.addEventListener('DOMContentLoaded', function() {
|
||||
// Initial stats load
|
||||
updatePerformanceStats();
|
||||
});
|
||||
// Update every 15 seconds
|
||||
setInterval(updateAdminSystemInfo, 15000);
|
||||
{% endif %}
|
||||
</script>
|
||||
</body>
|
||||
</html>
|
||||
@@ -92,16 +92,20 @@
|
||||
<div class="card-body">
|
||||
{% if content %}
|
||||
<!-- Bulk Actions Controls -->
|
||||
<div class="mb-3 d-flex flex-wrap align-items-center gap-2">
|
||||
<div class="form-check">
|
||||
<input class="form-check-input" type="checkbox" id="selectAllGroup">
|
||||
<label class="form-check-label" for="selectAllGroup">
|
||||
Select All
|
||||
</label>
|
||||
<div class="row mb-3">
|
||||
<div class="col-md-6">
|
||||
<div class="form-check">
|
||||
<input class="form-check-input" type="checkbox" id="selectAll">
|
||||
<label class="form-check-label" for="selectAll">
|
||||
Select All
|
||||
</label>
|
||||
</div>
|
||||
</div>
|
||||
<div class="col-md-6 text-end">
|
||||
<button type="button" class="btn btn-danger" id="bulkDeleteBtn" style="display:none;" onclick="confirmBulkDelete()">
|
||||
<i class="bi bi-trash"></i> Delete Selected
|
||||
</button>
|
||||
</div>
|
||||
<button id="deleteSelectedGroup" class="btn btn-danger" style="display: none;">
|
||||
<i class="bi bi-trash"></i> Delete Selected (<span id="selectedCountGroup">0</span>)
|
||||
</button>
|
||||
</div>
|
||||
|
||||
<ul class="list-group sortable-list" id="groupMediaList">
|
||||
@@ -112,7 +116,9 @@
|
||||
data-position="{{ loop.index0 }}">
|
||||
<!-- Checkbox for bulk selection -->
|
||||
<div class="me-2">
|
||||
<input type="checkbox" class="form-check-input group-media-checkbox"
|
||||
<input class="form-check-input media-checkbox"
|
||||
type="checkbox"
|
||||
name="selected_content"
|
||||
value="{{ media.id }}">
|
||||
</div>
|
||||
|
||||
@@ -239,70 +245,74 @@ document.addEventListener('DOMContentLoaded', function() {
|
||||
});
|
||||
}
|
||||
|
||||
// Bulk delete functionality
|
||||
const selectAllGroup = document.getElementById('selectAllGroup');
|
||||
const deleteSelectedGroup = document.getElementById('deleteSelectedGroup');
|
||||
const selectedCountGroup = document.getElementById('selectedCountGroup');
|
||||
const groupMediaCheckboxes = document.querySelectorAll('.group-media-checkbox');
|
||||
// Bulk selection functionality
|
||||
const selectAllCheckbox = document.getElementById('selectAll');
|
||||
const mediaCheckboxes = document.querySelectorAll('.media-checkbox');
|
||||
const bulkDeleteBtn = document.getElementById('bulkDeleteBtn');
|
||||
|
||||
// Update selected count and toggle delete button visibility
|
||||
function updateSelectedCount() {
|
||||
const selectedCount = document.querySelectorAll('.group-media-checkbox:checked').length;
|
||||
selectedCountGroup.textContent = selectedCount;
|
||||
deleteSelectedGroup.style.display = selectedCount > 0 ? 'inline-block' : 'none';
|
||||
// Select all functionality
|
||||
if (selectAllCheckbox) {
|
||||
selectAllCheckbox.addEventListener('change', function() {
|
||||
mediaCheckboxes.forEach(checkbox => {
|
||||
checkbox.checked = this.checked;
|
||||
});
|
||||
updateBulkDeleteButton();
|
||||
});
|
||||
}
|
||||
|
||||
// Select/Deselect all checkboxes
|
||||
selectAllGroup.addEventListener('change', function() {
|
||||
const isChecked = selectAllGroup.checked;
|
||||
groupMediaCheckboxes.forEach(checkbox => {
|
||||
checkbox.checked = isChecked;
|
||||
});
|
||||
updateSelectedCount();
|
||||
});
|
||||
|
||||
// Individual checkbox change
|
||||
groupMediaCheckboxes.forEach(checkbox => {
|
||||
checkbox.addEventListener('change', updateSelectedCount);
|
||||
mediaCheckboxes.forEach(checkbox => {
|
||||
checkbox.addEventListener('change', function() {
|
||||
updateSelectAllState();
|
||||
updateBulkDeleteButton();
|
||||
});
|
||||
});
|
||||
|
||||
// Delete selected button click
|
||||
deleteSelectedGroup.addEventListener('click', function() {
|
||||
const selectedIds = Array.from(groupMediaCheckboxes)
|
||||
.filter(checkbox => checkbox.checked)
|
||||
.map(checkbox => checkbox.value);
|
||||
|
||||
if (selectedIds.length === 0) {
|
||||
alert('No media selected for deletion.');
|
||||
return;
|
||||
function updateSelectAllState() {
|
||||
const checkedBoxes = Array.from(mediaCheckboxes).filter(cb => cb.checked);
|
||||
|
||||
if (selectAllCheckbox) {
|
||||
selectAllCheckbox.checked = checkedBoxes.length === mediaCheckboxes.length && mediaCheckboxes.length > 0;
|
||||
selectAllCheckbox.indeterminate = checkedBoxes.length > 0 && checkedBoxes.length < mediaCheckboxes.length;
|
||||
}
|
||||
}
|
||||
|
||||
if (confirm(`Are you sure you want to delete ${selectedIds.length} selected media items?`)) {
|
||||
// Send bulk delete request
|
||||
fetch('{{ url_for("bulk_delete_group_content", group_id=group.id) }}', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json',
|
||||
'X-CSRFToken': '{{ csrf_token() if csrf_token else "" }}'
|
||||
},
|
||||
body: JSON.stringify({content_ids: selectedIds})
|
||||
})
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
if (data.success) {
|
||||
alert(`Successfully deleted ${data.deleted_count} media items.`);
|
||||
location.reload(); // Reload the page to update the media list
|
||||
} else {
|
||||
alert('Error deleting media: ' + (data.error || 'Unknown error'));
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
console.error('Error:', error);
|
||||
alert('An error occurred while deleting the media.');
|
||||
});
|
||||
function updateBulkDeleteButton() {
|
||||
const checkedBoxes = Array.from(mediaCheckboxes).filter(cb => cb.checked);
|
||||
if (bulkDeleteBtn) {
|
||||
bulkDeleteBtn.style.display = checkedBoxes.length > 0 ? 'inline-block' : 'none';
|
||||
}
|
||||
});
|
||||
}
|
||||
});
|
||||
|
||||
function confirmBulkDelete() {
|
||||
const checkedBoxes = Array.from(document.querySelectorAll('.media-checkbox:checked'));
|
||||
if (checkedBoxes.length === 0) {
|
||||
alert('No media files selected.');
|
||||
return;
|
||||
}
|
||||
|
||||
const count = checkedBoxes.length;
|
||||
const message = `Are you sure you want to delete ${count} selected media file${count > 1 ? 's' : ''}? This action cannot be undone.`;
|
||||
|
||||
if (confirm(message)) {
|
||||
// Create a form with selected IDs
|
||||
const form = document.createElement('form');
|
||||
form.method = 'POST';
|
||||
form.action = '{{ url_for("bulk_delete_group_content", group_id=group.id) }}';
|
||||
|
||||
checkedBoxes.forEach(checkbox => {
|
||||
const input = document.createElement('input');
|
||||
input.type = 'hidden';
|
||||
input.name = 'selected_content';
|
||||
input.value = checkbox.value;
|
||||
form.appendChild(input);
|
||||
});
|
||||
|
||||
document.body.appendChild(form);
|
||||
form.submit();
|
||||
}
|
||||
}
|
||||
</script>
|
||||
</body>
|
||||
</html>
|
||||
@@ -94,18 +94,27 @@
|
||||
<div class="card-body">
|
||||
{% if content %}
|
||||
<!-- Bulk Actions Controls -->
|
||||
<div class="mb-3 d-flex flex-wrap align-items-center gap-2">
|
||||
<div class="form-check">
|
||||
<input class="form-check-input" type="checkbox" id="selectAll" {% if player.groups %}disabled{% endif %}>
|
||||
<label class="form-check-label" for="selectAll">
|
||||
Select All
|
||||
</label>
|
||||
<div class="row mb-3">
|
||||
<div class="col-md-6">
|
||||
<div class="form-check">
|
||||
<input class="form-check-input" type="checkbox" id="selectAll" {% if player.groups %}disabled{% endif %}>
|
||||
<label class="form-check-label" for="selectAll">
|
||||
Select All
|
||||
</label>
|
||||
</div>
|
||||
</div>
|
||||
<div class="col-md-6 text-end">
|
||||
<button type="button" class="btn btn-danger" id="bulkDeleteBtn" {% if player.groups %}disabled{% endif %} style="display:none;" onclick="confirmBulkDelete()">
|
||||
<i class="bi bi-trash"></i> Delete Selected
|
||||
</button>
|
||||
</div>
|
||||
<button id="deleteSelected" class="btn btn-danger" {% if player.groups %}disabled{% endif %} style="display: none;">
|
||||
<i class="bi bi-trash"></i> Delete Selected (<span id="selectedCount">0</span>)
|
||||
</button>
|
||||
</div>
|
||||
|
||||
<!-- Bulk Delete Form -->
|
||||
<form id="bulkDeleteForm" action="{{ url_for('bulk_delete_player_content', player_id=player.id) }}" method="post" style="display:none;">
|
||||
<input type="hidden" name="selected_content_ids" id="selectedContentIds">
|
||||
</form>
|
||||
|
||||
<ul class="list-group sortable-list" id="mediaList">
|
||||
{% for media in content %}
|
||||
<li class="list-group-item {% if theme == 'dark' %}dark-mode{% endif %}"
|
||||
@@ -115,8 +124,10 @@
|
||||
<div class="d-flex flex-column flex-md-row align-items-md-center">
|
||||
<!-- Checkbox for bulk selection -->
|
||||
<div class="me-2">
|
||||
<input type="checkbox" class="form-check-input media-checkbox"
|
||||
value="{{ media.id }}"
|
||||
<input class="form-check-input media-checkbox"
|
||||
type="checkbox"
|
||||
name="selected_content"
|
||||
value="{{ media.id }}"
|
||||
{% if player.groups %}disabled{% endif %}>
|
||||
</div>
|
||||
|
||||
@@ -256,75 +267,77 @@ document.addEventListener('DOMContentLoaded', function() {
|
||||
}
|
||||
}
|
||||
|
||||
// Bulk delete functionality
|
||||
// Bulk selection functionality
|
||||
const selectAllCheckbox = document.getElementById('selectAll');
|
||||
const mediaCheckboxes = document.querySelectorAll('.media-checkbox');
|
||||
const deleteSelectedButton = document.getElementById('deleteSelected');
|
||||
const selectedCountSpan = document.getElementById('selectedCount');
|
||||
const bulkDeleteBtn = document.getElementById('bulkDeleteBtn');
|
||||
|
||||
// Update selected count and toggle delete button visibility
|
||||
function updateSelectedCount() {
|
||||
const selectedCount = document.querySelectorAll('.media-checkbox:checked').length;
|
||||
selectedCountSpan.textContent = selectedCount;
|
||||
deleteSelectedButton.style.display = selectedCount > 0 ? 'inline-block' : 'none';
|
||||
}
|
||||
|
||||
// Select/Deselect all checkboxes
|
||||
selectAllCheckbox.addEventListener('change', function() {
|
||||
mediaCheckboxes.forEach(checkbox => {
|
||||
checkbox.checked = selectAllCheckbox.checked;
|
||||
// Select all functionality
|
||||
if (selectAllCheckbox) {
|
||||
selectAllCheckbox.addEventListener('change', function() {
|
||||
mediaCheckboxes.forEach(checkbox => {
|
||||
if (!checkbox.disabled) {
|
||||
checkbox.checked = this.checked;
|
||||
}
|
||||
});
|
||||
updateBulkDeleteButton();
|
||||
});
|
||||
updateSelectedCount();
|
||||
});
|
||||
}
|
||||
|
||||
// Individual checkbox change
|
||||
mediaCheckboxes.forEach(checkbox => {
|
||||
checkbox.addEventListener('change', function() {
|
||||
// Uncheck "Select All" if any checkbox is unchecked
|
||||
if (!this.checked) {
|
||||
selectAllCheckbox.checked = false;
|
||||
}
|
||||
updateSelectedCount();
|
||||
updateSelectAllState();
|
||||
updateBulkDeleteButton();
|
||||
});
|
||||
});
|
||||
|
||||
// Delete selected media
|
||||
deleteSelectedButton.addEventListener('click', function() {
|
||||
const selectedIds = Array.from(mediaCheckboxes)
|
||||
.filter(checkbox => checkbox.checked)
|
||||
.map(checkbox => checkbox.value);
|
||||
|
||||
if (selectedIds.length === 0) {
|
||||
alert('No media selected for deletion.');
|
||||
return;
|
||||
function updateSelectAllState() {
|
||||
const enabledCheckboxes = Array.from(mediaCheckboxes).filter(cb => !cb.disabled);
|
||||
const checkedBoxes = enabledCheckboxes.filter(cb => cb.checked);
|
||||
|
||||
if (selectAllCheckbox) {
|
||||
selectAllCheckbox.checked = checkedBoxes.length === enabledCheckboxes.length && enabledCheckboxes.length > 0;
|
||||
selectAllCheckbox.indeterminate = checkedBoxes.length > 0 && checkedBoxes.length < enabledCheckboxes.length;
|
||||
}
|
||||
}
|
||||
|
||||
if (confirm(`Are you sure you want to delete ${selectedIds.length} selected media items?`)) {
|
||||
// Send bulk delete request
|
||||
fetch('{{ url_for("bulk_delete_player_content", player_id=player.id) }}', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json',
|
||||
'X-CSRFToken': '{{ csrf_token() if csrf_token else "" }}'
|
||||
},
|
||||
body: JSON.stringify({content_ids: selectedIds})
|
||||
})
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
if (data.success) {
|
||||
alert('Selected media deleted successfully.');
|
||||
location.reload(); // Reload the page to update the media list
|
||||
} else {
|
||||
alert('Error deleting media: ' + (data.error || 'Unknown error'));
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
console.error('Error:', error);
|
||||
alert('An error occurred while deleting the media.');
|
||||
});
|
||||
function updateBulkDeleteButton() {
|
||||
const checkedBoxes = Array.from(mediaCheckboxes).filter(cb => cb.checked);
|
||||
if (bulkDeleteBtn) {
|
||||
bulkDeleteBtn.style.display = checkedBoxes.length > 0 ? 'inline-block' : 'none';
|
||||
}
|
||||
});
|
||||
}
|
||||
});
|
||||
|
||||
function confirmBulkDelete() {
|
||||
const checkedBoxes = Array.from(document.querySelectorAll('.media-checkbox:checked'));
|
||||
if (checkedBoxes.length === 0) {
|
||||
alert('No media files selected.');
|
||||
return;
|
||||
}
|
||||
|
||||
const count = checkedBoxes.length;
|
||||
const message = `Are you sure you want to delete ${count} selected media file${count > 1 ? 's' : ''}? This action cannot be undone.`;
|
||||
|
||||
if (confirm(message)) {
|
||||
// Create a form with selected IDs
|
||||
const form = document.createElement('form');
|
||||
form.method = 'POST';
|
||||
form.action = '{{ url_for("bulk_delete_player_content", player_id=player.id) }}';
|
||||
|
||||
checkedBoxes.forEach(checkbox => {
|
||||
const input = document.createElement('input');
|
||||
input.type = 'hidden';
|
||||
input.name = 'selected_content';
|
||||
input.value = checkbox.value;
|
||||
form.appendChild(input);
|
||||
});
|
||||
|
||||
document.body.appendChild(form);
|
||||
form.submit();
|
||||
}
|
||||
}
|
||||
</script>
|
||||
</body>
|
||||
</html>
|
||||
@@ -136,37 +136,82 @@
|
||||
</div>
|
||||
<div class="modal-body">
|
||||
<p id="status-message">Uploading and processing your files. Please wait...</p>
|
||||
<div class="progress mb-3">
|
||||
<div id="progress-bar" class="progress-bar progress-bar-striped progress-bar-animated" role="progressbar" style="width: 0%;" aria-valuenow="0" aria-valuemin="0" aria-valuemax="100"></div>
|
||||
</div>
|
||||
|
||||
<!-- Real-time performance monitoring during upload -->
|
||||
<div class="performance-monitor mt-4">
|
||||
<h6>System Load During Upload:</h6>
|
||||
<div class="row">
|
||||
<div class="col-4">
|
||||
<small>CPU Usage</small>
|
||||
<div class="progress mb-1" style="height: 20px;">
|
||||
<div id="modal-cpu-bar" class="progress-bar bg-info" style="width: 0%;">0%</div>
|
||||
</div>
|
||||
</div>
|
||||
<div class="col-4">
|
||||
<small>Memory Usage</small>
|
||||
<div class="progress mb-1" style="height: 20px;">
|
||||
<div id="modal-memory-bar" class="progress-bar bg-warning" style="width: 0%;">0%</div>
|
||||
</div>
|
||||
</div>
|
||||
<div class="col-4">
|
||||
<small>Disk Usage</small>
|
||||
<div class="progress mb-1" style="height: 20px;">
|
||||
<div id="modal-disk-bar" class="progress-bar bg-danger" style="width: 0%;">0%</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
<div class="mt-2">
|
||||
<small id="perf-stats" class="text-muted">Waiting for performance data...</small>
|
||||
<!-- File Processing Progress -->
|
||||
<div class="mb-3">
|
||||
<label class="form-label fw-bold">File Processing Progress</label>
|
||||
<div class="progress" style="height: 25px;">
|
||||
<div id="progress-bar" class="progress-bar progress-bar-striped progress-bar-animated" role="progressbar" style="width: 0%;" aria-valuenow="0" aria-valuemin="0" aria-valuemax="100"></div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- System Monitoring in Modal -->
|
||||
{% if system_info %}
|
||||
<div class="mt-4">
|
||||
<h6 class="mb-3">📊 Server Performance During Upload</h6>
|
||||
<div class="row">
|
||||
<!-- CPU Usage -->
|
||||
<div class="col-md-4 col-12 mb-3">
|
||||
<label class="form-label">CPU Usage</label>
|
||||
<div class="progress" style="height: 20px;">
|
||||
<div id="cpu-progress" class="progress-bar
|
||||
{% if system_info.cpu_percent < 50 %}bg-success
|
||||
{% elif system_info.cpu_percent < 80 %}bg-warning
|
||||
{% else %}bg-danger{% endif %}"
|
||||
role="progressbar"
|
||||
style="width: {{ system_info.cpu_percent }}%;">
|
||||
{{ system_info.cpu_percent }}%
|
||||
</div>
|
||||
</div>
|
||||
<small class="text-muted">{{ system_info.cpu_count }} cores available</small>
|
||||
</div>
|
||||
|
||||
<!-- Memory Usage -->
|
||||
<div class="col-md-4 col-12 mb-3">
|
||||
<label class="form-label">Memory Usage</label>
|
||||
<div class="progress" style="height: 20px;">
|
||||
<div id="memory-progress" class="progress-bar
|
||||
{% if system_info.memory_percent < 60 %}bg-success
|
||||
{% elif system_info.memory_percent < 85 %}bg-warning
|
||||
{% else %}bg-danger{% endif %}"
|
||||
role="progressbar"
|
||||
style="width: {{ system_info.memory_percent }}%;">
|
||||
{{ system_info.memory_percent }}%
|
||||
</div>
|
||||
</div>
|
||||
<small class="text-muted" id="memory-text">{{ system_info.memory_used }}GB / {{ system_info.memory_total }}GB</small>
|
||||
</div>
|
||||
|
||||
<!-- Disk Usage -->
|
||||
<div class="col-md-4 col-12 mb-3">
|
||||
<label class="form-label">Disk Space</label>
|
||||
<div class="progress" style="height: 20px;">
|
||||
<div id="disk-progress" class="progress-bar
|
||||
{% if system_info.disk_percent < 70 %}bg-success
|
||||
{% elif system_info.disk_percent < 90 %}bg-warning
|
||||
{% else %}bg-danger{% endif %}"
|
||||
role="progressbar"
|
||||
style="width: {{ system_info.disk_percent }}%;">
|
||||
{{ system_info.disk_percent }}%
|
||||
</div>
|
||||
</div>
|
||||
<small class="text-muted" id="disk-text">{{ system_info.disk_free }}GB free</small>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Storage Summary -->
|
||||
<div class="row mt-2">
|
||||
<div class="col-md-6 col-12 text-center">
|
||||
<strong>Current Media Storage:</strong>
|
||||
<span class="text-primary" id="storage-size">{{ system_info.upload_folder_size }}GB</span>
|
||||
</div>
|
||||
<div class="col-md-6 col-12 text-center">
|
||||
<strong>Last Updated:</strong>
|
||||
<span class="text-muted" id="modal-last-update">Just now</span>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
{% endif %}
|
||||
</div>
|
||||
<div class="modal-footer {{ 'dark-mode' if theme == 'dark' else '' }}">
|
||||
<button type="button" class="btn btn-secondary" data-bs-dismiss="modal" disabled>Close</button>
|
||||
@@ -178,6 +223,145 @@
|
||||
|
||||
<script src="https://cdn.jsdelivr.net/npm/bootstrap@5.3.0-alpha3/dist/js/bootstrap.bundle.min.js"></script>
|
||||
<script>
|
||||
function showStatusModal() {
|
||||
console.log("Processing popup triggered");
|
||||
const statusModal = new bootstrap.Modal(document.getElementById('statusModal'));
|
||||
statusModal.show();
|
||||
|
||||
// Update status message based on media type
|
||||
const mediaType = document.getElementById('media_type').value;
|
||||
const statusMessage = document.getElementById('status-message');
|
||||
|
||||
switch(mediaType) {
|
||||
case 'image':
|
||||
statusMessage.textContent = 'Uploading images...';
|
||||
break;
|
||||
case 'video':
|
||||
statusMessage.textContent = 'Uploading and processing video. This may take a while...';
|
||||
break;
|
||||
case 'pdf':
|
||||
statusMessage.textContent = 'Converting PDF to 4K images. This may take a while...';
|
||||
break;
|
||||
case 'ppt':
|
||||
statusMessage.textContent = 'Converting PowerPoint to 4K images. This may take a while...';
|
||||
break;
|
||||
default:
|
||||
statusMessage.textContent = 'Uploading and processing your files. Please wait...';
|
||||
}
|
||||
|
||||
// Start system monitoring updates in modal
|
||||
{% if system_info %}
|
||||
startModalSystemMonitoring();
|
||||
{% endif %}
|
||||
|
||||
// Simulate progress updates
|
||||
const progressBar = document.getElementById('progress-bar');
|
||||
let progress = 0;
|
||||
const interval = setInterval(() => {
|
||||
// For slow processes, increment more slowly
|
||||
const increment = (mediaType === 'image') ? 20 : 5;
|
||||
progress += increment;
|
||||
|
||||
if (progress >= 100) {
|
||||
clearInterval(interval);
|
||||
statusMessage.textContent = 'Files uploaded and processed successfully!';
|
||||
|
||||
// Stop system monitoring updates
|
||||
{% if system_info %}
|
||||
stopModalSystemMonitoring();
|
||||
{% endif %}
|
||||
|
||||
// Enable the close button
|
||||
document.querySelector('[data-bs-dismiss="modal"]').disabled = false;
|
||||
} else {
|
||||
progressBar.style.width = `${progress}%`;
|
||||
progressBar.setAttribute('aria-valuenow', progress);
|
||||
}
|
||||
}, 500);
|
||||
}
|
||||
|
||||
{% if system_info %}
|
||||
let modalSystemInterval;
|
||||
|
||||
function updateModalSystemInfo() {
|
||||
fetch('/api/system_info')
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
if (data.error) {
|
||||
console.warn('Could not fetch system info:', data.error);
|
||||
return;
|
||||
}
|
||||
|
||||
// Update CPU
|
||||
const cpuProgress = document.getElementById('cpu-progress');
|
||||
if (cpuProgress) {
|
||||
cpuProgress.style.width = data.cpu_percent + '%';
|
||||
cpuProgress.textContent = data.cpu_percent + '%';
|
||||
cpuProgress.className = 'progress-bar ' +
|
||||
(data.cpu_percent < 50 ? 'bg-success' :
|
||||
data.cpu_percent < 80 ? 'bg-warning' : 'bg-danger');
|
||||
}
|
||||
|
||||
// Update Memory
|
||||
const memoryProgress = document.getElementById('memory-progress');
|
||||
const memoryText = document.getElementById('memory-text');
|
||||
if (memoryProgress) {
|
||||
memoryProgress.style.width = data.memory_percent + '%';
|
||||
memoryProgress.textContent = data.memory_percent + '%';
|
||||
memoryProgress.className = 'progress-bar ' +
|
||||
(data.memory_percent < 60 ? 'bg-success' :
|
||||
data.memory_percent < 85 ? 'bg-warning' : 'bg-danger');
|
||||
}
|
||||
if (memoryText) {
|
||||
memoryText.textContent = data.memory_used + 'GB / ' + data.memory_total + 'GB';
|
||||
}
|
||||
|
||||
// Update Disk
|
||||
const diskProgress = document.getElementById('disk-progress');
|
||||
const diskText = document.getElementById('disk-text');
|
||||
if (diskProgress) {
|
||||
diskProgress.style.width = data.disk_percent + '%';
|
||||
diskProgress.textContent = data.disk_percent + '%';
|
||||
diskProgress.className = 'progress-bar ' +
|
||||
(data.disk_percent < 70 ? 'bg-success' :
|
||||
data.disk_percent < 90 ? 'bg-warning' : 'bg-danger');
|
||||
}
|
||||
if (diskText) {
|
||||
diskText.textContent = data.disk_free + 'GB free';
|
||||
}
|
||||
|
||||
// Update storage size
|
||||
const storageSize = document.getElementById('storage-size');
|
||||
if (storageSize) {
|
||||
storageSize.textContent = data.upload_folder_size + 'GB';
|
||||
}
|
||||
|
||||
// Update timestamp
|
||||
const lastUpdate = document.getElementById('modal-last-update');
|
||||
if (lastUpdate) {
|
||||
lastUpdate.textContent = new Date().toLocaleTimeString();
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
console.warn('Modal system monitoring update failed:', error);
|
||||
});
|
||||
}
|
||||
|
||||
function startModalSystemMonitoring() {
|
||||
// Update immediately
|
||||
updateModalSystemInfo();
|
||||
// Then update every 3 seconds for real-time monitoring during upload
|
||||
modalSystemInterval = setInterval(updateModalSystemInfo, 3000);
|
||||
}
|
||||
|
||||
function stopModalSystemMonitoring() {
|
||||
if (modalSystemInterval) {
|
||||
clearInterval(modalSystemInterval);
|
||||
modalSystemInterval = null;
|
||||
}
|
||||
}
|
||||
{% endif %}
|
||||
|
||||
function updateTargetIdOptions() {
|
||||
const targetType = document.getElementById('target_type').value;
|
||||
const targetIdSelect = document.getElementById('target_id');
|
||||
@@ -254,9 +438,6 @@
|
||||
statusMessage.textContent = 'Uploading and processing your files. Please wait...';
|
||||
}
|
||||
|
||||
// Start performance monitoring during upload
|
||||
startUploadMonitoring();
|
||||
|
||||
// Simulate progress updates
|
||||
const progressBar = document.getElementById('progress-bar');
|
||||
let progress = 0;
|
||||
@@ -267,7 +448,6 @@
|
||||
|
||||
if (progress >= 100) {
|
||||
clearInterval(interval);
|
||||
stopUploadMonitoring();
|
||||
statusMessage.textContent = 'Files uploaded and processed successfully!';
|
||||
|
||||
// Enable the close button
|
||||
@@ -278,82 +458,6 @@
|
||||
}
|
||||
}, 500);
|
||||
}
|
||||
|
||||
// Performance monitoring during upload
|
||||
let uploadMonitoringInterval = null;
|
||||
let startCpu = 0, startMemory = 0;
|
||||
let maxUploadCpu = 0, maxUploadMemory = 0;
|
||||
|
||||
function updateModalPerformance(data) {
|
||||
// Update CPU bar
|
||||
const cpuBar = document.getElementById('modal-cpu-bar');
|
||||
cpuBar.style.width = `${data.cpu_percent}%`;
|
||||
cpuBar.textContent = `${data.cpu_percent.toFixed(1)}%`;
|
||||
if (data.cpu_percent > 75) cpuBar.className = 'progress-bar bg-danger';
|
||||
else if (data.cpu_percent > 50) cpuBar.className = 'progress-bar bg-warning';
|
||||
else cpuBar.className = 'progress-bar bg-info';
|
||||
|
||||
// Update Memory bar
|
||||
const memoryBar = document.getElementById('modal-memory-bar');
|
||||
memoryBar.style.width = `${data.memory_percent}%`;
|
||||
memoryBar.textContent = `${data.memory_percent.toFixed(1)}%`;
|
||||
if (data.memory_percent > 75) memoryBar.className = 'progress-bar bg-danger';
|
||||
else if (data.memory_percent > 50) memoryBar.className = 'progress-bar bg-warning';
|
||||
else memoryBar.className = 'progress-bar bg-warning';
|
||||
|
||||
// Update Disk bar
|
||||
const diskBar = document.getElementById('modal-disk-bar');
|
||||
diskBar.style.width = `${data.disk_percent}%`;
|
||||
diskBar.textContent = `${data.disk_percent.toFixed(1)}%`;
|
||||
if (data.disk_percent > 85) diskBar.className = 'progress-bar bg-danger';
|
||||
else diskBar.className = 'progress-bar bg-danger';
|
||||
|
||||
// Track peaks
|
||||
if (data.cpu_percent > maxUploadCpu) maxUploadCpu = data.cpu_percent;
|
||||
if (data.memory_percent > maxUploadMemory) maxUploadMemory = data.memory_percent;
|
||||
|
||||
// Update stats text
|
||||
const perfStats = document.getElementById('perf-stats');
|
||||
const cpuChange = startCpu ? (data.cpu_percent - startCpu).toFixed(1) : '0.0';
|
||||
const memChange = startMemory ? (data.memory_percent - startMemory).toFixed(1) : '0.0';
|
||||
perfStats.innerHTML = `CPU: ${cpuChange > 0 ? '+' : ''}${cpuChange}% | Memory: ${memChange > 0 ? '+' : ''}${memChange}% | Peak CPU: ${maxUploadCpu.toFixed(1)}%`;
|
||||
}
|
||||
|
||||
function startUploadMonitoring() {
|
||||
// Get baseline performance
|
||||
fetch('/api/performance')
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
if (!data.error) {
|
||||
startCpu = data.cpu_percent;
|
||||
startMemory = data.memory_percent;
|
||||
maxUploadCpu = data.cpu_percent;
|
||||
maxUploadMemory = data.memory_percent;
|
||||
updateModalPerformance(data);
|
||||
}
|
||||
});
|
||||
|
||||
// Start monitoring every 1 second during upload
|
||||
uploadMonitoringInterval = setInterval(() => {
|
||||
fetch('/api/performance')
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
if (!data.error) {
|
||||
updateModalPerformance(data);
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
console.log('Performance monitoring error:', error);
|
||||
});
|
||||
}, 1000);
|
||||
}
|
||||
|
||||
function stopUploadMonitoring() {
|
||||
if (uploadMonitoringInterval) {
|
||||
clearInterval(uploadMonitoringInterval);
|
||||
uploadMonitoringInterval = null;
|
||||
}
|
||||
}
|
||||
</script>
|
||||
</body>
|
||||
</html>
|
||||
@@ -333,7 +333,7 @@ def edit_group_media(group_id, content_id, new_duration):
|
||||
|
||||
def delete_group_media(group_id, content_id):
|
||||
"""
|
||||
Delete a media item from all players in a group.
|
||||
Delete a media item from all players in a group and remove the physical file.
|
||||
|
||||
Args:
|
||||
group_id (int): ID of the group
|
||||
@@ -344,6 +344,8 @@ def delete_group_media(group_id, content_id):
|
||||
"""
|
||||
from models import Group, Content
|
||||
from extensions import db
|
||||
from flask import current_app
|
||||
import os
|
||||
|
||||
group = Group.query.get_or_404(group_id)
|
||||
content = Content.query.get(content_id)
|
||||
@@ -358,6 +360,19 @@ def delete_group_media(group_id, content_id):
|
||||
db.session.delete(content)
|
||||
count += 1
|
||||
|
||||
# Delete the physical file using absolute path
|
||||
upload_folder = current_app.config['UPLOAD_FOLDER']
|
||||
if not os.path.isabs(upload_folder):
|
||||
upload_folder = os.path.abspath(upload_folder)
|
||||
file_path = os.path.join(upload_folder, file_name)
|
||||
|
||||
if os.path.exists(file_path):
|
||||
try:
|
||||
os.remove(file_path)
|
||||
print(f"Deleted physical file: {file_path}")
|
||||
except OSError as e:
|
||||
print(f"Error deleting file {file_path}: {e}")
|
||||
|
||||
db.session.commit()
|
||||
|
||||
# Log the content deletion
|
||||
@@ -366,4 +381,5 @@ def delete_group_media(group_id, content_id):
|
||||
return True
|
||||
except Exception as e:
|
||||
db.session.rollback()
|
||||
print(f"Error in delete_group_media: {e}")
|
||||
return False
|
||||
86
app/utils/pptx_converter.py
Normal file
86
app/utils/pptx_converter.py
Normal file
@@ -0,0 +1,86 @@
|
||||
"""
|
||||
PPTX to PDF converter using LibreOffice for high-quality conversion
|
||||
This module provides the essential function to convert PowerPoint presentations to PDF
|
||||
using LibreOffice headless mode for professional-grade quality.
|
||||
|
||||
The converted PDF is then processed by the main upload workflow for 4K image generation.
|
||||
"""
|
||||
|
||||
import os
|
||||
import subprocess
|
||||
import logging
|
||||
|
||||
# Set up logging
|
||||
logging.basicConfig(level=logging.INFO)
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def pptx_to_pdf_libreoffice(pptx_path, output_dir):
|
||||
"""
|
||||
Convert PPTX to PDF using LibreOffice for highest quality.
|
||||
|
||||
This function is the core component of the PPTX processing workflow:
|
||||
PPTX → PDF (this function) → 4K JPG images (handled in uploads.py)
|
||||
|
||||
Args:
|
||||
pptx_path (str): Path to the PPTX file
|
||||
output_dir (str): Directory to save the PDF
|
||||
|
||||
Returns:
|
||||
str: Path to the generated PDF file, or None if conversion failed
|
||||
"""
|
||||
try:
|
||||
# Ensure output directory exists
|
||||
os.makedirs(output_dir, exist_ok=True)
|
||||
|
||||
# Use LibreOffice to convert PPTX to PDF
|
||||
cmd = [
|
||||
'libreoffice',
|
||||
'--headless',
|
||||
'--convert-to', 'pdf',
|
||||
'--outdir', output_dir,
|
||||
pptx_path
|
||||
]
|
||||
|
||||
logger.info(f"Converting PPTX to PDF using LibreOffice: {pptx_path}")
|
||||
result = subprocess.run(cmd, capture_output=True, text=True, timeout=120)
|
||||
|
||||
if result.returncode != 0:
|
||||
logger.error(f"LibreOffice conversion failed: {result.stderr}")
|
||||
return None
|
||||
|
||||
# Find the generated PDF file
|
||||
base_name = os.path.splitext(os.path.basename(pptx_path))[0]
|
||||
pdf_path = os.path.join(output_dir, f"{base_name}.pdf")
|
||||
|
||||
if os.path.exists(pdf_path):
|
||||
logger.info(f"PDF conversion successful: {pdf_path}")
|
||||
return pdf_path
|
||||
else:
|
||||
logger.error(f"PDF file not found after conversion: {pdf_path}")
|
||||
return None
|
||||
|
||||
except subprocess.TimeoutExpired:
|
||||
logger.error("LibreOffice conversion timed out (120s)")
|
||||
return None
|
||||
except Exception as e:
|
||||
logger.error(f"Error in PPTX to PDF conversion: {e}")
|
||||
return None
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
# Test the converter
|
||||
import sys
|
||||
if len(sys.argv) > 1:
|
||||
test_pptx = sys.argv[1]
|
||||
if os.path.exists(test_pptx):
|
||||
output_dir = "test_output"
|
||||
pdf_result = pptx_to_pdf_libreoffice(test_pptx, output_dir)
|
||||
if pdf_result:
|
||||
print(f"Successfully converted PPTX to PDF: {pdf_result}")
|
||||
else:
|
||||
print("PPTX to PDF conversion failed")
|
||||
else:
|
||||
print(f"File not found: {test_pptx}")
|
||||
else:
|
||||
print("Usage: python pptx_converter.py <pptx_file>")
|
||||
@@ -1,34 +1,36 @@
|
||||
import os
|
||||
import subprocess
|
||||
import signal
|
||||
import psutil
|
||||
import time
|
||||
from flask import Flask
|
||||
from werkzeug.utils import secure_filename
|
||||
from pdf2image import convert_from_path
|
||||
from pptx import Presentation
|
||||
from PIL import Image, ImageDraw, ImageFont
|
||||
import io
|
||||
from extensions import db
|
||||
from models import Content, Player, Group
|
||||
from utils.logger import log_content_added, log_upload, log_process
|
||||
|
||||
# Add timeout handling class
|
||||
class TimeoutError(Exception):
|
||||
pass
|
||||
|
||||
def timeout_handler(signum, frame):
|
||||
raise TimeoutError("Operation timed out")
|
||||
|
||||
# Function to add image to playlist
|
||||
def add_image_to_playlist(app, file, filename, duration, target_type, target_id):
|
||||
"""
|
||||
Save the image file and add it to the playlist database.
|
||||
"""
|
||||
file_path = os.path.join(app.config['UPLOAD_FOLDER'], filename)
|
||||
# Ensure we use absolute path for upload folder
|
||||
upload_folder = app.config['UPLOAD_FOLDER']
|
||||
if not os.path.isabs(upload_folder):
|
||||
upload_folder = os.path.abspath(upload_folder)
|
||||
|
||||
# Ensure upload folder exists
|
||||
if not os.path.exists(upload_folder):
|
||||
os.makedirs(upload_folder, exist_ok=True)
|
||||
print(f"Created upload folder: {upload_folder}")
|
||||
|
||||
file_path = os.path.join(upload_folder, filename)
|
||||
print(f"Saving image to: {file_path}")
|
||||
|
||||
# Only save if file does not already exist
|
||||
if not os.path.exists(file_path):
|
||||
file.save(file_path)
|
||||
print(f"Image saved successfully: {file_path}")
|
||||
else:
|
||||
print(f"File already exists: {file_path}")
|
||||
|
||||
print(f"Adding image to playlist: {filename}, Target Type: {target_type}, Target ID: {target_id}")
|
||||
|
||||
@@ -53,12 +55,19 @@ def convert_video(input_file, output_folder):
|
||||
"""
|
||||
Converts a video file to MP4 format with H.264 codec.
|
||||
"""
|
||||
# Ensure we use absolute path for output folder
|
||||
if not os.path.isabs(output_folder):
|
||||
output_folder = os.path.abspath(output_folder)
|
||||
print(f"Converted output folder to absolute path: {output_folder}")
|
||||
|
||||
if not os.path.exists(output_folder):
|
||||
os.makedirs(output_folder)
|
||||
os.makedirs(output_folder, exist_ok=True)
|
||||
print(f"Created output folder: {output_folder}")
|
||||
|
||||
# Generate the output file path
|
||||
base_name = os.path.splitext(os.path.basename(input_file))[0]
|
||||
output_file = os.path.join(output_folder, f"{base_name}.mp4")
|
||||
print(f"Converting video: {input_file} -> {output_file}")
|
||||
|
||||
# FFmpeg command to convert the video
|
||||
command = [
|
||||
@@ -88,7 +97,14 @@ def convert_video_and_update_playlist(app, file_path, original_filename, target_
|
||||
Converts a video and updates the playlist database.
|
||||
"""
|
||||
print(f"Starting video conversion for: {file_path}")
|
||||
converted_file = convert_video(file_path, app.config['UPLOAD_FOLDER'])
|
||||
|
||||
# Ensure we use absolute path for upload folder
|
||||
upload_folder = app.config['UPLOAD_FOLDER']
|
||||
if not os.path.isabs(upload_folder):
|
||||
upload_folder = os.path.abspath(upload_folder)
|
||||
print(f"Converted upload folder to absolute path: {upload_folder}")
|
||||
|
||||
converted_file = convert_video(file_path, upload_folder)
|
||||
if converted_file:
|
||||
converted_filename = os.path.basename(converted_file)
|
||||
print(f"Video converted successfully: {converted_filename}")
|
||||
@@ -118,39 +134,79 @@ def convert_video_and_update_playlist(app, file_path, original_filename, target_
|
||||
print(f"Video conversion failed for: {file_path}")
|
||||
|
||||
# PDF conversion functions
|
||||
def convert_pdf_to_images(pdf_file, output_folder, delete_pdf=True, dpi=600):
|
||||
def convert_pdf_to_images(pdf_file, output_folder, delete_pdf=True, dpi=300):
|
||||
"""
|
||||
Convert a PDF file to images in sequential order at high resolution (4K).
|
||||
Convert a PDF file to high-quality JPG images in sequential order.
|
||||
Uses standard 300 DPI for reliable conversion.
|
||||
"""
|
||||
print(f"Converting PDF to images: {pdf_file} at {dpi} DPI")
|
||||
print(f"Converting PDF to JPG images: {pdf_file} at {dpi} DPI")
|
||||
print(f"Original output folder: {output_folder}")
|
||||
|
||||
# Force absolute path resolution to ensure we use the app directory
|
||||
if not os.path.isabs(output_folder):
|
||||
# If relative path, resolve from the current working directory
|
||||
output_folder = os.path.abspath(output_folder)
|
||||
print(f"Converted relative path to absolute: {output_folder}")
|
||||
else:
|
||||
print(f"Using provided absolute path: {output_folder}")
|
||||
|
||||
# Ensure we're using the app static folder, not workspace root
|
||||
if output_folder.endswith('static/uploads'):
|
||||
# Check if we're accidentally using workspace root instead of app folder
|
||||
expected_app_path = '/opt/digiserver/app/static/uploads'
|
||||
if output_folder != expected_app_path:
|
||||
print(f"WARNING: Correcting path from {output_folder} to {expected_app_path}")
|
||||
output_folder = expected_app_path
|
||||
|
||||
print(f"Final output folder: {output_folder}")
|
||||
|
||||
try:
|
||||
# Convert PDF to images
|
||||
# Ensure output folder exists
|
||||
if not os.path.exists(output_folder):
|
||||
os.makedirs(output_folder, exist_ok=True)
|
||||
print(f"Created output folder: {output_folder}")
|
||||
|
||||
# Convert PDF to images using pdf2image
|
||||
print("Starting PDF conversion...")
|
||||
images = convert_from_path(pdf_file, dpi=dpi)
|
||||
print(f"PDF converted to {len(images)} page(s)")
|
||||
|
||||
if not images:
|
||||
print("ERROR: No images generated from PDF")
|
||||
return []
|
||||
|
||||
base_name = os.path.splitext(os.path.basename(pdf_file))[0]
|
||||
image_filenames = []
|
||||
|
||||
# Save each page as an image with zero-padded page numbers for proper sorting
|
||||
# Save each page as JPG image
|
||||
for i, image in enumerate(images):
|
||||
# Use consistent naming with zero-padded page numbers (e.g., page_001.jpg)
|
||||
# Convert to RGB if necessary
|
||||
if image.mode != 'RGB':
|
||||
image = image.convert('RGB')
|
||||
|
||||
# Simple naming with page numbers
|
||||
page_num = str(i + 1).zfill(3) # e.g., 001, 002, etc.
|
||||
image_filename = f"{base_name}_page_{page_num}.jpg"
|
||||
image_path = os.path.join(output_folder, image_filename)
|
||||
image.save(image_path, 'JPEG')
|
||||
|
||||
# Save as JPG
|
||||
image.save(image_path, 'JPEG', quality=85, optimize=True)
|
||||
image_filenames.append(image_filename)
|
||||
print(f"Saved page {i + 1} as image: {image_path}")
|
||||
print(f"Saved page {i + 1} to: {image_path}")
|
||||
|
||||
# Verify all pages were saved
|
||||
print(f"PDF conversion complete. {len(image_filenames)} pages saved.")
|
||||
print(f"Images in order: {image_filenames}")
|
||||
|
||||
# Delete the PDF file if requested
|
||||
if delete_pdf and os.path.exists(pdf_file):
|
||||
print(f"PDF conversion complete. {len(image_filenames)} JPG images saved to {output_folder}")
|
||||
|
||||
# Delete the PDF file if requested and conversion was successful
|
||||
if delete_pdf and os.path.exists(pdf_file) and image_filenames:
|
||||
os.remove(pdf_file)
|
||||
print(f"PDF file deleted: {pdf_file}")
|
||||
|
||||
|
||||
return image_filenames
|
||||
|
||||
except Exception as e:
|
||||
print(f"Error converting PDF to images: {e}")
|
||||
print(f"Error converting PDF to JPG images: {e}")
|
||||
import traceback
|
||||
traceback.print_exc()
|
||||
return []
|
||||
|
||||
def update_playlist_with_files(image_filenames, duration, target_type, target_id):
|
||||
@@ -207,126 +263,35 @@ def process_pdf(input_file, output_folder, duration, target_type, target_id):
|
||||
Returns:
|
||||
bool: True if successful, False otherwise
|
||||
"""
|
||||
print(f"Processing PDF file: {input_file}")
|
||||
print(f"Output folder: {output_folder}")
|
||||
|
||||
# Ensure we have absolute path for output folder
|
||||
if not os.path.isabs(output_folder):
|
||||
output_folder = os.path.abspath(output_folder)
|
||||
print(f"Converted output folder to absolute path: {output_folder}")
|
||||
|
||||
# Ensure output folder exists
|
||||
if not os.path.exists(output_folder):
|
||||
os.makedirs(output_folder)
|
||||
os.makedirs(output_folder, exist_ok=True)
|
||||
print(f"Created output folder: {output_folder}")
|
||||
|
||||
# Convert PDF to images
|
||||
image_filenames = convert_pdf_to_images(input_file, output_folder)
|
||||
# Convert PDF to images using standard quality (delete PDF after successful conversion)
|
||||
image_filenames = convert_pdf_to_images(input_file, output_folder, delete_pdf=True, dpi=300)
|
||||
|
||||
# Update playlist with generated images
|
||||
if image_filenames:
|
||||
return update_playlist_with_files(image_filenames, duration, target_type, target_id)
|
||||
return False
|
||||
success = update_playlist_with_files(image_filenames, duration, target_type, target_id)
|
||||
if success:
|
||||
print(f"Successfully processed PDF: {len(image_filenames)} images added to playlist")
|
||||
return success
|
||||
else:
|
||||
print("Failed to convert PDF to images")
|
||||
return False
|
||||
|
||||
def convert_pptx_to_images_direct(pptx_file, output_folder, delete_pptx=True, dpi=300):
|
||||
def process_pptx(input_file, output_folder, duration, target_type, target_id):
|
||||
"""
|
||||
Convert a PPTX file directly to images using python-pptx library.
|
||||
This eliminates the need for LibreOffice and provides more reliable conversion.
|
||||
|
||||
Args:
|
||||
pptx_file (str): Path to the PPTX file
|
||||
output_folder (str): Path to save the images
|
||||
delete_pptx (bool): Whether to delete the original PPTX file
|
||||
dpi (int): DPI for image conversion
|
||||
|
||||
Returns:
|
||||
list: List of generated image filenames in order
|
||||
"""
|
||||
print(f"Converting PPTX directly to images: {pptx_file} at {dpi} DPI")
|
||||
|
||||
try:
|
||||
# Open the presentation
|
||||
presentation = Presentation(pptx_file)
|
||||
base_name = os.path.splitext(os.path.basename(pptx_file))[0]
|
||||
image_filenames = []
|
||||
|
||||
print(f"PPTX has {len(presentation.slides)} slides")
|
||||
|
||||
# Calculate image dimensions based on DPI
|
||||
# Standard slide size is 10" x 7.5" (25.4cm x 19.05cm)
|
||||
width_px = int(10 * dpi) # 10 inches * DPI
|
||||
height_px = int(7.5 * dpi) # 7.5 inches * DPI
|
||||
|
||||
for i, slide in enumerate(presentation.slides):
|
||||
try:
|
||||
# Use zero-padded page numbers for proper sorting
|
||||
page_num = str(i + 1).zfill(3)
|
||||
image_filename = f"{base_name}_page_{page_num}.jpg"
|
||||
image_path = os.path.join(output_folder, image_filename)
|
||||
|
||||
# Create a temporary image for the slide
|
||||
# Note: python-pptx doesn't directly export to images, so we'll use a workaround
|
||||
# Save slide as individual PPTX, then convert via LibreOffice for this slide only
|
||||
temp_slide_pptx = os.path.join(output_folder, f"temp_slide_{i+1}.pptx")
|
||||
temp_slide_pdf = os.path.join(output_folder, f"temp_slide_{i+1}.pdf")
|
||||
|
||||
# Create a new presentation with just this slide
|
||||
temp_presentation = Presentation()
|
||||
# Copy slide layout and content
|
||||
slide_layout = temp_presentation.slide_layouts[0] # Use blank layout
|
||||
temp_slide = temp_presentation.slides.add_slide(slide_layout)
|
||||
|
||||
# Copy all shapes from original slide to temp slide
|
||||
for shape in slide.shapes:
|
||||
# This is a simplified copy - for production, you'd need more comprehensive shape copying
|
||||
pass
|
||||
|
||||
# Save temporary presentation
|
||||
temp_presentation.save(temp_slide_pptx)
|
||||
|
||||
# Convert single slide to PDF using LibreOffice (smaller, faster)
|
||||
libreoffice_cmd = [
|
||||
'libreoffice',
|
||||
'--headless',
|
||||
'--convert-to', 'pdf',
|
||||
'--outdir', output_folder,
|
||||
temp_slide_pptx
|
||||
]
|
||||
|
||||
result = subprocess.run(libreoffice_cmd, capture_output=True, text=True, timeout=60)
|
||||
|
||||
if result.returncode == 0 and os.path.exists(temp_slide_pdf):
|
||||
# Convert PDF to image
|
||||
images = convert_from_path(temp_slide_pdf, dpi=dpi)
|
||||
if images:
|
||||
images[0].save(image_path, 'JPEG', quality=85, optimize=True)
|
||||
image_filenames.append(image_filename)
|
||||
print(f"Saved slide {i + 1}/{len(presentation.slides)} as: {image_filename}")
|
||||
|
||||
# Clean up temporary files
|
||||
if os.path.exists(temp_slide_pdf):
|
||||
os.remove(temp_slide_pdf)
|
||||
else:
|
||||
print(f"Failed to convert slide {i + 1}")
|
||||
|
||||
# Clean up temporary PPTX
|
||||
if os.path.exists(temp_slide_pptx):
|
||||
os.remove(temp_slide_pptx)
|
||||
|
||||
except Exception as e:
|
||||
print(f"Error processing slide {i + 1}: {e}")
|
||||
continue
|
||||
|
||||
print(f"PPTX conversion complete. Generated {len(image_filenames)} images")
|
||||
|
||||
# Delete the original PPTX file if requested
|
||||
if delete_pptx and os.path.exists(pptx_file):
|
||||
pptx_size = os.path.getsize(pptx_file) / (1024*1024)
|
||||
os.remove(pptx_file)
|
||||
print(f"Original PPTX file deleted: {pptx_file} ({pptx_size:.2f} MB freed)")
|
||||
|
||||
return image_filenames
|
||||
|
||||
except Exception as e:
|
||||
print(f"Error converting PPTX to images: {e}")
|
||||
import traceback
|
||||
traceback.print_exc()
|
||||
return []
|
||||
|
||||
def process_pptx_improved(input_file, output_folder, duration, target_type, target_id):
|
||||
"""
|
||||
Improved PPTX processing function that's more reliable and faster.
|
||||
Process a PPTX file: convert to PDF first, then to JPG images (same workflow as PDF).
|
||||
|
||||
Args:
|
||||
input_file (str): Path to the PPTX file
|
||||
@@ -338,43 +303,49 @@ def process_pptx_improved(input_file, output_folder, duration, target_type, targ
|
||||
Returns:
|
||||
bool: True if successful, False otherwise
|
||||
"""
|
||||
print(f"=== Starting Improved PPTX Processing ===")
|
||||
print(f"Input file: {input_file}")
|
||||
print(f"Processing PPTX file using PDF workflow: {input_file}")
|
||||
print(f"Output folder: {output_folder}")
|
||||
|
||||
# Ensure we have absolute path for output folder
|
||||
if not os.path.isabs(output_folder):
|
||||
output_folder = os.path.abspath(output_folder)
|
||||
print(f"Converted output folder to absolute path: {output_folder}")
|
||||
|
||||
# Ensure output folder exists
|
||||
if not os.path.exists(output_folder):
|
||||
os.makedirs(output_folder, exist_ok=True)
|
||||
print(f"Created output folder: {output_folder}")
|
||||
|
||||
try:
|
||||
file_size = os.path.getsize(input_file) / (1024*1024)
|
||||
print(f"File size: {file_size:.2f} MB")
|
||||
# Step 1: Convert PPTX to PDF using LibreOffice for vector quality
|
||||
from utils.pptx_converter import pptx_to_pdf_libreoffice
|
||||
pdf_file = pptx_to_pdf_libreoffice(input_file, output_folder)
|
||||
|
||||
# Ensure output folder exists
|
||||
if not os.path.exists(output_folder):
|
||||
os.makedirs(output_folder)
|
||||
|
||||
# Check if LibreOffice is available
|
||||
try:
|
||||
result = subprocess.run(['libreoffice', '--version'], capture_output=True, text=True, timeout=10)
|
||||
if result.returncode != 0:
|
||||
print("LibreOffice not available, falling back to basic conversion")
|
||||
return False
|
||||
except:
|
||||
print("LibreOffice not available, falling back to basic conversion")
|
||||
return False
|
||||
|
||||
# Convert PPTX directly to images
|
||||
image_filenames = convert_pptx_to_images_direct(input_file, output_folder, True, dpi=300)
|
||||
|
||||
# Verify we got images
|
||||
if not image_filenames:
|
||||
print("Error: No images were generated from the PPTX")
|
||||
if not pdf_file:
|
||||
print("Error: Failed to convert PPTX to PDF")
|
||||
return False
|
||||
|
||||
print(f"Generated {len(image_filenames)} images for PPTX")
|
||||
print(f"PPTX successfully converted to PDF: {pdf_file}")
|
||||
|
||||
# Update playlist with generated images in sequential order
|
||||
# Step 2: Use the same PDF to images workflow as direct PDF uploads
|
||||
# Convert PDF to JPG images (300 DPI, same as PDF workflow)
|
||||
image_filenames = convert_pdf_to_images(pdf_file, output_folder, delete_pdf=True, dpi=300)
|
||||
|
||||
if not image_filenames:
|
||||
print("Error: Failed to convert PDF to images")
|
||||
return False
|
||||
|
||||
print(f"Generated {len(image_filenames)} JPG images from PPTX → PDF")
|
||||
|
||||
# Step 3: Delete the original PPTX file after successful conversion
|
||||
if os.path.exists(input_file):
|
||||
os.remove(input_file)
|
||||
print(f"Original PPTX file deleted: {input_file}")
|
||||
|
||||
# Step 4: Update playlist with generated images in sequential order
|
||||
success = update_playlist_with_files(image_filenames, duration, target_type, target_id)
|
||||
|
||||
print(f"=== PPTX Processing Complete ===")
|
||||
print(f"Successfully processed {len(image_filenames)} slides")
|
||||
|
||||
if success:
|
||||
print(f"Successfully processed PPTX: {len(image_filenames)} images added to playlist")
|
||||
return success
|
||||
|
||||
except Exception as e:
|
||||
@@ -405,8 +376,20 @@ def process_uploaded_files(app, files, media_type, duration, target_type, target
|
||||
try:
|
||||
# Generate a secure filename and save the file
|
||||
filename = secure_filename(file.filename)
|
||||
file_path = os.path.join(app.config['UPLOAD_FOLDER'], filename)
|
||||
|
||||
# Ensure we use absolute path for upload folder
|
||||
upload_folder = app.config['UPLOAD_FOLDER']
|
||||
if not os.path.isabs(upload_folder):
|
||||
upload_folder = os.path.abspath(upload_folder)
|
||||
|
||||
# Ensure upload folder exists
|
||||
if not os.path.exists(upload_folder):
|
||||
os.makedirs(upload_folder, exist_ok=True)
|
||||
print(f"Created upload folder: {upload_folder}")
|
||||
|
||||
file_path = os.path.join(upload_folder, filename)
|
||||
file.save(file_path)
|
||||
print(f"File saved to: {file_path}")
|
||||
|
||||
print(f"Processing file: {filename}, Media Type: {media_type}")
|
||||
result = {'filename': filename, 'success': True, 'message': ''}
|
||||
@@ -432,7 +415,7 @@ def process_uploaded_files(app, files, media_type, duration, target_type, target
|
||||
player.playlist_version += 1
|
||||
|
||||
db.session.commit()
|
||||
# Start background conversion
|
||||
# Start background conversion using absolute path
|
||||
import threading
|
||||
threading.Thread(target=convert_video_and_update_playlist,
|
||||
args=(app, file_path, filename, target_type, target_id, duration)).start()
|
||||
@@ -440,8 +423,8 @@ def process_uploaded_files(app, files, media_type, duration, target_type, target
|
||||
log_upload('video', filename, target_type, target_id)
|
||||
|
||||
elif media_type == 'pdf':
|
||||
# For PDFs, convert to images and update playlist
|
||||
success = process_pdf(file_path, app.config['UPLOAD_FOLDER'],
|
||||
# For PDFs, convert to images and update playlist using absolute path
|
||||
success = process_pdf(file_path, upload_folder,
|
||||
duration, target_type, target_id)
|
||||
if success:
|
||||
result['message'] = f"PDF {filename} processed successfully"
|
||||
@@ -451,8 +434,8 @@ def process_uploaded_files(app, files, media_type, duration, target_type, target
|
||||
result['message'] = f"Error processing PDF file: {filename}"
|
||||
|
||||
elif media_type == 'ppt':
|
||||
# For PPT/PPTX, convert to PDF, then to images, and update playlist
|
||||
success = process_pptx_improved(file_path, app.config['UPLOAD_FOLDER'],
|
||||
# For PPT/PPTX, convert to PDF, then to images, and update playlist using absolute path
|
||||
success = process_pptx(file_path, upload_folder,
|
||||
duration, target_type, target_id)
|
||||
if success:
|
||||
result['message'] = f"PowerPoint {filename} processed successfully"
|
||||
64
cleanup-docker.sh
Executable file
64
cleanup-docker.sh
Executable file
@@ -0,0 +1,64 @@
|
||||
#!/bin/bash
|
||||
# DigiServer Docker Cleanup Script
|
||||
# Version: 1.1.0
|
||||
|
||||
set -e
|
||||
|
||||
echo "🧹 DigiServer Docker Cleanup"
|
||||
echo "============================"
|
||||
|
||||
# Colors for output
|
||||
RED='\033[0;31m'
|
||||
GREEN='\033[0;32m'
|
||||
YELLOW='\033[1;33m'
|
||||
BLUE='\033[0;34m'
|
||||
NC='\033[0m' # No Color
|
||||
|
||||
# Function to print colored output
|
||||
print_status() {
|
||||
echo -e "${BLUE}[INFO]${NC} $1"
|
||||
}
|
||||
|
||||
print_success() {
|
||||
echo -e "${GREEN}[SUCCESS]${NC} $1"
|
||||
}
|
||||
|
||||
print_warning() {
|
||||
echo -e "${YELLOW}[WARNING]${NC} $1"
|
||||
}
|
||||
|
||||
print_error() {
|
||||
echo -e "${RED}[ERROR]${NC} $1"
|
||||
}
|
||||
|
||||
# Confirm cleanup
|
||||
print_warning "This will stop and remove all DigiServer containers and images."
|
||||
print_warning "Your data in the ./data directory will be preserved."
|
||||
echo ""
|
||||
read -p "Are you sure you want to continue? (y/N): " -n 1 -r
|
||||
echo
|
||||
if [[ ! $REPLY =~ ^[Yy]$ ]]; then
|
||||
print_status "Cleanup cancelled."
|
||||
exit 0
|
||||
fi
|
||||
|
||||
# Stop and remove containers
|
||||
print_status "Stopping DigiServer containers..."
|
||||
docker compose down
|
||||
|
||||
# Remove DigiServer images
|
||||
print_status "Removing DigiServer images..."
|
||||
docker rmi digiserver:latest 2>/dev/null || print_warning "DigiServer image not found"
|
||||
|
||||
# Clean up unused Docker resources
|
||||
print_status "Cleaning up unused Docker resources..."
|
||||
docker system prune -f
|
||||
|
||||
# Clean up development cache files
|
||||
print_status "Cleaning up development cache files..."
|
||||
find ./app -name "__pycache__" -type d -exec rm -rf {} + 2>/dev/null || true
|
||||
find ./app -name "*.pyc" -delete 2>/dev/null || true
|
||||
|
||||
print_success "Cleanup completed!"
|
||||
print_status "Data directory preserved at: ./data"
|
||||
print_status "To redeploy, run: ./deploy-docker.sh"
|
||||
109
deploy-docker.sh
Executable file
109
deploy-docker.sh
Executable file
@@ -0,0 +1,109 @@
|
||||
#!/bin/bash
|
||||
# DigiServer Docker Deployment Script
|
||||
# Version: 1.1.0
|
||||
|
||||
set -e
|
||||
|
||||
echo "🚀 DigiServer Docker Deployment"
|
||||
echo "================================"
|
||||
|
||||
# Colors for output
|
||||
RED='\033[0;31m'
|
||||
GREEN='\033[0;32m'
|
||||
YELLOW='\033[1;33m'
|
||||
BLUE='\033[0;34m'
|
||||
NC='\033[0m' # No Color
|
||||
|
||||
# Function to print colored output
|
||||
print_status() {
|
||||
echo -e "${BLUE}[INFO]${NC} $1"
|
||||
}
|
||||
|
||||
print_success() {
|
||||
echo -e "${GREEN}[SUCCESS]${NC} $1"
|
||||
}
|
||||
|
||||
print_warning() {
|
||||
echo -e "${YELLOW}[WARNING]${NC} $1"
|
||||
}
|
||||
|
||||
print_error() {
|
||||
echo -e "${RED}[ERROR]${NC} $1"
|
||||
}
|
||||
|
||||
# Check if Docker is running
|
||||
if ! docker info >/dev/null 2>&1; then
|
||||
print_error "Docker is not running. Please start Docker and try again."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
print_status "Docker is running ✓"
|
||||
|
||||
# Check if docker compose is available
|
||||
if ! docker compose version >/dev/null 2>&1; then
|
||||
print_error "docker compose is not available. Please install Docker Compose and try again."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
print_status "docker compose is available ✓"
|
||||
|
||||
# Stop existing containers if running
|
||||
print_status "Stopping existing containers..."
|
||||
docker compose down 2>/dev/null || true
|
||||
|
||||
# Remove old images (optional)
|
||||
read -p "Do you want to remove old DigiServer images? (y/N): " -n 1 -r
|
||||
echo
|
||||
if [[ $REPLY =~ ^[Yy]$ ]]; then
|
||||
print_status "Removing old images..."
|
||||
docker image prune -f --filter label=app=digiserver 2>/dev/null || true
|
||||
docker rmi digiserver:latest 2>/dev/null || true
|
||||
fi
|
||||
|
||||
# Create data directories if they don't exist
|
||||
print_status "Creating data directories..."
|
||||
mkdir -p data/instance data/uploads data/resurse
|
||||
|
||||
# Build the Docker image
|
||||
print_status "Building DigiServer Docker image..."
|
||||
docker compose build
|
||||
|
||||
# Check if build was successful
|
||||
if [ $? -eq 0 ]; then
|
||||
print_success "Docker image built successfully!"
|
||||
else
|
||||
print_error "Docker build failed!"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Start the containers
|
||||
print_status "Starting DigiServer containers..."
|
||||
docker compose up -d
|
||||
|
||||
# Wait a moment for containers to start
|
||||
sleep 10
|
||||
|
||||
# Check if containers are running
|
||||
if docker compose ps | grep -q "Up"; then
|
||||
print_success "DigiServer is now running!"
|
||||
echo ""
|
||||
echo "🌐 Access your DigiServer at: http://localhost:8880"
|
||||
echo "📊 Admin Panel: http://localhost:8880/admin"
|
||||
echo ""
|
||||
echo "Default credentials:"
|
||||
echo "Username: admin"
|
||||
echo "Password: Initial01!"
|
||||
echo ""
|
||||
print_warning "Please change the default password after first login!"
|
||||
echo ""
|
||||
echo "📝 To view logs: docker compose logs -f"
|
||||
echo "🛑 To stop: docker compose down"
|
||||
echo "📊 To check status: docker compose ps"
|
||||
else
|
||||
print_error "Failed to start DigiServer containers!"
|
||||
echo ""
|
||||
echo "Check logs with: docker compose logs"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
print_success "Deployment completed successfully! 🎉"
|
||||
32
docker-compose.dev.yml
Normal file
32
docker-compose.dev.yml
Normal file
@@ -0,0 +1,32 @@
|
||||
# Development Docker Compose Configuration
|
||||
# Use this for development with hot reloading
|
||||
|
||||
services:
|
||||
digiserver-dev:
|
||||
build: .
|
||||
image: digiserver:dev
|
||||
container_name: digiserver-dev
|
||||
ports:
|
||||
- "5000:5000"
|
||||
environment:
|
||||
- FLASK_APP=app.py
|
||||
- FLASK_RUN_HOST=0.0.0.0
|
||||
- FLASK_ENV=development
|
||||
- FLASK_DEBUG=1
|
||||
- ADMIN_USER=admin
|
||||
- ADMIN_PASSWORD=Initial01!
|
||||
- SECRET_KEY=Ma_Duc_Dupa_Merele_Lui_Ana
|
||||
volumes:
|
||||
# Mount app code for hot reloading
|
||||
- ./app:/app
|
||||
# Persistent data volumes
|
||||
- ./data/instance:/app/instance
|
||||
- ./data/uploads:/app/static/uploads
|
||||
- ./data/resurse:/app/static/resurse
|
||||
restart: unless-stopped
|
||||
networks:
|
||||
- digiserver-dev-network
|
||||
|
||||
networks:
|
||||
digiserver-dev-network:
|
||||
driver: bridge
|
||||
@@ -1,33 +1,35 @@
|
||||
version: "3.8"
|
||||
# DigiServer - Digital Signage Management Platform
|
||||
# Version: 1.1.0
|
||||
# Build Date: 2025-06-29
|
||||
|
||||
services:
|
||||
web:
|
||||
digiserver:
|
||||
build: .
|
||||
image: digi-server:latest
|
||||
image: digiserver:latest
|
||||
container_name: digiserver
|
||||
ports:
|
||||
- "80:5000"
|
||||
- "8880:5000"
|
||||
environment:
|
||||
- FLASK_APP
|
||||
- FLASK_RUN_HOST
|
||||
- ADMIN_USER
|
||||
- ADMIN_PASSWORD
|
||||
- SECRET_KEY
|
||||
env_file:
|
||||
- .env
|
||||
- FLASK_APP=app.py
|
||||
- FLASK_RUN_HOST=0.0.0.0
|
||||
- DEFAULT_USER=admin
|
||||
- DEFAULT_PASSWORD=Initial01!
|
||||
- SECRET_KEY=Ma_Duc_Dupa_Merele_Lui_Ana
|
||||
volumes:
|
||||
- /home/pi/Desktop/digi-server/instance:/app/instance
|
||||
- /home/pi/Desktop/digi-server/uploads:/app/static/uploads
|
||||
# Persistent data volumes
|
||||
- ./data/instance:/app/instance
|
||||
- ./data/uploads:/app/static/uploads
|
||||
- ./data/resurse:/app/static/resurse
|
||||
restart: unless-stopped
|
||||
deploy:
|
||||
resources:
|
||||
limits:
|
||||
memory: 512M
|
||||
cpus: '1.0'
|
||||
reservations:
|
||||
memory: 128M
|
||||
cpus: '0.25'
|
||||
healthcheck:
|
||||
test: ["CMD", "curl", "-f", "http://localhost:5000/"]
|
||||
interval: 30s
|
||||
timeout: 10s
|
||||
retries: 3
|
||||
start_period: 40s
|
||||
networks:
|
||||
- digiserver-network
|
||||
|
||||
networks:
|
||||
digiserver-network:
|
||||
driver: bridge
|
||||
|
||||
@@ -1,25 +0,0 @@
|
||||
python3 -m venv digiscreen
|
||||
|
||||
source digiscreen/bin/activate
|
||||
|
||||
pip install flask sqlalchemy flask-sqlalchemy
|
||||
|
||||
pip install flask-login flask-bcrypt
|
||||
|
||||
python3 setup.py sdist
|
||||
|
||||
python3 setup.py bdist_wheel flask
|
||||
|
||||
|
||||
for installing all the requirements
|
||||
pip install -r requirements.txt
|
||||
|
||||
|
||||
|
||||
sudo apt-get update
|
||||
sudo apt-get install -y \
|
||||
ffmpeg \
|
||||
libpoppler-cpp-dev \
|
||||
poppler-utils \
|
||||
libreoffice \
|
||||
libmagic1
|
||||
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
|
Before Width: | Height: | Size: 361 KiB |
Binary file not shown.
Binary file not shown.
|
Before Width: | Height: | Size: 794 KiB |
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Reference in New Issue
Block a user