Compare commits
30 Commits
main
...
02-2025-06
Author | SHA1 | Date | |
---|---|---|---|
e5a274d191 | |||
c97d025acb | |||
d8d330fd9d | |||
fa951c3141 | |||
d818ec7d73 | |||
b5f2b53262 | |||
bcbac5c8b3 | |||
e5333d8410 | |||
77f31cd694 | |||
92fd390866 | |||
1b9ad03400 | |||
92b7c41f61 | |||
aec2d3b0aa | |||
30edf8d272 | |||
eb4f5752ee | |||
9170198c6c | |||
a4ef775731 | |||
d010ae9b74 | |||
17e0f2f03d | |||
b08574efe5 | |||
dc3e67e17b | |||
343bfbf1de | |||
fed54259ca | |||
aedface2a9 | |||
3804897e2b | |||
0c9186e57e | |||
829e8c3978 | |||
08791d2fed | |||
05ad940079 | |||
80e724cf86 |
71
.dockerignore
Normal file
71
.dockerignore
Normal file
@ -0,0 +1,71 @@
|
||||
# Git
|
||||
.git
|
||||
.gitignore
|
||||
|
||||
# Python
|
||||
__pycache__
|
||||
*.pyc
|
||||
*.pyo
|
||||
*.pyd
|
||||
.Python
|
||||
venv/
|
||||
env/
|
||||
.venv
|
||||
pip-log.txt
|
||||
pip-delete-this-directory.txt
|
||||
.tox/
|
||||
.coverage
|
||||
.coverage.*
|
||||
.cache
|
||||
*.egg-info/
|
||||
.pytest_cache/
|
||||
|
||||
# Node
|
||||
node_modules/
|
||||
npm-debug.log*
|
||||
yarn-debug.log*
|
||||
yarn-error.log*
|
||||
|
||||
# IDE
|
||||
.vscode/
|
||||
.idea/
|
||||
*.swp
|
||||
*.swo
|
||||
*~
|
||||
|
||||
# OS
|
||||
.DS_Store
|
||||
.DS_Store?
|
||||
._*
|
||||
.Spotlight-V100
|
||||
.Trashes
|
||||
ehthumbs.db
|
||||
Thumbs.db
|
||||
|
||||
# Project specific
|
||||
logs/
|
||||
*.log
|
||||
.env
|
||||
.env.*
|
||||
!.env.production
|
||||
*.db
|
||||
*.sqlite
|
||||
/tmp
|
||||
/temp
|
||||
test_*.py
|
||||
tests/
|
||||
|
||||
# Documentation
|
||||
*.md
|
||||
!README.md
|
||||
docs/
|
||||
|
||||
# CI/CD
|
||||
.github/
|
||||
.gitlab-ci.yml
|
||||
.travis.yml
|
||||
|
||||
# Development files
|
||||
deploy.sh
|
||||
Makefile
|
||||
docker-compose.override.yml
|
73
.env.example
Normal file
73
.env.example
Normal file
@ -0,0 +1,73 @@
|
||||
# Talk2Me Environment Configuration
|
||||
# Copy this file to .env and fill in your values
|
||||
|
||||
# Flask Configuration
|
||||
FLASK_ENV=development
|
||||
FLASK_SECRET_KEY=your-secret-key-here-change-in-production
|
||||
FLASK_DEBUG=False
|
||||
|
||||
# Server Configuration
|
||||
HOST=0.0.0.0
|
||||
PORT=5005
|
||||
|
||||
# Database Configuration
|
||||
DATABASE_URL=postgresql://user:password@localhost:5432/talk2me
|
||||
REDIS_URL=redis://localhost:6379/0
|
||||
|
||||
# Ollama Configuration
|
||||
OLLAMA_BASE_URL=http://localhost:11434
|
||||
OLLAMA_MODEL=gemma2:9b
|
||||
OLLAMA_LARGE_MODEL=gemma3:27b
|
||||
|
||||
# TTS Configuration
|
||||
TTS_SERVER_URL=http://localhost:8000
|
||||
TTS_API_KEY=your-tts-api-key-here
|
||||
|
||||
# Security Configuration
|
||||
JWT_SECRET_KEY=your-jwt-secret-key-here
|
||||
JWT_ACCESS_TOKEN_EXPIRES=3600
|
||||
JWT_REFRESH_TOKEN_EXPIRES=2592000
|
||||
|
||||
# Admin Configuration
|
||||
ADMIN_USERNAME=admin
|
||||
ADMIN_PASSWORD=change-this-password
|
||||
ADMIN_EMAIL=admin@example.com
|
||||
|
||||
# Rate Limiting
|
||||
RATE_LIMIT_PER_MINUTE=60
|
||||
RATE_LIMIT_PER_HOUR=1000
|
||||
|
||||
# Session Configuration
|
||||
SESSION_LIFETIME=86400
|
||||
SESSION_CLEANUP_INTERVAL=3600
|
||||
|
||||
# Logging
|
||||
LOG_LEVEL=INFO
|
||||
LOG_FORMAT=json
|
||||
|
||||
# CORS Configuration
|
||||
CORS_ORIGINS=http://localhost:3000,http://localhost:5005
|
||||
|
||||
# Feature Flags
|
||||
ENABLE_ANALYTICS=true
|
||||
ENABLE_RATE_LIMITING=true
|
||||
ENABLE_SESSION_MANAGEMENT=true
|
||||
ENABLE_ERROR_TRACKING=true
|
||||
|
||||
# Performance Settings
|
||||
MAX_CONTENT_LENGTH=16777216
|
||||
REQUEST_TIMEOUT=300
|
||||
WHISPER_MODEL=base
|
||||
WHISPER_DEVICE=auto
|
||||
|
||||
# Email Configuration (Optional)
|
||||
SMTP_HOST=smtp.gmail.com
|
||||
SMTP_PORT=587
|
||||
SMTP_USER=your-email@example.com
|
||||
SMTP_PASSWORD=your-email-password
|
||||
SMTP_FROM=noreply@example.com
|
||||
|
||||
# External Services (Optional)
|
||||
SENTRY_DSN=
|
||||
DATADOG_API_KEY=
|
||||
NEWRELIC_LICENSE_KEY=
|
54
.env.template
Normal file
54
.env.template
Normal file
@ -0,0 +1,54 @@
|
||||
# Talk2Me Environment Configuration Template
|
||||
# Copy this file to .env and update with your values
|
||||
|
||||
# Flask Configuration
|
||||
FLASK_ENV=production
|
||||
SECRET_KEY=your-secret-key-here-change-this
|
||||
|
||||
# Security Settings for HTTPS/Reverse Proxy
|
||||
SESSION_COOKIE_SECURE=true
|
||||
SESSION_COOKIE_SAMESITE=Lax
|
||||
PREFERRED_URL_SCHEME=https
|
||||
|
||||
# TTS Server Configuration
|
||||
TTS_SERVER_URL=http://localhost:5050/v1/audio/speech
|
||||
TTS_API_KEY=your-tts-api-key-here
|
||||
|
||||
# Whisper Configuration
|
||||
WHISPER_MODEL_SIZE=base
|
||||
WHISPER_DEVICE=auto
|
||||
|
||||
# Ollama Configuration
|
||||
OLLAMA_HOST=http://localhost:11434
|
||||
OLLAMA_MODEL=gemma3:27b
|
||||
|
||||
# Admin Configuration
|
||||
ADMIN_TOKEN=your-admin-token-here-change-this
|
||||
|
||||
# CORS Configuration (comma-separated)
|
||||
CORS_ORIGINS=https://talk2me.dr74.net,http://localhost:5000
|
||||
ADMIN_CORS_ORIGINS=https://talk2me.dr74.net
|
||||
|
||||
# Rate Limiting
|
||||
RATE_LIMIT_ENABLED=true
|
||||
RATE_LIMIT_STORAGE_URL=memory://
|
||||
|
||||
# Feature Flags
|
||||
ENABLE_PUSH_NOTIFICATIONS=true
|
||||
ENABLE_OFFLINE_MODE=true
|
||||
ENABLE_STREAMING=true
|
||||
ENABLE_MULTI_SPEAKER=true
|
||||
|
||||
# Logging
|
||||
LOG_LEVEL=INFO
|
||||
LOG_FILE=logs/talk2me.log
|
||||
|
||||
# Upload Configuration
|
||||
UPLOAD_FOLDER=/tmp/talk2me_uploads
|
||||
MAX_CONTENT_LENGTH=52428800
|
||||
MAX_AUDIO_SIZE=26214400
|
||||
MAX_JSON_SIZE=1048576
|
||||
|
||||
# Worker Configuration (for Gunicorn)
|
||||
WORKER_CONNECTIONS=1000
|
||||
WORKER_TIMEOUT=120
|
80
.gitignore
vendored
80
.gitignore
vendored
@ -1 +1,81 @@
|
||||
# Python
|
||||
__pycache__/
|
||||
*.py[cod]
|
||||
*$py.class
|
||||
*.so
|
||||
.Python
|
||||
venv/
|
||||
env/
|
||||
ENV/
|
||||
.venv
|
||||
.env
|
||||
|
||||
# Flask
|
||||
instance/
|
||||
.webassets-cache
|
||||
|
||||
# IDE
|
||||
.vscode/
|
||||
.idea/
|
||||
*.swp
|
||||
*.swo
|
||||
*~
|
||||
|
||||
# OS
|
||||
.DS_Store
|
||||
.DS_Store?
|
||||
._*
|
||||
.Spotlight-V100
|
||||
.Trashes
|
||||
ehthumbs.db
|
||||
Thumbs.db
|
||||
|
||||
# Node.js
|
||||
node_modules/
|
||||
npm-debug.log*
|
||||
yarn-debug.log*
|
||||
yarn-error.log*
|
||||
|
||||
# TypeScript
|
||||
static/js/dist/
|
||||
*.tsbuildinfo
|
||||
|
||||
# Temporary files
|
||||
*.log
|
||||
*.tmp
|
||||
temp/
|
||||
tmp/
|
||||
|
||||
# Audio files (for testing)
|
||||
*.mp3
|
||||
*.wav
|
||||
*.ogg
|
||||
|
||||
# Local environment
|
||||
.env.local
|
||||
.env.*.local
|
||||
.env.production
|
||||
.env.development
|
||||
.env.staging
|
||||
|
||||
# VAPID keys
|
||||
vapid_private.pem
|
||||
vapid_public.pem
|
||||
|
||||
# Secrets management
|
||||
.secrets.json
|
||||
.master_key
|
||||
secrets.db
|
||||
*.key
|
||||
|
||||
# Test files
|
||||
test_*.py
|
||||
*_test_output.*
|
||||
test-*.html
|
||||
*-debug-script.py
|
||||
|
||||
# Claude IDE
|
||||
.claude/
|
||||
|
||||
# Standalone compiled JS (use dist/ instead)
|
||||
static/js/app.js
|
||||
|
221
ADMIN_DASHBOARD.md
Normal file
221
ADMIN_DASHBOARD.md
Normal file
@ -0,0 +1,221 @@
|
||||
# Talk2Me Admin Analytics Dashboard
|
||||
|
||||
A comprehensive analytics dashboard for monitoring and managing the Talk2Me application.
|
||||
|
||||
## Features
|
||||
|
||||
### Real-time Monitoring
|
||||
- **Request Volume**: Track requests per minute, hour, and day
|
||||
- **Active Sessions**: Monitor current active user sessions
|
||||
- **Error Rates**: Real-time error tracking and analysis
|
||||
- **System Health**: Monitor Redis, PostgreSQL, and ML services status
|
||||
|
||||
### Analytics & Insights
|
||||
- **Translation & Transcription Metrics**: Usage statistics by operation type
|
||||
- **Language Pair Analysis**: Most popular translation combinations
|
||||
- **Response Time Monitoring**: Track performance across all operations
|
||||
- **Cache Performance**: Monitor cache hit rates for optimization
|
||||
|
||||
### Performance Metrics
|
||||
- **Response Time Percentiles**: P95 and P99 latency tracking
|
||||
- **Throughput Analysis**: Requests per minute visualization
|
||||
- **Slow Request Detection**: Identify and analyze performance bottlenecks
|
||||
- **Resource Usage**: Memory and GPU utilization tracking
|
||||
|
||||
### Data Management
|
||||
- **Export Capabilities**: Download analytics data in JSON format
|
||||
- **Historical Data**: View trends over time (daily, weekly, monthly)
|
||||
- **Error Logs**: Detailed error tracking with stack traces
|
||||
- **Session Management**: Track and manage user sessions
|
||||
|
||||
## Setup
|
||||
|
||||
### 1. Database Setup
|
||||
|
||||
Initialize the analytics database tables:
|
||||
|
||||
```bash
|
||||
python init_analytics_db.py
|
||||
```
|
||||
|
||||
This creates the following tables:
|
||||
- `error_logs`: Detailed error tracking
|
||||
- `request_logs`: Request-level analytics
|
||||
- `translation_logs`: Translation operation metrics
|
||||
- `transcription_logs`: Transcription operation metrics
|
||||
- `tts_logs`: Text-to-speech operation metrics
|
||||
- `daily_stats`: Aggregated daily statistics
|
||||
|
||||
### 2. Configuration
|
||||
|
||||
Set the following environment variables:
|
||||
|
||||
```bash
|
||||
# Admin access token (required)
|
||||
export ADMIN_TOKEN="your-secure-admin-token"
|
||||
|
||||
# Database configuration
|
||||
export DATABASE_URL="postgresql://user:password@localhost/talk2me"
|
||||
|
||||
# Redis configuration
|
||||
export REDIS_URL="redis://localhost:6379/0"
|
||||
```
|
||||
|
||||
### 3. Access the Dashboard
|
||||
|
||||
1. Navigate to: `http://your-domain/admin`
|
||||
2. Enter your admin token
|
||||
3. Access the analytics dashboard
|
||||
|
||||
## Dashboard Sections
|
||||
|
||||
### Overview Cards
|
||||
- Total Requests (all-time and today)
|
||||
- Active Sessions (real-time)
|
||||
- Error Rate (24-hour percentage)
|
||||
- Cache Hit Rate (performance metric)
|
||||
|
||||
### Charts & Visualizations
|
||||
|
||||
#### Request Volume Chart
|
||||
- Toggle between minute, hour, and day views
|
||||
- Real-time updates every 5 seconds
|
||||
- Historical data for trend analysis
|
||||
|
||||
#### Language Pairs Donut Chart
|
||||
- Top 6 most used language combinations
|
||||
- Visual breakdown of translation patterns
|
||||
|
||||
#### Operations Bar Chart
|
||||
- Daily translation and transcription counts
|
||||
- 7-day historical view
|
||||
|
||||
#### Response Time Line Chart
|
||||
- Average, P95, and P99 response times
|
||||
- Broken down by operation type
|
||||
|
||||
#### Error Analysis
|
||||
- Error type distribution pie chart
|
||||
- Recent errors list with details
|
||||
- Timeline of error occurrences
|
||||
|
||||
### Performance Table
|
||||
- Detailed metrics for each operation type
|
||||
- Average response times
|
||||
- 95th and 99th percentile latencies
|
||||
|
||||
## Real-time Updates
|
||||
|
||||
The dashboard uses Server-Sent Events (SSE) for real-time updates:
|
||||
- Automatic refresh every 5 seconds
|
||||
- Connection status indicator
|
||||
- Automatic reconnection on disconnect
|
||||
|
||||
## Data Export
|
||||
|
||||
Export analytics data for external analysis:
|
||||
|
||||
1. Click "Export Data" in the navigation
|
||||
2. Choose data type:
|
||||
- `requests`: Request and operation counts
|
||||
- `errors`: Error logs and details
|
||||
- `performance`: Response time metrics
|
||||
- `all`: Complete data export
|
||||
|
||||
## API Endpoints
|
||||
|
||||
The admin dashboard provides the following API endpoints:
|
||||
|
||||
### Authentication Required
|
||||
All endpoints require the `X-Admin-Token` header.
|
||||
|
||||
### Available Endpoints
|
||||
|
||||
#### Overview Stats
|
||||
```
|
||||
GET /admin/api/stats/overview
|
||||
```
|
||||
Returns overall system statistics
|
||||
|
||||
#### Request Statistics
|
||||
```
|
||||
GET /admin/api/stats/requests/{timeframe}
|
||||
```
|
||||
Timeframes: `minute`, `hour`, `day`
|
||||
|
||||
#### Operation Statistics
|
||||
```
|
||||
GET /admin/api/stats/operations
|
||||
```
|
||||
Translation and transcription metrics
|
||||
|
||||
#### Error Statistics
|
||||
```
|
||||
GET /admin/api/stats/errors
|
||||
```
|
||||
Error types, timeline, and recent errors
|
||||
|
||||
#### Performance Statistics
|
||||
```
|
||||
GET /admin/api/stats/performance
|
||||
```
|
||||
Response times and throughput metrics
|
||||
|
||||
#### Data Export
|
||||
```
|
||||
GET /admin/api/export/{data_type}
|
||||
```
|
||||
Data types: `requests`, `errors`, `performance`, `all`
|
||||
|
||||
#### Real-time Updates
|
||||
```
|
||||
GET /admin/api/stream/updates
|
||||
```
|
||||
Server-Sent Events stream for real-time updates
|
||||
|
||||
## Mobile Optimization
|
||||
|
||||
The dashboard is fully responsive and optimized for mobile devices:
|
||||
- Touch-friendly controls
|
||||
- Responsive charts that adapt to screen size
|
||||
- Collapsible navigation for small screens
|
||||
- Optimized data tables for mobile viewing
|
||||
|
||||
## Security
|
||||
|
||||
- Admin token authentication required
|
||||
- Session-based authentication after login
|
||||
- Separate CORS configuration for admin endpoints
|
||||
- All sensitive data masked in exports
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Dashboard Not Loading
|
||||
1. Check Redis and PostgreSQL connections
|
||||
2. Verify admin token is set correctly
|
||||
3. Check browser console for JavaScript errors
|
||||
|
||||
### Missing Data
|
||||
1. Ensure analytics middleware is initialized
|
||||
2. Check database tables are created
|
||||
3. Verify Redis is running and accessible
|
||||
|
||||
### Real-time Updates Not Working
|
||||
1. Check SSE support in your reverse proxy
|
||||
2. Ensure `X-Accel-Buffering: no` header is set
|
||||
3. Verify firewall allows SSE connections
|
||||
|
||||
## Performance Considerations
|
||||
|
||||
- Charts limited to reasonable data points for performance
|
||||
- Automatic data aggregation for historical views
|
||||
- Efficient database queries with proper indexing
|
||||
- Client-side caching for static data
|
||||
|
||||
## Future Enhancements
|
||||
|
||||
- WebSocket support for lower latency updates
|
||||
- Customizable dashboards and widgets
|
||||
- Alert configuration for thresholds
|
||||
- Integration with external monitoring tools
|
||||
- Machine learning for anomaly detection
|
315
AUTHENTICATION.md
Normal file
315
AUTHENTICATION.md
Normal file
@ -0,0 +1,315 @@
|
||||
# Talk2Me Authentication System
|
||||
|
||||
This document describes the comprehensive user authentication and authorization system implemented for the Talk2Me application.
|
||||
|
||||
## Overview
|
||||
|
||||
The authentication system provides:
|
||||
- User account management with roles (admin, user)
|
||||
- JWT-based API authentication
|
||||
- Session management for web interface
|
||||
- API key authentication for programmatic access
|
||||
- User-specific rate limiting
|
||||
- Admin dashboard for user management
|
||||
- Secure password hashing with bcrypt
|
||||
|
||||
## Features
|
||||
|
||||
### 1. User Management
|
||||
|
||||
#### User Account Model
|
||||
- **Email & Username**: Unique identifiers for each user
|
||||
- **Password**: Securely hashed using bcrypt
|
||||
- **API Key**: Unique key for each user (format: `tk_<random_string>`)
|
||||
- **Roles**: `admin` or `user`
|
||||
- **Account Status**: Active, verified, suspended
|
||||
- **Rate Limits**: Configurable per-user limits
|
||||
- **Usage Tracking**: Tracks requests, translations, transcriptions, and TTS usage
|
||||
|
||||
#### Admin Features
|
||||
- Create, update, delete users
|
||||
- Suspend/unsuspend accounts
|
||||
- Reset passwords
|
||||
- Manage user permissions
|
||||
- View login history
|
||||
- Monitor active sessions
|
||||
- Bulk operations
|
||||
|
||||
### 2. Authentication Methods
|
||||
|
||||
#### JWT Authentication
|
||||
- Access tokens (1 hour expiration)
|
||||
- Refresh tokens (30 days expiration)
|
||||
- Token blacklisting for revocation
|
||||
- Secure token storage
|
||||
|
||||
#### API Key Authentication
|
||||
- Bearer token in `X-API-Key` header
|
||||
- Query parameter fallback: `?api_key=tk_xxx`
|
||||
- Per-key rate limiting
|
||||
|
||||
#### Session Management
|
||||
- Track active sessions per user
|
||||
- Session expiration handling
|
||||
- Multi-device support
|
||||
- Session revocation
|
||||
|
||||
### 3. Security Features
|
||||
|
||||
#### Password Security
|
||||
- Bcrypt hashing with salt
|
||||
- Minimum 8 character requirement
|
||||
- Password change tracking
|
||||
- Failed login attempt tracking
|
||||
- Account lockout after 5 failed attempts (30 minutes)
|
||||
|
||||
#### Rate Limiting
|
||||
- User-specific limits (per minute/hour/day)
|
||||
- IP-based fallback for unauthenticated requests
|
||||
- Admin users get 10x higher limits
|
||||
- Endpoint-specific overrides
|
||||
|
||||
#### Audit Trail
|
||||
- Login history with IP and user agent
|
||||
- Success/failure tracking
|
||||
- Suspicious activity flagging
|
||||
- Security event logging
|
||||
|
||||
## Database Schema
|
||||
|
||||
### Users Table
|
||||
```sql
|
||||
- id (UUID, primary key)
|
||||
- email (unique)
|
||||
- username (unique)
|
||||
- password_hash
|
||||
- api_key (unique)
|
||||
- role (admin/user)
|
||||
- is_active, is_verified, is_suspended
|
||||
- rate limits (per_minute, per_hour, per_day)
|
||||
- usage stats (total_requests, translations, etc.)
|
||||
- timestamps (created_at, updated_at, last_login_at)
|
||||
```
|
||||
|
||||
### Login History Table
|
||||
```sql
|
||||
- id (UUID)
|
||||
- user_id (foreign key)
|
||||
- login_at, logout_at
|
||||
- login_method (password/api_key/jwt)
|
||||
- success (boolean)
|
||||
- ip_address, user_agent
|
||||
- session_id, jwt_jti
|
||||
```
|
||||
|
||||
### User Sessions Table
|
||||
```sql
|
||||
- id (UUID)
|
||||
- session_id (unique)
|
||||
- user_id (foreign key)
|
||||
- access_token_jti, refresh_token_jti
|
||||
- created_at, last_active_at, expires_at
|
||||
- ip_address, user_agent
|
||||
```
|
||||
|
||||
### Revoked Tokens Table
|
||||
```sql
|
||||
- id (UUID)
|
||||
- jti (unique, token ID)
|
||||
- token_type (access/refresh)
|
||||
- user_id
|
||||
- revoked_at, expires_at
|
||||
- reason
|
||||
```
|
||||
|
||||
## API Endpoints
|
||||
|
||||
### Authentication Endpoints
|
||||
|
||||
#### POST /api/auth/login
|
||||
Login with username/email and password.
|
||||
```json
|
||||
{
|
||||
"username": "user@example.com",
|
||||
"password": "password123"
|
||||
}
|
||||
```
|
||||
|
||||
Response:
|
||||
```json
|
||||
{
|
||||
"success": true,
|
||||
"user": { ... },
|
||||
"tokens": {
|
||||
"access_token": "eyJ...",
|
||||
"refresh_token": "eyJ...",
|
||||
"expires_in": 3600
|
||||
},
|
||||
"session_id": "uuid"
|
||||
}
|
||||
```
|
||||
|
||||
#### POST /api/auth/logout
|
||||
Logout and revoke current token.
|
||||
|
||||
#### POST /api/auth/refresh
|
||||
Refresh access token using refresh token.
|
||||
|
||||
#### GET /api/auth/profile
|
||||
Get current user profile.
|
||||
|
||||
#### PUT /api/auth/profile
|
||||
Update user profile (name, settings).
|
||||
|
||||
#### POST /api/auth/change-password
|
||||
Change user password.
|
||||
|
||||
#### POST /api/auth/regenerate-api-key
|
||||
Generate new API key.
|
||||
|
||||
### Admin User Management
|
||||
|
||||
#### GET /api/auth/admin/users
|
||||
List all users with filtering and pagination.
|
||||
|
||||
#### POST /api/auth/admin/users
|
||||
Create new user (admin only).
|
||||
|
||||
#### GET /api/auth/admin/users/:id
|
||||
Get user details with login history.
|
||||
|
||||
#### PUT /api/auth/admin/users/:id
|
||||
Update user details.
|
||||
|
||||
#### DELETE /api/auth/admin/users/:id
|
||||
Delete user account.
|
||||
|
||||
#### POST /api/auth/admin/users/:id/suspend
|
||||
Suspend user account.
|
||||
|
||||
#### POST /api/auth/admin/users/:id/reset-password
|
||||
Reset user password.
|
||||
|
||||
## Usage Examples
|
||||
|
||||
### Authenticating Requests
|
||||
|
||||
#### Using JWT Token
|
||||
```bash
|
||||
curl -H "Authorization: Bearer eyJ..." \
|
||||
https://api.talk2me.app/translate
|
||||
```
|
||||
|
||||
#### Using API Key
|
||||
```bash
|
||||
curl -H "X-API-Key: tk_your_api_key" \
|
||||
https://api.talk2me.app/translate
|
||||
```
|
||||
|
||||
### Python Client Example
|
||||
```python
|
||||
import requests
|
||||
|
||||
# Login and get token
|
||||
response = requests.post('https://api.talk2me.app/api/auth/login', json={
|
||||
'username': 'user@example.com',
|
||||
'password': 'password123'
|
||||
})
|
||||
tokens = response.json()['tokens']
|
||||
|
||||
# Use token for requests
|
||||
headers = {'Authorization': f"Bearer {tokens['access_token']}"}
|
||||
translation = requests.post(
|
||||
'https://api.talk2me.app/translate',
|
||||
headers=headers,
|
||||
json={'text': 'Hello', 'target_lang': 'Spanish'}
|
||||
)
|
||||
```
|
||||
|
||||
## Setup Instructions
|
||||
|
||||
### 1. Install Dependencies
|
||||
```bash
|
||||
pip install -r requirements.txt
|
||||
```
|
||||
|
||||
### 2. Initialize Database
|
||||
```bash
|
||||
python init_auth_db.py
|
||||
```
|
||||
|
||||
This will:
|
||||
- Create all database tables
|
||||
- Prompt you to create an admin user
|
||||
- Display the admin's API key
|
||||
|
||||
### 3. Configure Environment
|
||||
Add to your `.env` file:
|
||||
```env
|
||||
JWT_SECRET_KEY=your-secret-key-change-in-production
|
||||
DATABASE_URL=postgresql://user:pass@localhost/talk2me
|
||||
```
|
||||
|
||||
### 4. Run Migrations (if needed)
|
||||
```bash
|
||||
alembic upgrade head
|
||||
```
|
||||
|
||||
## Security Best Practices
|
||||
|
||||
1. **JWT Secret**: Use a strong, random secret key in production
|
||||
2. **HTTPS Only**: Always use HTTPS in production
|
||||
3. **Rate Limiting**: Configure appropriate limits per user role
|
||||
4. **Password Policy**: Enforce strong passwords
|
||||
5. **Session Timeout**: Configure appropriate session durations
|
||||
6. **Audit Logging**: Monitor login attempts and suspicious activity
|
||||
7. **API Key Rotation**: Encourage regular API key rotation
|
||||
8. **Database Security**: Use encrypted connections to database
|
||||
|
||||
## Admin Dashboard
|
||||
|
||||
Access the admin dashboard at `/admin/users` (requires admin login).
|
||||
|
||||
Features:
|
||||
- User list with search and filters
|
||||
- User details with usage statistics
|
||||
- Create/edit/delete users
|
||||
- Suspend/unsuspend accounts
|
||||
- View login history
|
||||
- Monitor active sessions
|
||||
- Bulk operations
|
||||
|
||||
## Rate Limiting
|
||||
|
||||
Default limits:
|
||||
- **Regular Users**: 30/min, 500/hour, 5000/day
|
||||
- **Admin Users**: 300/min, 5000/hour, 50000/day
|
||||
|
||||
Endpoint-specific limits are configured in `user_rate_limiter.py`.
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Common Issues
|
||||
|
||||
1. **"Token expired"**: Refresh token using `/api/auth/refresh`
|
||||
2. **"Account locked"**: Wait 30 minutes or contact admin
|
||||
3. **"Rate limit exceeded"**: Check your usage limits
|
||||
4. **"Invalid API key"**: Regenerate key in profile settings
|
||||
|
||||
### Debug Mode
|
||||
Enable debug logging:
|
||||
```python
|
||||
import logging
|
||||
logging.getLogger('auth').setLevel(logging.DEBUG)
|
||||
```
|
||||
|
||||
## Future Enhancements
|
||||
|
||||
- [ ] OAuth2 integration (Google, GitHub)
|
||||
- [ ] Two-factor authentication
|
||||
- [ ] Email verification workflow
|
||||
- [ ] Password reset via email
|
||||
- [ ] User groups and team management
|
||||
- [ ] Fine-grained permissions
|
||||
- [ ] API key scopes
|
||||
- [ ] Usage quotas and billing
|
302
DATABASE_INTEGRATION.md
Normal file
302
DATABASE_INTEGRATION.md
Normal file
@ -0,0 +1,302 @@
|
||||
# Database Integration Guide
|
||||
|
||||
This guide explains the Redis and PostgreSQL integration for the Talk2Me application.
|
||||
|
||||
## Overview
|
||||
|
||||
The Talk2Me application now uses:
|
||||
- **PostgreSQL**: For persistent storage of translations, transcriptions, user preferences, and analytics
|
||||
- **Redis**: For caching, session management, and rate limiting
|
||||
|
||||
## Architecture
|
||||
|
||||
### PostgreSQL Database Schema
|
||||
|
||||
1. **translations** - Stores translation history
|
||||
- Source and target text
|
||||
- Languages
|
||||
- Translation time and model used
|
||||
- Session and user tracking
|
||||
|
||||
2. **transcriptions** - Stores transcription history
|
||||
- Transcribed text
|
||||
- Detected language
|
||||
- Audio metadata
|
||||
- Performance metrics
|
||||
|
||||
3. **user_preferences** - Stores user settings
|
||||
- Preferred languages
|
||||
- Voice preferences
|
||||
- Usage statistics
|
||||
|
||||
4. **usage_analytics** - Aggregated analytics
|
||||
- Hourly and daily metrics
|
||||
- Service performance
|
||||
- Language pair statistics
|
||||
|
||||
5. **api_keys** - API key management
|
||||
- Rate limits
|
||||
- Permissions
|
||||
- Usage tracking
|
||||
|
||||
### Redis Usage
|
||||
|
||||
1. **Translation Cache**
|
||||
- Key: `translation:{source_lang}:{target_lang}:{text_hash}`
|
||||
- Expires: 24 hours
|
||||
- Reduces API calls to Ollama
|
||||
|
||||
2. **Session Management**
|
||||
- Key: `session:{session_id}`
|
||||
- Stores session data and resources
|
||||
- Expires: 1 hour (configurable)
|
||||
|
||||
3. **Rate Limiting**
|
||||
- Token bucket implementation
|
||||
- Per-client and global limits
|
||||
- Sliding window tracking
|
||||
|
||||
4. **Push Subscriptions**
|
||||
- Set: `push_subscriptions`
|
||||
- Individual subscriptions: `push_subscription:{id}`
|
||||
|
||||
## Setup Instructions
|
||||
|
||||
### Prerequisites
|
||||
|
||||
1. Install PostgreSQL:
|
||||
```bash
|
||||
# Ubuntu/Debian
|
||||
sudo apt-get install postgresql postgresql-contrib
|
||||
|
||||
# MacOS
|
||||
brew install postgresql
|
||||
```
|
||||
|
||||
2. Install Redis:
|
||||
```bash
|
||||
# Ubuntu/Debian
|
||||
sudo apt-get install redis-server
|
||||
|
||||
# MacOS
|
||||
brew install redis
|
||||
```
|
||||
|
||||
3. Install Python dependencies:
|
||||
```bash
|
||||
pip install -r requirements.txt
|
||||
```
|
||||
|
||||
### Quick Setup
|
||||
|
||||
Run the setup script:
|
||||
```bash
|
||||
./setup_databases.sh
|
||||
```
|
||||
|
||||
### Manual Setup
|
||||
|
||||
1. Create PostgreSQL database:
|
||||
```bash
|
||||
createdb talk2me
|
||||
```
|
||||
|
||||
2. Start Redis:
|
||||
```bash
|
||||
redis-server
|
||||
```
|
||||
|
||||
3. Create .env file with database URLs:
|
||||
```env
|
||||
DATABASE_URL=postgresql://username@localhost/talk2me
|
||||
REDIS_URL=redis://localhost:6379/0
|
||||
```
|
||||
|
||||
4. Initialize database:
|
||||
```bash
|
||||
python database_init.py
|
||||
```
|
||||
|
||||
5. Run migrations:
|
||||
```bash
|
||||
python migrations.py init
|
||||
python migrations.py create "Initial migration"
|
||||
python migrations.py run
|
||||
```
|
||||
|
||||
## Configuration
|
||||
|
||||
### Environment Variables
|
||||
|
||||
```env
|
||||
# PostgreSQL
|
||||
DATABASE_URL=postgresql://username:password@host:port/database
|
||||
SQLALCHEMY_DATABASE_URI=${DATABASE_URL}
|
||||
SQLALCHEMY_ENGINE_OPTIONS_POOL_SIZE=10
|
||||
SQLALCHEMY_ENGINE_OPTIONS_POOL_RECYCLE=3600
|
||||
|
||||
# Redis
|
||||
REDIS_URL=redis://localhost:6379/0
|
||||
REDIS_DECODE_RESPONSES=false
|
||||
REDIS_MAX_CONNECTIONS=50
|
||||
REDIS_SOCKET_TIMEOUT=5
|
||||
|
||||
# Session Management
|
||||
MAX_SESSION_DURATION=3600
|
||||
MAX_SESSION_IDLE_TIME=900
|
||||
MAX_RESOURCES_PER_SESSION=100
|
||||
MAX_BYTES_PER_SESSION=104857600
|
||||
```
|
||||
|
||||
## Migration from In-Memory to Database
|
||||
|
||||
### What Changed
|
||||
|
||||
1. **Rate Limiting**
|
||||
- Before: In-memory dictionaries
|
||||
- After: Redis sorted sets and hashes
|
||||
|
||||
2. **Session Management**
|
||||
- Before: In-memory session storage
|
||||
- After: Redis with automatic expiration
|
||||
|
||||
3. **Translation Cache**
|
||||
- Before: Client-side IndexedDB only
|
||||
- After: Server-side Redis cache + client cache
|
||||
|
||||
4. **Analytics**
|
||||
- Before: No persistent analytics
|
||||
- After: PostgreSQL aggregated metrics
|
||||
|
||||
### Migration Steps
|
||||
|
||||
1. Backup current app.py:
|
||||
```bash
|
||||
cp app.py app_backup.py
|
||||
```
|
||||
|
||||
2. Use the new app with database support:
|
||||
```bash
|
||||
cp app_with_db.py app.py
|
||||
```
|
||||
|
||||
3. Update any custom configurations in the new app.py
|
||||
|
||||
## API Changes
|
||||
|
||||
### New Endpoints
|
||||
|
||||
- `/api/history/translations` - Get translation history
|
||||
- `/api/history/transcriptions` - Get transcription history
|
||||
- `/api/preferences` - Get/update user preferences
|
||||
- `/api/analytics` - Get usage analytics
|
||||
|
||||
### Enhanced Features
|
||||
|
||||
1. **Translation Caching**
|
||||
- Automatic server-side caching
|
||||
- Reduced response time for repeated translations
|
||||
|
||||
2. **Session Persistence**
|
||||
- Sessions survive server restarts
|
||||
- Better resource tracking
|
||||
|
||||
3. **Improved Rate Limiting**
|
||||
- Distributed rate limiting across multiple servers
|
||||
- More accurate tracking
|
||||
|
||||
## Performance Considerations
|
||||
|
||||
1. **Database Indexes**
|
||||
- Indexes on session_id, user_id, languages
|
||||
- Composite indexes for common queries
|
||||
|
||||
2. **Redis Memory Usage**
|
||||
- Monitor with: `redis-cli info memory`
|
||||
- Configure maxmemory policy
|
||||
|
||||
3. **Connection Pooling**
|
||||
- PostgreSQL: 10 connections default
|
||||
- Redis: 50 connections default
|
||||
|
||||
## Monitoring
|
||||
|
||||
### PostgreSQL
|
||||
```sql
|
||||
-- Check database size
|
||||
SELECT pg_database_size('talk2me');
|
||||
|
||||
-- Active connections
|
||||
SELECT count(*) FROM pg_stat_activity;
|
||||
|
||||
-- Slow queries
|
||||
SELECT * FROM pg_stat_statements ORDER BY mean_time DESC LIMIT 10;
|
||||
```
|
||||
|
||||
### Redis
|
||||
```bash
|
||||
# Memory usage
|
||||
redis-cli info memory
|
||||
|
||||
# Connected clients
|
||||
redis-cli info clients
|
||||
|
||||
# Monitor commands
|
||||
redis-cli monitor
|
||||
```
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Common Issues
|
||||
|
||||
1. **PostgreSQL Connection Failed**
|
||||
- Check if PostgreSQL is running: `sudo systemctl status postgresql`
|
||||
- Verify DATABASE_URL in .env
|
||||
- Check pg_hba.conf for authentication
|
||||
|
||||
2. **Redis Connection Failed**
|
||||
- Check if Redis is running: `redis-cli ping`
|
||||
- Verify REDIS_URL in .env
|
||||
- Check Redis logs: `sudo journalctl -u redis`
|
||||
|
||||
3. **Migration Errors**
|
||||
- Drop and recreate database if needed
|
||||
- Check migration files in `migrations/`
|
||||
- Run `python migrations.py init` to reinitialize
|
||||
|
||||
## Backup and Restore
|
||||
|
||||
### PostgreSQL Backup
|
||||
```bash
|
||||
# Backup
|
||||
pg_dump talk2me > talk2me_backup.sql
|
||||
|
||||
# Restore
|
||||
psql talk2me < talk2me_backup.sql
|
||||
```
|
||||
|
||||
### Redis Backup
|
||||
```bash
|
||||
# Backup (if persistence enabled)
|
||||
redis-cli BGSAVE
|
||||
|
||||
# Copy dump.rdb file
|
||||
cp /var/lib/redis/dump.rdb redis_backup.rdb
|
||||
```
|
||||
|
||||
## Security Notes
|
||||
|
||||
1. **Database Credentials**
|
||||
- Never commit .env file
|
||||
- Use strong passwords
|
||||
- Limit database user permissions
|
||||
|
||||
2. **Redis Security**
|
||||
- Consider enabling Redis AUTH
|
||||
- Bind to localhost only
|
||||
- Use SSL for remote connections
|
||||
|
||||
3. **Data Privacy**
|
||||
- Translations/transcriptions contain user data
|
||||
- Implement data retention policies
|
||||
- Consider encryption at rest
|
46
Dockerfile
Normal file
46
Dockerfile
Normal file
@ -0,0 +1,46 @@
|
||||
# Production Dockerfile for Talk2Me
|
||||
FROM python:3.10-slim
|
||||
|
||||
# Install system dependencies
|
||||
RUN apt-get update && apt-get install -y \
|
||||
build-essential \
|
||||
curl \
|
||||
ffmpeg \
|
||||
git \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
|
||||
# Create non-root user
|
||||
RUN useradd -m -u 1000 talk2me
|
||||
|
||||
# Set working directory
|
||||
WORKDIR /app
|
||||
|
||||
# Copy requirements first for better caching
|
||||
COPY requirements.txt requirements-prod.txt ./
|
||||
RUN pip install --no-cache-dir -r requirements-prod.txt
|
||||
|
||||
# Copy application code
|
||||
COPY --chown=talk2me:talk2me . .
|
||||
|
||||
# Create necessary directories
|
||||
RUN mkdir -p logs /tmp/talk2me_uploads && \
|
||||
chown -R talk2me:talk2me logs /tmp/talk2me_uploads
|
||||
|
||||
# Switch to non-root user
|
||||
USER talk2me
|
||||
|
||||
# Set environment variables
|
||||
ENV FLASK_ENV=production \
|
||||
PYTHONUNBUFFERED=1 \
|
||||
UPLOAD_FOLDER=/tmp/talk2me_uploads \
|
||||
LOGS_DIR=/app/logs
|
||||
|
||||
# Health check
|
||||
HEALTHCHECK --interval=30s --timeout=10s --start-period=40s --retries=3 \
|
||||
CMD curl -f http://localhost:5005/health || exit 1
|
||||
|
||||
# Expose port
|
||||
EXPOSE 5005
|
||||
|
||||
# Run with gunicorn
|
||||
CMD ["gunicorn", "--config", "gunicorn_config.py", "wsgi:application"]
|
874
README.md
874
README.md
@ -1,9 +1,31 @@
|
||||
# Voice Language Translator
|
||||
# Talk2Me - Real-Time Voice Language Translator
|
||||
|
||||
A mobile-friendly web application that translates spoken language between multiple languages using:
|
||||
- Gemma 3 open-source LLM via Ollama for translation
|
||||
- OpenAI Whisper for speech-to-text
|
||||
- OpenAI Edge TTS for text-to-speech
|
||||
A production-ready, mobile-friendly web application that provides real-time translation of spoken language between multiple languages.
|
||||
|
||||
## Features
|
||||
|
||||
- **Real-time Speech Recognition**: Powered by OpenAI Whisper with GPU acceleration
|
||||
- **Advanced Translation**: Using Gemma 3 open-source LLM via Ollama
|
||||
- **Natural Text-to-Speech**: OpenAI Edge TTS for lifelike voice output
|
||||
- **Progressive Web App**: Full offline support with service workers
|
||||
- **Multi-Speaker Support**: Track and translate conversations with multiple participants
|
||||
- **Enterprise Security**: Comprehensive rate limiting, session management, and encrypted secrets
|
||||
- **Production Ready**: Docker support, load balancing, and extensive monitoring
|
||||
- **Admin Dashboard**: Real-time analytics, performance monitoring, and system health tracking
|
||||
|
||||
## Table of Contents
|
||||
|
||||
- [Supported Languages](#supported-languages)
|
||||
- [Quick Start](#quick-start)
|
||||
- [Installation](#installation)
|
||||
- [Configuration](#configuration)
|
||||
- [Security Features](#security-features)
|
||||
- [Production Deployment](#production-deployment)
|
||||
- [API Documentation](#api-documentation)
|
||||
- [Development](#development)
|
||||
- [Monitoring & Operations](#monitoring--operations)
|
||||
- [Troubleshooting](#troubleshooting)
|
||||
- [Contributing](#contributing)
|
||||
|
||||
## Supported Languages
|
||||
|
||||
@ -22,48 +44,830 @@ A mobile-friendly web application that translates spoken language between multip
|
||||
- Turkish
|
||||
- Uzbek
|
||||
|
||||
## Setup Instructions
|
||||
## Quick Start
|
||||
|
||||
1. Install the required Python packages:
|
||||
```
|
||||
```bash
|
||||
# Clone the repository
|
||||
git clone https://github.com/yourusername/talk2me.git
|
||||
cd talk2me
|
||||
|
||||
# Install dependencies
|
||||
pip install -r requirements.txt
|
||||
npm install
|
||||
|
||||
# Initialize secure configuration
|
||||
python manage_secrets.py init
|
||||
python manage_secrets.py set TTS_API_KEY your-api-key-here
|
||||
|
||||
# Ensure Ollama is running with Gemma
|
||||
ollama pull gemma2:9b
|
||||
ollama pull gemma3:27b
|
||||
|
||||
# Start the application
|
||||
python app.py
|
||||
```
|
||||
|
||||
Open your browser and navigate to `http://localhost:5005`
|
||||
|
||||
## Installation
|
||||
|
||||
### Prerequisites
|
||||
|
||||
- Python 3.8+
|
||||
- Node.js 14+
|
||||
- Ollama (for LLM translation)
|
||||
- OpenAI Edge TTS server
|
||||
- Optional: NVIDIA GPU with CUDA, AMD GPU with ROCm, or Apple Silicon
|
||||
|
||||
### Detailed Setup
|
||||
|
||||
1. **Install Python dependencies**:
|
||||
```bash
|
||||
python -m venv venv
|
||||
source venv/bin/activate # On Windows: venv\Scripts\activate
|
||||
pip install -r requirements.txt
|
||||
```
|
||||
|
||||
2. Make sure you have Ollama installed and the Gemma 3 model loaded:
|
||||
```
|
||||
ollama pull gemma3
|
||||
2. **Install Node.js dependencies**:
|
||||
```bash
|
||||
npm install
|
||||
npm run build # Build TypeScript files
|
||||
```
|
||||
|
||||
3. Ensure your OpenAI Edge TTS server is running on port 5050.
|
||||
|
||||
4. Run the application:
|
||||
```
|
||||
python app.py
|
||||
3. **Configure GPU Support** (Optional):
|
||||
```bash
|
||||
# For NVIDIA GPUs
|
||||
pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118
|
||||
|
||||
# For AMD GPUs (ROCm)
|
||||
pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm5.4.2
|
||||
|
||||
# For Apple Silicon
|
||||
pip install torch torchvision torchaudio
|
||||
```
|
||||
|
||||
5. Open your browser and navigate to:
|
||||
```
|
||||
http://localhost:8000
|
||||
4. **Set up Ollama**:
|
||||
```bash
|
||||
# Install Ollama (https://ollama.ai)
|
||||
curl -fsSL https://ollama.ai/install.sh | sh
|
||||
|
||||
# Pull required models
|
||||
ollama pull gemma2:9b # Faster, for streaming
|
||||
ollama pull gemma3:27b # Better quality
|
||||
```
|
||||
|
||||
## Usage
|
||||
5. **Configure TTS Server**:
|
||||
Ensure your OpenAI Edge TTS server is running. Default expected at `http://localhost:5050`
|
||||
|
||||
1. Select your source language from the dropdown menu
|
||||
2. Press the microphone button and speak
|
||||
3. Press the button again to stop recording
|
||||
4. Wait for the transcription to complete
|
||||
5. Select your target language
|
||||
6. Press the "Translate" button
|
||||
7. Use the play buttons to hear the original or translated text
|
||||
## Configuration
|
||||
|
||||
## Technical Details
|
||||
### Environment Variables
|
||||
|
||||
- The app uses Flask for the web server
|
||||
- Audio is processed client-side using the MediaRecorder API
|
||||
- Whisper for speech recognition with language hints
|
||||
- Ollama provides access to the Gemma 3 model for translation
|
||||
- OpenAI Edge TTS delivers natural-sounding speech output
|
||||
Talk2Me uses encrypted secrets management for sensitive configuration. You can use either the secure secrets system or traditional environment variables.
|
||||
|
||||
## Mobile Support
|
||||
#### Using Secure Secrets Management (Recommended)
|
||||
|
||||
The interface is fully responsive and designed to work well on mobile devices.
|
||||
```bash
|
||||
# Initialize the secrets system
|
||||
python manage_secrets.py init
|
||||
|
||||
# Set required secrets
|
||||
python manage_secrets.py set TTS_API_KEY
|
||||
python manage_secrets.py set TTS_SERVER_URL
|
||||
python manage_secrets.py set ADMIN_TOKEN
|
||||
|
||||
# List all secrets
|
||||
python manage_secrets.py list
|
||||
|
||||
# Rotate encryption keys
|
||||
python manage_secrets.py rotate
|
||||
```
|
||||
|
||||
#### Using Environment Variables
|
||||
|
||||
Create a `.env` file by copying `.env.example`:
|
||||
|
||||
```bash
|
||||
cp .env.example .env
|
||||
```
|
||||
|
||||
Key environment variables:
|
||||
|
||||
```env
|
||||
# Core Configuration (REQUIRED for production)
|
||||
FLASK_SECRET_KEY=your-secret-key-here-change-in-production # IMPORTANT: Set this for production!
|
||||
TTS_API_KEY=your-api-key-here
|
||||
TTS_SERVER_URL=http://localhost:5050/v1/audio/speech
|
||||
ADMIN_TOKEN=your-secure-admin-token
|
||||
|
||||
# CORS Configuration
|
||||
CORS_ORIGINS=https://yourdomain.com,https://app.yourdomain.com
|
||||
ADMIN_CORS_ORIGINS=https://admin.yourdomain.com
|
||||
|
||||
# Security Settings
|
||||
SECRET_KEY=your-secret-key-here
|
||||
MAX_CONTENT_LENGTH=52428800 # 50MB
|
||||
SESSION_LIFETIME=3600 # 1 hour
|
||||
RATE_LIMIT_STORAGE_URL=redis://localhost:6379/0
|
||||
|
||||
# Performance Tuning
|
||||
WHISPER_MODEL_SIZE=base
|
||||
GPU_MEMORY_THRESHOLD_MB=2048
|
||||
MEMORY_CLEANUP_INTERVAL=30
|
||||
```
|
||||
|
||||
**Important**: Always set `FLASK_SECRET_KEY` to a secure, random value in production. You can generate one using:
|
||||
|
||||
```bash
|
||||
python -c "import secrets; print(secrets.token_hex(32))"
|
||||
```
|
||||
|
||||
### Advanced Configuration
|
||||
|
||||
#### CORS Settings
|
||||
|
||||
```bash
|
||||
# Development (allow all origins)
|
||||
export CORS_ORIGINS="*"
|
||||
|
||||
# Production (restrict to specific domains)
|
||||
export CORS_ORIGINS="https://yourdomain.com,https://app.yourdomain.com"
|
||||
export ADMIN_CORS_ORIGINS="https://admin.yourdomain.com"
|
||||
```
|
||||
|
||||
#### Rate Limiting
|
||||
|
||||
Configure per-endpoint rate limits:
|
||||
|
||||
```python
|
||||
# In your config or via admin API
|
||||
RATE_LIMITS = {
|
||||
'default': {'requests_per_minute': 30, 'requests_per_hour': 500},
|
||||
'transcribe': {'requests_per_minute': 10, 'requests_per_hour': 100},
|
||||
'translate': {'requests_per_minute': 20, 'requests_per_hour': 300}
|
||||
}
|
||||
```
|
||||
|
||||
#### Session Management
|
||||
|
||||
```python
|
||||
SESSION_CONFIG = {
|
||||
'max_file_size_mb': 100,
|
||||
'max_files_per_session': 100,
|
||||
'idle_timeout_minutes': 15,
|
||||
'max_lifetime_minutes': 60
|
||||
}
|
||||
```
|
||||
|
||||
## Security Features
|
||||
|
||||
### 1. Rate Limiting
|
||||
|
||||
Comprehensive DoS protection with:
|
||||
- Token bucket algorithm with sliding window
|
||||
- Per-endpoint configurable limits
|
||||
- Automatic IP blocking for abusive clients
|
||||
- Request size validation
|
||||
|
||||
```bash
|
||||
# Check rate limit status
|
||||
curl -H "X-Admin-Token: $ADMIN_TOKEN" http://localhost:5005/admin/rate-limits
|
||||
|
||||
# Block an IP
|
||||
curl -X POST -H "X-Admin-Token: $ADMIN_TOKEN" \
|
||||
-H "Content-Type: application/json" \
|
||||
-d '{"ip": "192.168.1.100", "duration": 3600}' \
|
||||
http://localhost:5005/admin/block-ip
|
||||
```
|
||||
|
||||
### 2. Secrets Management
|
||||
|
||||
- AES-128 encryption for sensitive data
|
||||
- Automatic key rotation
|
||||
- Audit logging
|
||||
- Platform-specific secure storage
|
||||
|
||||
```bash
|
||||
# View audit log
|
||||
python manage_secrets.py audit
|
||||
|
||||
# Backup secrets
|
||||
python manage_secrets.py export --output backup.enc
|
||||
|
||||
# Restore from backup
|
||||
python manage_secrets.py import --input backup.enc
|
||||
```
|
||||
|
||||
### 3. Session Management
|
||||
|
||||
- Automatic resource tracking
|
||||
- Per-session limits (100 files, 100MB)
|
||||
- Idle session cleanup (15 minutes)
|
||||
- Real-time monitoring
|
||||
|
||||
```bash
|
||||
# View active sessions
|
||||
curl -H "X-Admin-Token: $ADMIN_TOKEN" http://localhost:5005/admin/sessions
|
||||
|
||||
# Clean up specific session
|
||||
curl -X POST -H "X-Admin-Token: $ADMIN_TOKEN" \
|
||||
http://localhost:5005/admin/sessions/SESSION_ID/cleanup
|
||||
```
|
||||
|
||||
### 4. Request Size Limits
|
||||
|
||||
- Global limit: 50MB
|
||||
- Audio files: 25MB
|
||||
- JSON payloads: 1MB
|
||||
- Dynamic configuration
|
||||
|
||||
```bash
|
||||
# Update size limits
|
||||
curl -X POST -H "X-Admin-Token: $ADMIN_TOKEN" \
|
||||
-H "Content-Type: application/json" \
|
||||
-d '{"max_audio_size": "30MB"}' \
|
||||
http://localhost:5005/admin/size-limits
|
||||
```
|
||||
|
||||
## Production Deployment
|
||||
|
||||
### Docker Deployment
|
||||
|
||||
```bash
|
||||
# Build and run with Docker Compose (CPU only)
|
||||
docker-compose up -d
|
||||
|
||||
# With NVIDIA GPU support
|
||||
docker-compose -f docker-compose.yml -f docker-compose.nvidia.yml up -d
|
||||
|
||||
# With AMD GPU support (ROCm)
|
||||
docker-compose -f docker-compose.yml -f docker-compose.amd.yml up -d
|
||||
|
||||
# With Apple Silicon support
|
||||
docker-compose -f docker-compose.yml -f docker-compose.apple.yml up -d
|
||||
|
||||
# Scale web workers
|
||||
docker-compose up -d --scale talk2me=4
|
||||
|
||||
# View logs
|
||||
docker-compose logs -f talk2me
|
||||
```
|
||||
|
||||
### Docker Compose Configuration
|
||||
|
||||
Choose the appropriate configuration based on your GPU:
|
||||
|
||||
#### NVIDIA GPU Configuration
|
||||
|
||||
```yaml
|
||||
version: '3.8'
|
||||
services:
|
||||
web:
|
||||
build: .
|
||||
ports:
|
||||
- "5005:5005"
|
||||
environment:
|
||||
- GUNICORN_WORKERS=4
|
||||
- GUNICORN_THREADS=2
|
||||
volumes:
|
||||
- ./logs:/app/logs
|
||||
- whisper-cache:/root/.cache/whisper
|
||||
deploy:
|
||||
resources:
|
||||
limits:
|
||||
memory: 4G
|
||||
reservations:
|
||||
devices:
|
||||
- driver: nvidia
|
||||
count: 1
|
||||
capabilities: [gpu]
|
||||
```
|
||||
|
||||
#### AMD GPU Configuration (ROCm)
|
||||
|
||||
```yaml
|
||||
version: '3.8'
|
||||
services:
|
||||
web:
|
||||
build: .
|
||||
ports:
|
||||
- "5005:5005"
|
||||
environment:
|
||||
- GUNICORN_WORKERS=4
|
||||
- GUNICORN_THREADS=2
|
||||
- HSA_OVERRIDE_GFX_VERSION=10.3.0 # Adjust for your GPU
|
||||
volumes:
|
||||
- ./logs:/app/logs
|
||||
- whisper-cache:/root/.cache/whisper
|
||||
- /dev/kfd:/dev/kfd # ROCm KFD interface
|
||||
- /dev/dri:/dev/dri # Direct Rendering Interface
|
||||
devices:
|
||||
- /dev/kfd
|
||||
- /dev/dri
|
||||
group_add:
|
||||
- video
|
||||
- render
|
||||
deploy:
|
||||
resources:
|
||||
limits:
|
||||
memory: 4G
|
||||
```
|
||||
|
||||
#### Apple Silicon Configuration
|
||||
|
||||
```yaml
|
||||
version: '3.8'
|
||||
services:
|
||||
web:
|
||||
build: .
|
||||
platform: linux/arm64/v8 # For M1/M2 Macs
|
||||
ports:
|
||||
- "5005:5005"
|
||||
environment:
|
||||
- GUNICORN_WORKERS=4
|
||||
- GUNICORN_THREADS=2
|
||||
- PYTORCH_ENABLE_MPS_FALLBACK=1 # Enable MPS fallback
|
||||
volumes:
|
||||
- ./logs:/app/logs
|
||||
- whisper-cache:/root/.cache/whisper
|
||||
deploy:
|
||||
resources:
|
||||
limits:
|
||||
memory: 4G
|
||||
```
|
||||
|
||||
#### CPU-Only Configuration
|
||||
|
||||
```yaml
|
||||
version: '3.8'
|
||||
services:
|
||||
web:
|
||||
build: .
|
||||
ports:
|
||||
- "5005:5005"
|
||||
environment:
|
||||
- GUNICORN_WORKERS=4
|
||||
- GUNICORN_THREADS=2
|
||||
- OMP_NUM_THREADS=4 # OpenMP threads for CPU
|
||||
volumes:
|
||||
- ./logs:/app/logs
|
||||
- whisper-cache:/root/.cache/whisper
|
||||
deploy:
|
||||
resources:
|
||||
limits:
|
||||
memory: 4G
|
||||
cpus: '4.0'
|
||||
```
|
||||
|
||||
### Nginx Configuration
|
||||
|
||||
```nginx
|
||||
upstream talk2me {
|
||||
least_conn;
|
||||
server web1:5005 weight=1 max_fails=3 fail_timeout=30s;
|
||||
server web2:5005 weight=1 max_fails=3 fail_timeout=30s;
|
||||
}
|
||||
|
||||
server {
|
||||
listen 443 ssl http2;
|
||||
server_name talk2me.yourdomain.com;
|
||||
|
||||
ssl_certificate /etc/ssl/certs/talk2me.crt;
|
||||
ssl_certificate_key /etc/ssl/private/talk2me.key;
|
||||
|
||||
client_max_body_size 50M;
|
||||
|
||||
location / {
|
||||
proxy_pass http://talk2me;
|
||||
proxy_set_header X-Real-IP $remote_addr;
|
||||
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
||||
proxy_set_header Host $host;
|
||||
|
||||
# WebSocket support
|
||||
proxy_http_version 1.1;
|
||||
proxy_set_header Upgrade $http_upgrade;
|
||||
proxy_set_header Connection "upgrade";
|
||||
}
|
||||
|
||||
# Cache static assets
|
||||
location /static/ {
|
||||
alias /app/static/;
|
||||
expires 30d;
|
||||
add_header Cache-Control "public, immutable";
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### Systemd Service
|
||||
|
||||
```ini
|
||||
[Unit]
|
||||
Description=Talk2Me Translation Service
|
||||
After=network.target
|
||||
|
||||
[Service]
|
||||
Type=notify
|
||||
User=talk2me
|
||||
Group=talk2me
|
||||
WorkingDirectory=/opt/talk2me
|
||||
Environment="PATH=/opt/talk2me/venv/bin"
|
||||
ExecStart=/opt/talk2me/venv/bin/gunicorn \
|
||||
--config gunicorn_config.py \
|
||||
--bind 0.0.0.0:5005 \
|
||||
app:app
|
||||
Restart=always
|
||||
RestartSec=10
|
||||
|
||||
[Install]
|
||||
WantedBy=multi-user.target
|
||||
```
|
||||
|
||||
## API Documentation
|
||||
|
||||
### Core Endpoints
|
||||
|
||||
#### Transcribe Audio
|
||||
```http
|
||||
POST /transcribe
|
||||
Content-Type: multipart/form-data
|
||||
|
||||
audio: (binary)
|
||||
source_lang: auto|language_code
|
||||
```
|
||||
|
||||
#### Translate Text
|
||||
```http
|
||||
POST /translate
|
||||
Content-Type: application/json
|
||||
|
||||
{
|
||||
"text": "Hello world",
|
||||
"source_lang": "English",
|
||||
"target_lang": "Spanish"
|
||||
}
|
||||
```
|
||||
|
||||
#### Streaming Translation
|
||||
```http
|
||||
POST /translate/stream
|
||||
Content-Type: application/json
|
||||
|
||||
{
|
||||
"text": "Long text to translate",
|
||||
"source_lang": "auto",
|
||||
"target_lang": "French"
|
||||
}
|
||||
|
||||
Response: Server-Sent Events stream
|
||||
```
|
||||
|
||||
#### Text-to-Speech
|
||||
```http
|
||||
POST /speak
|
||||
Content-Type: application/json
|
||||
|
||||
{
|
||||
"text": "Hola mundo",
|
||||
"language": "Spanish"
|
||||
}
|
||||
```
|
||||
|
||||
### Admin Endpoints
|
||||
|
||||
All admin endpoints require `X-Admin-Token` header.
|
||||
|
||||
#### Health & Monitoring
|
||||
- `GET /health` - Basic health check
|
||||
- `GET /health/detailed` - Component status
|
||||
- `GET /metrics` - Prometheus metrics
|
||||
- `GET /admin/memory` - Memory usage stats
|
||||
|
||||
#### Session Management
|
||||
- `GET /admin/sessions` - List active sessions
|
||||
- `GET /admin/sessions/:id` - Session details
|
||||
- `POST /admin/sessions/:id/cleanup` - Manual cleanup
|
||||
|
||||
#### Security Controls
|
||||
- `GET /admin/rate-limits` - View rate limits
|
||||
- `POST /admin/block-ip` - Block IP address
|
||||
- `GET /admin/logs/security` - Security events
|
||||
|
||||
## Admin Dashboard
|
||||
|
||||
Talk2Me includes a comprehensive admin analytics dashboard for monitoring and managing the application.
|
||||
|
||||
### Features
|
||||
|
||||
- **Real-time Analytics**: Monitor requests, active sessions, and error rates
|
||||
- **Performance Metrics**: Track response times, throughput, and resource usage
|
||||
- **System Health**: Monitor Redis, PostgreSQL, and ML services status
|
||||
- **Language Analytics**: View popular language pairs and usage patterns
|
||||
- **Error Analysis**: Detailed error tracking with types and trends
|
||||
- **Data Export**: Download analytics data in JSON format
|
||||
|
||||
### Setup
|
||||
|
||||
1. **Initialize Database**:
|
||||
```bash
|
||||
python init_analytics_db.py
|
||||
```
|
||||
|
||||
2. **Configure Admin Token**:
|
||||
```bash
|
||||
export ADMIN_TOKEN="your-secure-admin-token"
|
||||
```
|
||||
|
||||
3. **Access Dashboard**:
|
||||
- Navigate to `https://yourdomain.com/admin`
|
||||
- Enter your admin token
|
||||
- View real-time analytics
|
||||
|
||||
### Dashboard Sections
|
||||
|
||||
- **Overview Cards**: Key metrics at a glance
|
||||
- **Request Volume**: Visualize traffic patterns
|
||||
- **Operations**: Translation and transcription statistics
|
||||
- **Performance**: Response time percentiles (P95, P99)
|
||||
- **Error Tracking**: Error types and recent issues
|
||||
- **System Health**: Component status monitoring
|
||||
|
||||
For detailed admin documentation, see [ADMIN_DASHBOARD.md](ADMIN_DASHBOARD.md).
|
||||
|
||||
## Development
|
||||
|
||||
### TypeScript Development
|
||||
|
||||
```bash
|
||||
# Install dependencies
|
||||
npm install
|
||||
|
||||
# Development mode with auto-compilation
|
||||
npm run dev
|
||||
|
||||
# Build for production
|
||||
npm run build
|
||||
|
||||
# Type checking
|
||||
npm run typecheck
|
||||
```
|
||||
|
||||
### Project Structure
|
||||
|
||||
```
|
||||
talk2me/
|
||||
├── app.py # Main Flask application
|
||||
├── config.py # Configuration management
|
||||
├── requirements.txt # Python dependencies
|
||||
├── package.json # Node.js dependencies
|
||||
├── tsconfig.json # TypeScript configuration
|
||||
├── gunicorn_config.py # Production server config
|
||||
├── docker-compose.yml # Container orchestration
|
||||
├── static/
|
||||
│ ├── js/
|
||||
│ │ ├── src/ # TypeScript source files
|
||||
│ │ └── dist/ # Compiled JavaScript
|
||||
│ ├── css/ # Stylesheets
|
||||
│ └── icons/ # PWA icons
|
||||
├── templates/ # HTML templates
|
||||
├── logs/ # Application logs
|
||||
└── tests/ # Test suite
|
||||
```
|
||||
|
||||
### Key Components
|
||||
|
||||
1. **Connection Management** (`connectionManager.ts`)
|
||||
- Automatic retry with exponential backoff
|
||||
- Request queuing during offline periods
|
||||
- Connection status monitoring
|
||||
|
||||
2. **Translation Cache** (`translationCache.ts`)
|
||||
- IndexedDB for offline support
|
||||
- LRU eviction policy
|
||||
- Automatic cache size management
|
||||
|
||||
3. **Speaker Management** (`speakerManager.ts`)
|
||||
- Multi-speaker conversation tracking
|
||||
- Speaker-specific audio handling
|
||||
- Conversation export functionality
|
||||
|
||||
4. **Error Handling** (`errorBoundary.ts`)
|
||||
- Global error catching
|
||||
- Automatic error reporting
|
||||
- User-friendly error messages
|
||||
|
||||
### Running Tests
|
||||
|
||||
```bash
|
||||
# Python tests
|
||||
pytest tests/ -v
|
||||
|
||||
# TypeScript tests
|
||||
npm test
|
||||
|
||||
# Integration tests
|
||||
python test_integration.py
|
||||
```
|
||||
|
||||
## Monitoring & Operations
|
||||
|
||||
### Logging System
|
||||
|
||||
Talk2Me uses structured JSON logging with multiple streams:
|
||||
|
||||
```bash
|
||||
logs/
|
||||
├── talk2me.log # General application log
|
||||
├── errors.log # Error-specific log
|
||||
├── access.log # HTTP access log
|
||||
├── security.log # Security events
|
||||
└── performance.log # Performance metrics
|
||||
```
|
||||
|
||||
View logs:
|
||||
```bash
|
||||
# Recent errors
|
||||
tail -f logs/errors.log | jq '.'
|
||||
|
||||
# Security events
|
||||
grep "rate_limit_exceeded" logs/security.log | jq '.'
|
||||
|
||||
# Slow requests
|
||||
jq 'select(.extra_fields.duration_ms > 1000)' logs/performance.log
|
||||
```
|
||||
|
||||
### Memory Management
|
||||
|
||||
Talk2Me includes comprehensive memory leak prevention:
|
||||
|
||||
1. **Backend Memory Management**
|
||||
- GPU memory monitoring
|
||||
- Automatic model reloading
|
||||
- Temporary file cleanup
|
||||
|
||||
2. **Frontend Memory Management**
|
||||
- Audio blob cleanup
|
||||
- WebRTC resource management
|
||||
- Event listener cleanup
|
||||
|
||||
Monitor memory:
|
||||
```bash
|
||||
# Check memory stats
|
||||
curl -H "X-Admin-Token: $ADMIN_TOKEN" http://localhost:5005/admin/memory
|
||||
|
||||
# Trigger manual cleanup
|
||||
curl -X POST -H "X-Admin-Token: $ADMIN_TOKEN" \
|
||||
http://localhost:5005/admin/memory/cleanup
|
||||
```
|
||||
|
||||
### Performance Tuning
|
||||
|
||||
#### GPU Optimization
|
||||
|
||||
```python
|
||||
# config.py or environment
|
||||
GPU_OPTIMIZATIONS = {
|
||||
'enabled': True,
|
||||
'fp16': True, # Half precision for 2x speedup
|
||||
'batch_size': 1, # Adjust based on GPU memory
|
||||
'num_workers': 2, # Parallel data loading
|
||||
'pin_memory': True # Faster GPU transfer
|
||||
}
|
||||
```
|
||||
|
||||
#### Whisper Optimization
|
||||
|
||||
```python
|
||||
TRANSCRIBE_OPTIONS = {
|
||||
'beam_size': 1, # Faster inference
|
||||
'best_of': 1, # Disable multiple attempts
|
||||
'temperature': 0, # Deterministic output
|
||||
'compression_ratio_threshold': 2.4,
|
||||
'logprob_threshold': -1.0,
|
||||
'no_speech_threshold': 0.6
|
||||
}
|
||||
```
|
||||
|
||||
### Scaling Considerations
|
||||
|
||||
1. **Horizontal Scaling**
|
||||
- Use Redis for shared rate limiting
|
||||
- Configure sticky sessions for WebSocket
|
||||
- Share audio files via object storage
|
||||
|
||||
2. **Vertical Scaling**
|
||||
- Increase worker processes
|
||||
- Tune thread pool size
|
||||
- Allocate more GPU memory
|
||||
|
||||
3. **Caching Strategy**
|
||||
- Cache translations in Redis
|
||||
- Use CDN for static assets
|
||||
- Enable HTTP caching headers
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Common Issues
|
||||
|
||||
#### GPU Not Detected
|
||||
|
||||
```bash
|
||||
# Check CUDA availability
|
||||
python -c "import torch; print(torch.cuda.is_available())"
|
||||
|
||||
# Check GPU memory
|
||||
nvidia-smi
|
||||
|
||||
# For AMD GPUs
|
||||
rocm-smi
|
||||
|
||||
# For Apple Silicon
|
||||
python -c "import torch; print(torch.backends.mps.is_available())"
|
||||
```
|
||||
|
||||
#### High Memory Usage
|
||||
|
||||
```bash
|
||||
# Check for memory leaks
|
||||
curl -H "X-Admin-Token: $ADMIN_TOKEN" http://localhost:5005/health/storage
|
||||
|
||||
# Manual cleanup
|
||||
curl -X POST -H "X-Admin-Token: $ADMIN_TOKEN" \
|
||||
http://localhost:5005/admin/cleanup
|
||||
```
|
||||
|
||||
#### CORS Issues
|
||||
|
||||
```bash
|
||||
# Test CORS configuration
|
||||
curl -X OPTIONS http://localhost:5005/api/transcribe \
|
||||
-H "Origin: https://yourdomain.com" \
|
||||
-H "Access-Control-Request-Method: POST"
|
||||
```
|
||||
|
||||
#### TTS Server Connection
|
||||
|
||||
```bash
|
||||
# Check TTS server status
|
||||
curl http://localhost:5005/check_tts_server
|
||||
|
||||
# Update TTS configuration
|
||||
curl -X POST http://localhost:5005/update_tts_config \
|
||||
-H "Content-Type: application/json" \
|
||||
-d '{"server_url": "http://localhost:5050/v1/audio/speech", "api_key": "new-key"}'
|
||||
```
|
||||
|
||||
### Debug Mode
|
||||
|
||||
Enable debug logging:
|
||||
```bash
|
||||
export FLASK_ENV=development
|
||||
export LOG_LEVEL=DEBUG
|
||||
python app.py
|
||||
```
|
||||
|
||||
### Performance Profiling
|
||||
|
||||
```bash
|
||||
# Enable performance logging
|
||||
export ENABLE_PROFILING=true
|
||||
|
||||
# View slow requests
|
||||
jq 'select(.duration_ms > 1000)' logs/performance.log
|
||||
```
|
||||
|
||||
## Contributing
|
||||
|
||||
We welcome contributions! Please see our [Contributing Guidelines](CONTRIBUTING.md) for details.
|
||||
|
||||
### Development Setup
|
||||
|
||||
1. Fork the repository
|
||||
2. Create a feature branch (`git checkout -b feature/amazing-feature`)
|
||||
3. Make your changes
|
||||
4. Run tests (`pytest && npm test`)
|
||||
5. Commit your changes (`git commit -m 'Add amazing feature'`)
|
||||
6. Push to the branch (`git push origin feature/amazing-feature`)
|
||||
7. Open a Pull Request
|
||||
|
||||
### Code Style
|
||||
|
||||
- Python: Follow PEP 8
|
||||
- TypeScript: Use ESLint configuration
|
||||
- Commit messages: Use conventional commits
|
||||
|
||||
## License
|
||||
|
||||
This project is licensed under the MIT License - see the [LICENSE](LICENSE) file for details.
|
||||
|
||||
## Acknowledgments
|
||||
|
||||
- OpenAI Whisper team for the amazing speech recognition model
|
||||
- Ollama team for making LLMs accessible
|
||||
- All contributors who have helped improve Talk2Me
|
||||
|
||||
## Support
|
||||
|
||||
- **Documentation**: Full docs at [docs.talk2me.app](https://docs.talk2me.app)
|
||||
- **Issues**: [GitHub Issues](https://github.com/yourusername/talk2me/issues)
|
||||
- **Discussions**: [GitHub Discussions](https://github.com/yourusername/talk2me/discussions)
|
||||
- **Security**: Please report security vulnerabilities to security@talk2me.app
|
155
REVERSE_PROXY.md
Normal file
155
REVERSE_PROXY.md
Normal file
@ -0,0 +1,155 @@
|
||||
# Nginx Reverse Proxy Configuration for Talk2Me
|
||||
|
||||
## Nginx Configuration
|
||||
|
||||
Add the following to your Nginx configuration for the domain `talk2me.dr74.net`:
|
||||
|
||||
```nginx
|
||||
server {
|
||||
listen 443 ssl http2;
|
||||
server_name talk2me.dr74.net;
|
||||
|
||||
# SSL configuration
|
||||
ssl_certificate /path/to/ssl/cert.pem;
|
||||
ssl_certificate_key /path/to/ssl/key.pem;
|
||||
|
||||
# Security headers
|
||||
add_header X-Content-Type-Options nosniff;
|
||||
add_header X-Frame-Options SAMEORIGIN;
|
||||
add_header X-XSS-Protection "1; mode=block";
|
||||
add_header Referrer-Policy "no-referrer-when-downgrade";
|
||||
|
||||
# Proxy settings
|
||||
location / {
|
||||
proxy_pass http://localhost:5000; # Adjust port as needed
|
||||
proxy_http_version 1.1;
|
||||
|
||||
# Important headers for PWA and WebSocket support
|
||||
proxy_set_header Host $host;
|
||||
proxy_set_header X-Real-IP $remote_addr;
|
||||
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
||||
proxy_set_header X-Forwarded-Proto $scheme;
|
||||
proxy_set_header X-Forwarded-Host $host;
|
||||
|
||||
# WebSocket support
|
||||
proxy_set_header Upgrade $http_upgrade;
|
||||
proxy_set_header Connection "upgrade";
|
||||
|
||||
# Timeouts for long-running requests
|
||||
proxy_connect_timeout 60s;
|
||||
proxy_send_timeout 60s;
|
||||
proxy_read_timeout 60s;
|
||||
|
||||
# Buffer settings
|
||||
proxy_buffering off;
|
||||
proxy_buffer_size 4k;
|
||||
proxy_buffers 8 4k;
|
||||
proxy_busy_buffers_size 8k;
|
||||
|
||||
# Disable cache for dynamic content
|
||||
proxy_cache_bypass 1;
|
||||
proxy_no_cache 1;
|
||||
}
|
||||
|
||||
# Static files with caching
|
||||
location /static/ {
|
||||
proxy_pass http://localhost:5000/static/;
|
||||
proxy_set_header Host $host;
|
||||
proxy_set_header X-Real-IP $remote_addr;
|
||||
|
||||
# Cache static files
|
||||
expires 30d;
|
||||
add_header Cache-Control "public, immutable";
|
||||
}
|
||||
|
||||
# Service worker needs special handling
|
||||
location /service-worker.js {
|
||||
proxy_pass http://localhost:5000/service-worker.js;
|
||||
proxy_set_header Host $host;
|
||||
proxy_set_header X-Real-IP $remote_addr;
|
||||
|
||||
# No cache for service worker
|
||||
add_header Cache-Control "no-cache, no-store, must-revalidate";
|
||||
add_header Pragma "no-cache";
|
||||
expires 0;
|
||||
}
|
||||
|
||||
# Manifest file
|
||||
location /static/manifest.json {
|
||||
proxy_pass http://localhost:5000/static/manifest.json;
|
||||
proxy_set_header Host $host;
|
||||
proxy_set_header X-Real-IP $remote_addr;
|
||||
|
||||
# Allow manifest to be cached briefly
|
||||
expires 1h;
|
||||
add_header Cache-Control "public";
|
||||
add_header Content-Type "application/manifest+json";
|
||||
}
|
||||
}
|
||||
|
||||
# Redirect HTTP to HTTPS
|
||||
server {
|
||||
listen 80;
|
||||
server_name talk2me.dr74.net;
|
||||
return 301 https://$server_name$request_uri;
|
||||
}
|
||||
```
|
||||
|
||||
## Flask Application Configuration
|
||||
|
||||
Set these environment variables for the Talk2Me application:
|
||||
|
||||
```bash
|
||||
# Add to .env file or set as environment variables
|
||||
FLASK_ENV=production
|
||||
SESSION_COOKIE_SECURE=true
|
||||
SESSION_COOKIE_SAMESITE=Lax
|
||||
PREFERRED_URL_SCHEME=https
|
||||
|
||||
# If using a non-standard port
|
||||
# SERVER_NAME=talk2me.dr74.net
|
||||
```
|
||||
|
||||
## Testing the Configuration
|
||||
|
||||
1. **Check SSL Certificate**:
|
||||
```bash
|
||||
curl -I https://talk2me.dr74.net
|
||||
```
|
||||
|
||||
2. **Verify Service Worker**:
|
||||
```bash
|
||||
curl https://talk2me.dr74.net/service-worker.js
|
||||
```
|
||||
|
||||
3. **Check Manifest**:
|
||||
```bash
|
||||
curl https://talk2me.dr74.net/static/manifest.json
|
||||
```
|
||||
|
||||
4. **Test PWA Installation**:
|
||||
- Visit https://talk2me.dr74.net in Chrome
|
||||
- Open Developer Tools (F12)
|
||||
- Go to Application tab
|
||||
- Check "Manifest" section for any errors
|
||||
- Check "Service Workers" section
|
||||
|
||||
## Common Issues and Solutions
|
||||
|
||||
### Issue: Icons not loading
|
||||
- Ensure static files are being served correctly
|
||||
- Check Nginx error logs: `tail -f /var/log/nginx/error.log`
|
||||
|
||||
### Issue: Service Worker not registering
|
||||
- Verify HTTPS is working correctly
|
||||
- Check browser console for errors
|
||||
- Ensure service worker scope is correct
|
||||
|
||||
### Issue: "Add to Home Screen" not appearing
|
||||
- Clear browser cache and data
|
||||
- Ensure all manifest requirements are met
|
||||
- Check Chrome's PWA criteria in DevTools Lighthouse
|
||||
|
||||
### Issue: WebSocket connections failing
|
||||
- Verify Nginx has WebSocket headers configured
|
||||
- Check if firewall allows WebSocket connections
|
667
admin/__init__.py
Normal file
667
admin/__init__.py
Normal file
@ -0,0 +1,667 @@
|
||||
from flask import Blueprint, request, jsonify, render_template, redirect, url_for, session, current_app
|
||||
from functools import wraps
|
||||
import os
|
||||
import logging
|
||||
import json
|
||||
from datetime import datetime, timedelta
|
||||
import redis
|
||||
import psycopg2
|
||||
from psycopg2.extras import RealDictCursor
|
||||
import time
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
# Create admin blueprint
|
||||
admin_bp = Blueprint('admin', __name__,
|
||||
template_folder='templates',
|
||||
static_folder='static',
|
||||
static_url_path='/admin/static')
|
||||
|
||||
# Initialize Redis and PostgreSQL connections
|
||||
redis_client = None
|
||||
pg_conn = None
|
||||
|
||||
def init_admin(app):
|
||||
"""Initialize admin module with app configuration"""
|
||||
global redis_client, pg_conn
|
||||
|
||||
try:
|
||||
# Initialize Redis
|
||||
redis_client = redis.from_url(
|
||||
app.config.get('REDIS_URL', 'redis://localhost:6379/0'),
|
||||
decode_responses=True
|
||||
)
|
||||
redis_client.ping()
|
||||
logger.info("Redis connection established for admin dashboard")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to connect to Redis: {e}")
|
||||
redis_client = None
|
||||
|
||||
try:
|
||||
# Initialize PostgreSQL
|
||||
pg_conn = psycopg2.connect(
|
||||
app.config.get('DATABASE_URL', 'postgresql://localhost/talk2me')
|
||||
)
|
||||
logger.info("PostgreSQL connection established for admin dashboard")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to connect to PostgreSQL: {e}")
|
||||
pg_conn = None
|
||||
|
||||
def admin_required(f):
|
||||
"""Decorator to require admin authentication"""
|
||||
@wraps(f)
|
||||
def decorated_function(*args, **kwargs):
|
||||
# Check if user is logged in with admin role (from unified login)
|
||||
user_role = session.get('user_role')
|
||||
if user_role == 'admin':
|
||||
return f(*args, **kwargs)
|
||||
|
||||
# Also support the old admin token for backward compatibility
|
||||
auth_token = request.headers.get('X-Admin-Token')
|
||||
session_token = session.get('admin_token')
|
||||
expected_token = os.environ.get('ADMIN_TOKEN', 'default-admin-token')
|
||||
|
||||
if auth_token == expected_token or session_token == expected_token:
|
||||
if auth_token == expected_token:
|
||||
session['admin_token'] = expected_token
|
||||
return f(*args, **kwargs)
|
||||
|
||||
# For API endpoints, return JSON error
|
||||
if request.path.startswith('/admin/api/'):
|
||||
return jsonify({'error': 'Unauthorized'}), 401
|
||||
|
||||
# For web pages, redirect to unified login
|
||||
return redirect(url_for('login', next=request.url))
|
||||
|
||||
return decorated_function
|
||||
|
||||
@admin_bp.route('/login', methods=['GET', 'POST'])
|
||||
def login():
|
||||
"""Admin login - redirect to main login page"""
|
||||
# Redirect to the unified login page
|
||||
next_url = request.args.get('next', url_for('admin.dashboard'))
|
||||
return redirect(url_for('login', next=next_url))
|
||||
|
||||
@admin_bp.route('/logout')
|
||||
def logout():
|
||||
"""Admin logout - redirect to main logout"""
|
||||
# Clear all session data
|
||||
session.clear()
|
||||
return redirect(url_for('index'))
|
||||
|
||||
@admin_bp.route('/')
|
||||
@admin_bp.route('/dashboard')
|
||||
@admin_required
|
||||
def dashboard():
|
||||
"""Main admin dashboard"""
|
||||
return render_template('dashboard.html')
|
||||
|
||||
@admin_bp.route('/users')
|
||||
@admin_required
|
||||
def users():
|
||||
"""User management page"""
|
||||
# The template is in the main templates folder, not admin/templates
|
||||
return render_template('admin_users.html')
|
||||
|
||||
# Analytics API endpoints
|
||||
@admin_bp.route('/api/stats/overview')
|
||||
@admin_required
|
||||
def get_overview_stats():
|
||||
"""Get overview statistics"""
|
||||
try:
|
||||
stats = {
|
||||
'requests': {'total': 0, 'today': 0, 'hour': 0},
|
||||
'translations': {'total': 0, 'today': 0},
|
||||
'transcriptions': {'total': 0, 'today': 0},
|
||||
'active_sessions': 0,
|
||||
'error_rate': 0,
|
||||
'cache_hit_rate': 0,
|
||||
'system_health': check_system_health()
|
||||
}
|
||||
|
||||
# Get data from Redis
|
||||
if redis_client:
|
||||
try:
|
||||
# Request counts
|
||||
stats['requests']['total'] = int(redis_client.get('stats:requests:total') or 0)
|
||||
stats['requests']['today'] = int(redis_client.get(f'stats:requests:daily:{datetime.now().strftime("%Y-%m-%d")}') or 0)
|
||||
stats['requests']['hour'] = int(redis_client.get(f'stats:requests:hourly:{datetime.now().strftime("%Y-%m-%d-%H")}') or 0)
|
||||
|
||||
# Operation counts
|
||||
stats['translations']['total'] = int(redis_client.get('stats:translations:total') or 0)
|
||||
stats['translations']['today'] = int(redis_client.get(f'stats:translations:daily:{datetime.now().strftime("%Y-%m-%d")}') or 0)
|
||||
stats['transcriptions']['total'] = int(redis_client.get('stats:transcriptions:total') or 0)
|
||||
stats['transcriptions']['today'] = int(redis_client.get(f'stats:transcriptions:daily:{datetime.now().strftime("%Y-%m-%d")}') or 0)
|
||||
|
||||
# Active sessions
|
||||
stats['active_sessions'] = len(redis_client.keys('session:*'))
|
||||
|
||||
# Cache stats
|
||||
cache_hits = int(redis_client.get('stats:cache:hits') or 0)
|
||||
cache_misses = int(redis_client.get('stats:cache:misses') or 0)
|
||||
if cache_hits + cache_misses > 0:
|
||||
stats['cache_hit_rate'] = round((cache_hits / (cache_hits + cache_misses)) * 100, 2)
|
||||
|
||||
# Error rate
|
||||
total_requests = stats['requests']['today']
|
||||
errors_today = int(redis_client.get(f'stats:errors:daily:{datetime.now().strftime("%Y-%m-%d")}') or 0)
|
||||
if total_requests > 0:
|
||||
stats['error_rate'] = round((errors_today / total_requests) * 100, 2)
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error fetching Redis stats: {e}")
|
||||
|
||||
return jsonify(stats)
|
||||
except Exception as e:
|
||||
logger.error(f"Error in get_overview_stats: {e}")
|
||||
return jsonify({'error': str(e)}), 500
|
||||
|
||||
@admin_bp.route('/api/stats/requests/<timeframe>')
|
||||
@admin_required
|
||||
def get_request_stats(timeframe):
|
||||
"""Get request statistics for different timeframes"""
|
||||
try:
|
||||
if timeframe not in ['minute', 'hour', 'day']:
|
||||
return jsonify({'error': 'Invalid timeframe'}), 400
|
||||
|
||||
data = []
|
||||
labels = []
|
||||
|
||||
if redis_client:
|
||||
now = datetime.now()
|
||||
|
||||
if timeframe == 'minute':
|
||||
# Last 60 minutes
|
||||
for i in range(59, -1, -1):
|
||||
time_key = (now - timedelta(minutes=i)).strftime('%Y-%m-%d-%H-%M')
|
||||
count = int(redis_client.get(f'stats:requests:minute:{time_key}') or 0)
|
||||
data.append(count)
|
||||
labels.append((now - timedelta(minutes=i)).strftime('%H:%M'))
|
||||
|
||||
elif timeframe == 'hour':
|
||||
# Last 24 hours
|
||||
for i in range(23, -1, -1):
|
||||
time_key = (now - timedelta(hours=i)).strftime('%Y-%m-%d-%H')
|
||||
count = int(redis_client.get(f'stats:requests:hourly:{time_key}') or 0)
|
||||
data.append(count)
|
||||
labels.append((now - timedelta(hours=i)).strftime('%H:00'))
|
||||
|
||||
elif timeframe == 'day':
|
||||
# Last 30 days
|
||||
for i in range(29, -1, -1):
|
||||
time_key = (now - timedelta(days=i)).strftime('%Y-%m-%d')
|
||||
count = int(redis_client.get(f'stats:requests:daily:{time_key}') or 0)
|
||||
data.append(count)
|
||||
labels.append((now - timedelta(days=i)).strftime('%m/%d'))
|
||||
|
||||
return jsonify({
|
||||
'labels': labels,
|
||||
'data': data,
|
||||
'timeframe': timeframe
|
||||
})
|
||||
except Exception as e:
|
||||
logger.error(f"Error in get_request_stats: {e}")
|
||||
return jsonify({'error': str(e)}), 500
|
||||
|
||||
@admin_bp.route('/api/stats/operations')
|
||||
@admin_required
|
||||
def get_operation_stats():
|
||||
"""Get translation and transcription statistics"""
|
||||
try:
|
||||
stats = {
|
||||
'translations': {'data': [], 'labels': []},
|
||||
'transcriptions': {'data': [], 'labels': []},
|
||||
'language_pairs': {},
|
||||
'response_times': {'translation': [], 'transcription': []}
|
||||
}
|
||||
|
||||
if redis_client:
|
||||
now = datetime.now()
|
||||
|
||||
# Get daily stats for last 7 days
|
||||
for i in range(6, -1, -1):
|
||||
date_key = (now - timedelta(days=i)).strftime('%Y-%m-%d')
|
||||
date_label = (now - timedelta(days=i)).strftime('%m/%d')
|
||||
|
||||
# Translation counts
|
||||
trans_count = int(redis_client.get(f'stats:translations:daily:{date_key}') or 0)
|
||||
stats['translations']['data'].append(trans_count)
|
||||
stats['translations']['labels'].append(date_label)
|
||||
|
||||
# Transcription counts
|
||||
transcr_count = int(redis_client.get(f'stats:transcriptions:daily:{date_key}') or 0)
|
||||
stats['transcriptions']['data'].append(transcr_count)
|
||||
stats['transcriptions']['labels'].append(date_label)
|
||||
|
||||
# Get language pair statistics
|
||||
lang_pairs = redis_client.hgetall('stats:language_pairs') or {}
|
||||
stats['language_pairs'] = {k: int(v) for k, v in lang_pairs.items()}
|
||||
|
||||
# Get response times (last 100 operations)
|
||||
trans_times = redis_client.lrange('stats:response_times:translation', 0, 99)
|
||||
transcr_times = redis_client.lrange('stats:response_times:transcription', 0, 99)
|
||||
|
||||
stats['response_times']['translation'] = [float(t) for t in trans_times[:20]]
|
||||
stats['response_times']['transcription'] = [float(t) for t in transcr_times[:20]]
|
||||
|
||||
return jsonify(stats)
|
||||
except Exception as e:
|
||||
logger.error(f"Error in get_operation_stats: {e}")
|
||||
return jsonify({'error': str(e)}), 500
|
||||
|
||||
@admin_bp.route('/api/stats/errors')
|
||||
@admin_required
|
||||
def get_error_stats():
|
||||
"""Get error statistics"""
|
||||
try:
|
||||
stats = {
|
||||
'error_types': {},
|
||||
'error_timeline': {'data': [], 'labels': []},
|
||||
'recent_errors': []
|
||||
}
|
||||
|
||||
if pg_conn:
|
||||
try:
|
||||
with pg_conn.cursor(cursor_factory=RealDictCursor) as cursor:
|
||||
# Get error types distribution
|
||||
cursor.execute("""
|
||||
SELECT error_type, COUNT(*) as count
|
||||
FROM error_logs
|
||||
WHERE created_at > NOW() - INTERVAL '24 hours'
|
||||
GROUP BY error_type
|
||||
ORDER BY count DESC
|
||||
LIMIT 10
|
||||
""")
|
||||
error_types = cursor.fetchall()
|
||||
stats['error_types'] = {row['error_type']: row['count'] for row in error_types}
|
||||
|
||||
# Get error timeline (hourly for last 24 hours)
|
||||
cursor.execute("""
|
||||
SELECT
|
||||
DATE_TRUNC('hour', created_at) as hour,
|
||||
COUNT(*) as count
|
||||
FROM error_logs
|
||||
WHERE created_at > NOW() - INTERVAL '24 hours'
|
||||
GROUP BY hour
|
||||
ORDER BY hour
|
||||
""")
|
||||
timeline = cursor.fetchall()
|
||||
|
||||
for row in timeline:
|
||||
stats['error_timeline']['labels'].append(row['hour'].strftime('%H:00'))
|
||||
stats['error_timeline']['data'].append(row['count'])
|
||||
|
||||
# Get recent errors
|
||||
cursor.execute("""
|
||||
SELECT
|
||||
error_type,
|
||||
error_message,
|
||||
endpoint,
|
||||
created_at
|
||||
FROM error_logs
|
||||
ORDER BY created_at DESC
|
||||
LIMIT 10
|
||||
""")
|
||||
recent = cursor.fetchall()
|
||||
stats['recent_errors'] = [
|
||||
{
|
||||
'type': row['error_type'],
|
||||
'message': row['error_message'][:100],
|
||||
'endpoint': row['endpoint'],
|
||||
'time': row['created_at'].isoformat()
|
||||
}
|
||||
for row in recent
|
||||
]
|
||||
except Exception as e:
|
||||
logger.error(f"Error querying PostgreSQL: {e}")
|
||||
|
||||
# Fallback to Redis if PostgreSQL fails
|
||||
if not stats['error_types'] and redis_client:
|
||||
error_types = redis_client.hgetall('stats:error_types') or {}
|
||||
stats['error_types'] = {k: int(v) for k, v in error_types.items()}
|
||||
|
||||
# Get hourly error counts
|
||||
now = datetime.now()
|
||||
for i in range(23, -1, -1):
|
||||
hour_key = (now - timedelta(hours=i)).strftime('%Y-%m-%d-%H')
|
||||
count = int(redis_client.get(f'stats:errors:hourly:{hour_key}') or 0)
|
||||
stats['error_timeline']['data'].append(count)
|
||||
stats['error_timeline']['labels'].append((now - timedelta(hours=i)).strftime('%H:00'))
|
||||
|
||||
return jsonify(stats)
|
||||
except Exception as e:
|
||||
logger.error(f"Error in get_error_stats: {e}")
|
||||
return jsonify({'error': str(e)}), 500
|
||||
|
||||
@admin_bp.route('/api/stats/performance')
|
||||
@admin_required
|
||||
def get_performance_stats():
|
||||
"""Get performance metrics"""
|
||||
try:
|
||||
stats = {
|
||||
'response_times': {
|
||||
'translation': {'avg': 0, 'p95': 0, 'p99': 0},
|
||||
'transcription': {'avg': 0, 'p95': 0, 'p99': 0},
|
||||
'tts': {'avg': 0, 'p95': 0, 'p99': 0}
|
||||
},
|
||||
'throughput': {'data': [], 'labels': []},
|
||||
'slow_requests': []
|
||||
}
|
||||
|
||||
if redis_client:
|
||||
# Calculate response time percentiles
|
||||
for operation in ['translation', 'transcription', 'tts']:
|
||||
times = redis_client.lrange(f'stats:response_times:{operation}', 0, -1)
|
||||
if times:
|
||||
times = sorted([float(t) for t in times])
|
||||
stats['response_times'][operation]['avg'] = round(sum(times) / len(times), 2)
|
||||
stats['response_times'][operation]['p95'] = round(times[int(len(times) * 0.95)], 2)
|
||||
stats['response_times'][operation]['p99'] = round(times[int(len(times) * 0.99)], 2)
|
||||
|
||||
# Get throughput (requests per minute for last hour)
|
||||
now = datetime.now()
|
||||
for i in range(59, -1, -1):
|
||||
time_key = (now - timedelta(minutes=i)).strftime('%Y-%m-%d-%H-%M')
|
||||
count = int(redis_client.get(f'stats:requests:minute:{time_key}') or 0)
|
||||
stats['throughput']['data'].append(count)
|
||||
stats['throughput']['labels'].append((now - timedelta(minutes=i)).strftime('%H:%M'))
|
||||
|
||||
# Get slow requests
|
||||
slow_requests = redis_client.lrange('stats:slow_requests', 0, 9)
|
||||
stats['slow_requests'] = [json.loads(req) for req in slow_requests if req]
|
||||
|
||||
return jsonify(stats)
|
||||
except Exception as e:
|
||||
logger.error(f"Error in get_performance_stats: {e}")
|
||||
return jsonify({'error': str(e)}), 500
|
||||
|
||||
@admin_bp.route('/api/export/<data_type>')
|
||||
@admin_required
|
||||
def export_data(data_type):
|
||||
"""Export analytics data"""
|
||||
try:
|
||||
if data_type not in ['requests', 'errors', 'performance', 'all']:
|
||||
return jsonify({'error': 'Invalid data type'}), 400
|
||||
|
||||
export_data = {
|
||||
'export_time': datetime.now().isoformat(),
|
||||
'data_type': data_type
|
||||
}
|
||||
|
||||
if data_type in ['requests', 'all']:
|
||||
# Export request data
|
||||
request_data = []
|
||||
if redis_client:
|
||||
# Get daily stats for last 30 days
|
||||
now = datetime.now()
|
||||
for i in range(29, -1, -1):
|
||||
date_key = (now - timedelta(days=i)).strftime('%Y-%m-%d')
|
||||
request_data.append({
|
||||
'date': date_key,
|
||||
'requests': int(redis_client.get(f'stats:requests:daily:{date_key}') or 0),
|
||||
'translations': int(redis_client.get(f'stats:translations:daily:{date_key}') or 0),
|
||||
'transcriptions': int(redis_client.get(f'stats:transcriptions:daily:{date_key}') or 0),
|
||||
'errors': int(redis_client.get(f'stats:errors:daily:{date_key}') or 0)
|
||||
})
|
||||
export_data['requests'] = request_data
|
||||
|
||||
if data_type in ['errors', 'all']:
|
||||
# Export error data from PostgreSQL
|
||||
error_data = []
|
||||
if pg_conn:
|
||||
try:
|
||||
with pg_conn.cursor(cursor_factory=RealDictCursor) as cursor:
|
||||
cursor.execute("""
|
||||
SELECT * FROM error_logs
|
||||
WHERE created_at > NOW() - INTERVAL '7 days'
|
||||
ORDER BY created_at DESC
|
||||
""")
|
||||
errors = cursor.fetchall()
|
||||
error_data = [dict(row) for row in errors]
|
||||
except Exception as e:
|
||||
logger.error(f"Error exporting from PostgreSQL: {e}")
|
||||
export_data['errors'] = error_data
|
||||
|
||||
if data_type in ['performance', 'all']:
|
||||
# Export performance data
|
||||
perf_data = {
|
||||
'response_times': {},
|
||||
'slow_requests': []
|
||||
}
|
||||
if redis_client:
|
||||
for op in ['translation', 'transcription', 'tts']:
|
||||
times = redis_client.lrange(f'stats:response_times:{op}', 0, -1)
|
||||
perf_data['response_times'][op] = [float(t) for t in times]
|
||||
|
||||
slow_reqs = redis_client.lrange('stats:slow_requests', 0, -1)
|
||||
perf_data['slow_requests'] = [json.loads(req) for req in slow_reqs if req]
|
||||
|
||||
export_data['performance'] = perf_data
|
||||
|
||||
# Return as downloadable JSON
|
||||
response = jsonify(export_data)
|
||||
response.headers['Content-Disposition'] = f'attachment; filename=talk2me_analytics_{data_type}_{datetime.now().strftime("%Y%m%d_%H%M%S")}.json'
|
||||
return response
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error in export_data: {e}")
|
||||
return jsonify({'error': str(e)}), 500
|
||||
|
||||
def check_system_health():
|
||||
"""Check health of system components"""
|
||||
health = {
|
||||
'redis': 'unknown',
|
||||
'postgresql': 'unknown',
|
||||
'tts': 'unknown',
|
||||
'overall': 'healthy'
|
||||
}
|
||||
|
||||
# Check Redis
|
||||
if redis_client:
|
||||
try:
|
||||
redis_client.ping()
|
||||
health['redis'] = 'healthy'
|
||||
except:
|
||||
health['redis'] = 'unhealthy'
|
||||
health['overall'] = 'degraded'
|
||||
else:
|
||||
health['redis'] = 'not_configured'
|
||||
health['overall'] = 'degraded'
|
||||
|
||||
# Check PostgreSQL
|
||||
if pg_conn:
|
||||
try:
|
||||
with pg_conn.cursor() as cursor:
|
||||
cursor.execute("SELECT 1")
|
||||
cursor.fetchone()
|
||||
health['postgresql'] = 'healthy'
|
||||
except:
|
||||
health['postgresql'] = 'unhealthy'
|
||||
health['overall'] = 'degraded'
|
||||
else:
|
||||
health['postgresql'] = 'not_configured'
|
||||
health['overall'] = 'degraded'
|
||||
|
||||
# Check TTS Server
|
||||
tts_server_url = current_app.config.get('TTS_SERVER_URL')
|
||||
if tts_server_url:
|
||||
try:
|
||||
import requests
|
||||
# Extract base URL from the speech endpoint
|
||||
base_url = tts_server_url.rsplit('/v1/audio/speech', 1)[0] if '/v1/audio/speech' in tts_server_url else tts_server_url
|
||||
health_url = f"{base_url}/health" if not tts_server_url.endswith('/health') else tts_server_url
|
||||
response = requests.get(health_url, timeout=2)
|
||||
if response.status_code == 200:
|
||||
health['tts'] = 'healthy'
|
||||
health['tts_details'] = response.json() if response.headers.get('content-type') == 'application/json' else {}
|
||||
elif response.status_code == 404:
|
||||
# Try voices endpoint as fallback
|
||||
voices_url = f"{base_url}/voices" if base_url else f"{tts_server_url.rsplit('/speech', 1)[0]}/voices"
|
||||
voices_response = requests.get(voices_url, timeout=2)
|
||||
if voices_response.status_code == 200:
|
||||
health['tts'] = 'healthy'
|
||||
else:
|
||||
health['tts'] = 'unhealthy'
|
||||
health['overall'] = 'degraded'
|
||||
else:
|
||||
health['tts'] = 'unhealthy'
|
||||
health['overall'] = 'degraded'
|
||||
except requests.exceptions.RequestException:
|
||||
health['tts'] = 'unreachable'
|
||||
health['overall'] = 'degraded'
|
||||
except Exception as e:
|
||||
health['tts'] = 'error'
|
||||
health['overall'] = 'degraded'
|
||||
logger.error(f"TTS health check error: {e}")
|
||||
else:
|
||||
health['tts'] = 'not_configured'
|
||||
# TTS is optional, so don't degrade overall health
|
||||
|
||||
return health
|
||||
|
||||
# TTS Server Status endpoint
|
||||
@admin_bp.route('/api/tts/status')
|
||||
@admin_required
|
||||
def get_tts_status():
|
||||
"""Get detailed TTS server status"""
|
||||
try:
|
||||
tts_info = {
|
||||
'configured': False,
|
||||
'status': 'not_configured',
|
||||
'server_url': None,
|
||||
'api_key_configured': False,
|
||||
'details': {}
|
||||
}
|
||||
|
||||
# Check configuration
|
||||
tts_server_url = current_app.config.get('TTS_SERVER_URL')
|
||||
tts_api_key = current_app.config.get('TTS_API_KEY')
|
||||
|
||||
if tts_server_url:
|
||||
tts_info['configured'] = True
|
||||
tts_info['server_url'] = tts_server_url
|
||||
tts_info['api_key_configured'] = bool(tts_api_key)
|
||||
|
||||
# Try to get detailed status
|
||||
try:
|
||||
import requests
|
||||
headers = {}
|
||||
if tts_api_key:
|
||||
headers['Authorization'] = f'Bearer {tts_api_key}'
|
||||
|
||||
# Check health endpoint
|
||||
# Extract base URL from the speech endpoint
|
||||
base_url = tts_server_url.rsplit('/v1/audio/speech', 1)[0] if '/v1/audio/speech' in tts_server_url else tts_server_url
|
||||
health_url = f"{base_url}/health" if not tts_server_url.endswith('/health') else tts_server_url
|
||||
|
||||
response = requests.get(health_url, headers=headers, timeout=3)
|
||||
if response.status_code == 200:
|
||||
tts_info['status'] = 'healthy'
|
||||
if response.headers.get('content-type') == 'application/json':
|
||||
tts_info['details'] = response.json()
|
||||
else:
|
||||
tts_info['status'] = 'unhealthy'
|
||||
tts_info['details']['error'] = f'Health check returned status {response.status_code}'
|
||||
|
||||
# Try to get voice list
|
||||
try:
|
||||
voices_url = f"{base_url}/voices" if base_url else f"{tts_server_url.rsplit('/speech', 1)[0]}/voices"
|
||||
voices_response = requests.get(voices_url, headers=headers, timeout=3)
|
||||
if voices_response.status_code == 200 and voices_response.headers.get('content-type') == 'application/json':
|
||||
voices_data = voices_response.json()
|
||||
tts_info['details']['available_voices'] = voices_data.get('voices', [])
|
||||
tts_info['details']['voice_count'] = len(voices_data.get('voices', []))
|
||||
# If we can get voices, consider the server healthy even if health endpoint doesn't exist
|
||||
if tts_info['status'] == 'unhealthy' and response.status_code == 404:
|
||||
tts_info['status'] = 'healthy'
|
||||
tts_info['details'].pop('error', None)
|
||||
except:
|
||||
pass
|
||||
|
||||
except requests.exceptions.ConnectionError:
|
||||
tts_info['status'] = 'unreachable'
|
||||
tts_info['details']['error'] = 'Cannot connect to TTS server'
|
||||
except requests.exceptions.Timeout:
|
||||
tts_info['status'] = 'timeout'
|
||||
tts_info['details']['error'] = 'TTS server request timed out'
|
||||
except Exception as e:
|
||||
tts_info['status'] = 'error'
|
||||
tts_info['details']['error'] = str(e)
|
||||
|
||||
# Get recent TTS usage stats from Redis
|
||||
if redis_client:
|
||||
try:
|
||||
now = datetime.now()
|
||||
tts_info['usage'] = {
|
||||
'total': int(redis_client.get('stats:tts:total') or 0),
|
||||
'today': int(redis_client.get(f'stats:tts:daily:{now.strftime("%Y-%m-%d")}') or 0),
|
||||
'this_hour': int(redis_client.get(f'stats:tts:hourly:{now.strftime("%Y-%m-%d-%H")}') or 0)
|
||||
}
|
||||
|
||||
# Get recent response times
|
||||
response_times = redis_client.lrange('stats:response_times:tts', -100, -1)
|
||||
if response_times:
|
||||
times = [float(t) for t in response_times]
|
||||
tts_info['performance'] = {
|
||||
'avg_response_time': round(sum(times) / len(times), 2),
|
||||
'min_response_time': round(min(times), 2),
|
||||
'max_response_time': round(max(times), 2)
|
||||
}
|
||||
except Exception as e:
|
||||
logger.error(f"Error getting TTS stats from Redis: {e}")
|
||||
|
||||
return jsonify(tts_info)
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error in get_tts_status: {e}")
|
||||
return jsonify({'error': str(e)}), 500
|
||||
|
||||
# WebSocket support for real-time updates (using Server-Sent Events as fallback)
|
||||
@admin_bp.route('/api/stream/updates')
|
||||
@admin_required
|
||||
def stream_updates():
|
||||
"""Stream real-time updates using Server-Sent Events"""
|
||||
def generate():
|
||||
last_update = time.time()
|
||||
|
||||
while True:
|
||||
# Send update every 5 seconds
|
||||
if time.time() - last_update > 5:
|
||||
try:
|
||||
# Get current stats
|
||||
stats = {
|
||||
'timestamp': datetime.now().isoformat(),
|
||||
'requests_per_minute': 0,
|
||||
'active_sessions': 0,
|
||||
'recent_errors': 0
|
||||
}
|
||||
|
||||
if redis_client:
|
||||
# Current requests per minute
|
||||
current_minute = datetime.now().strftime('%Y-%m-%d-%H-%M')
|
||||
stats['requests_per_minute'] = int(redis_client.get(f'stats:requests:minute:{current_minute}') or 0)
|
||||
|
||||
# Active sessions
|
||||
stats['active_sessions'] = len(redis_client.keys('session:*'))
|
||||
|
||||
# Recent errors
|
||||
current_hour = datetime.now().strftime('%Y-%m-%d-%H')
|
||||
stats['recent_errors'] = int(redis_client.get(f'stats:errors:hourly:{current_hour}') or 0)
|
||||
|
||||
yield f"data: {json.dumps(stats)}\n\n"
|
||||
last_update = time.time()
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error in stream_updates: {e}")
|
||||
yield f"data: {json.dumps({'error': str(e)})}\n\n"
|
||||
|
||||
time.sleep(1)
|
||||
|
||||
return current_app.response_class(
|
||||
generate(),
|
||||
mimetype='text/event-stream',
|
||||
headers={
|
||||
'Cache-Control': 'no-cache',
|
||||
'X-Accel-Buffering': 'no'
|
||||
}
|
||||
)
|
192
admin/static/css/admin.css
Normal file
192
admin/static/css/admin.css
Normal file
@ -0,0 +1,192 @@
|
||||
/* Admin Dashboard Styles */
|
||||
|
||||
body {
|
||||
background-color: #f8f9fa;
|
||||
padding-top: 56px; /* For fixed navbar */
|
||||
}
|
||||
|
||||
/* Cards */
|
||||
.card {
|
||||
border: none;
|
||||
box-shadow: 0 0.125rem 0.25rem rgba(0, 0, 0, 0.075);
|
||||
transition: transform 0.2s;
|
||||
}
|
||||
|
||||
.card:hover {
|
||||
transform: translateY(-2px);
|
||||
box-shadow: 0 0.5rem 1rem rgba(0, 0, 0, 0.15);
|
||||
}
|
||||
|
||||
.card-header {
|
||||
background-color: #fff;
|
||||
border-bottom: 1px solid #e3e6f0;
|
||||
font-weight: 600;
|
||||
}
|
||||
|
||||
/* Status Badges */
|
||||
.badge {
|
||||
padding: 0.375rem 0.75rem;
|
||||
font-weight: normal;
|
||||
}
|
||||
|
||||
.badge.bg-success {
|
||||
background-color: #1cc88a !important;
|
||||
}
|
||||
|
||||
.badge.bg-warning {
|
||||
background-color: #f6c23e !important;
|
||||
color: #000;
|
||||
}
|
||||
|
||||
.badge.bg-danger {
|
||||
background-color: #e74a3b !important;
|
||||
}
|
||||
|
||||
/* Charts */
|
||||
canvas {
|
||||
max-width: 100%;
|
||||
}
|
||||
|
||||
/* Tables */
|
||||
.table {
|
||||
font-size: 0.875rem;
|
||||
}
|
||||
|
||||
.table th {
|
||||
font-weight: 600;
|
||||
text-transform: uppercase;
|
||||
font-size: 0.75rem;
|
||||
color: #6c757d;
|
||||
}
|
||||
|
||||
/* Login Page */
|
||||
.login-container {
|
||||
min-height: 100vh;
|
||||
display: flex;
|
||||
align-items: center;
|
||||
justify-content: center;
|
||||
}
|
||||
|
||||
/* Responsive adjustments */
|
||||
@media (max-width: 768px) {
|
||||
.card-body h2 {
|
||||
font-size: 1.5rem;
|
||||
}
|
||||
|
||||
.btn-group {
|
||||
display: flex;
|
||||
flex-direction: column;
|
||||
}
|
||||
|
||||
.btn-group .btn {
|
||||
border-radius: 0.25rem !important;
|
||||
margin: 2px 0;
|
||||
}
|
||||
}
|
||||
|
||||
/* Loading spinners */
|
||||
.spinner-border-sm {
|
||||
width: 1rem;
|
||||
height: 1rem;
|
||||
}
|
||||
|
||||
/* Error list */
|
||||
.error-item {
|
||||
padding: 0.5rem;
|
||||
border-bottom: 1px solid #dee2e6;
|
||||
}
|
||||
|
||||
.error-item:last-child {
|
||||
border-bottom: none;
|
||||
}
|
||||
|
||||
.error-type {
|
||||
font-weight: 600;
|
||||
color: #e74a3b;
|
||||
}
|
||||
|
||||
.error-time {
|
||||
font-size: 0.75rem;
|
||||
color: #6c757d;
|
||||
}
|
||||
|
||||
/* Toast notifications */
|
||||
.toast {
|
||||
background-color: white;
|
||||
box-shadow: 0 0.5rem 1rem rgba(0, 0, 0, 0.15);
|
||||
}
|
||||
|
||||
/* Animations */
|
||||
@keyframes pulse {
|
||||
0% {
|
||||
opacity: 1;
|
||||
}
|
||||
50% {
|
||||
opacity: 0.5;
|
||||
}
|
||||
100% {
|
||||
opacity: 1;
|
||||
}
|
||||
}
|
||||
|
||||
.updating {
|
||||
animation: pulse 1s infinite;
|
||||
}
|
||||
|
||||
/* Dark mode support */
|
||||
@media (prefers-color-scheme: dark) {
|
||||
body {
|
||||
background-color: #1a1a1a;
|
||||
color: #e0e0e0;
|
||||
}
|
||||
|
||||
.card {
|
||||
background-color: #2a2a2a;
|
||||
color: #e0e0e0;
|
||||
}
|
||||
|
||||
.card-header {
|
||||
background-color: #2a2a2a;
|
||||
border-bottom-color: #3a3a3a;
|
||||
}
|
||||
|
||||
.table {
|
||||
color: #e0e0e0;
|
||||
}
|
||||
|
||||
.table-striped tbody tr:nth-of-type(odd) {
|
||||
background-color: rgba(255, 255, 255, 0.05);
|
||||
}
|
||||
|
||||
.form-control {
|
||||
background-color: #3a3a3a;
|
||||
border-color: #4a4a4a;
|
||||
color: #e0e0e0;
|
||||
}
|
||||
}
|
||||
|
||||
/* Performance optimization */
|
||||
.chart-container {
|
||||
position: relative;
|
||||
height: 40vh;
|
||||
width: 100%;
|
||||
}
|
||||
|
||||
/* Scrollbar styling */
|
||||
::-webkit-scrollbar {
|
||||
width: 8px;
|
||||
height: 8px;
|
||||
}
|
||||
|
||||
::-webkit-scrollbar-track {
|
||||
background: #f1f1f1;
|
||||
}
|
||||
|
||||
::-webkit-scrollbar-thumb {
|
||||
background: #888;
|
||||
border-radius: 4px;
|
||||
}
|
||||
|
||||
::-webkit-scrollbar-thumb:hover {
|
||||
background: #555;
|
||||
}
|
261
admin/static/js/admin.js
Normal file
261
admin/static/js/admin.js
Normal file
@ -0,0 +1,261 @@
|
||||
// Admin Dashboard JavaScript
|
||||
$(document).ready(function() {
|
||||
// Load initial data
|
||||
loadOverviewStats();
|
||||
loadSystemHealth();
|
||||
loadTTSStatus();
|
||||
loadRequestChart('hour');
|
||||
loadOperationStats();
|
||||
loadLanguagePairs();
|
||||
loadRecentErrors();
|
||||
loadActiveSessions();
|
||||
|
||||
// Set up auto-refresh
|
||||
setInterval(loadOverviewStats, 30000); // Every 30 seconds
|
||||
setInterval(loadSystemHealth, 60000); // Every minute
|
||||
setInterval(loadTTSStatus, 60000); // Every minute
|
||||
|
||||
// Set up real-time updates if available
|
||||
initializeEventStream();
|
||||
});
|
||||
|
||||
// Charts
|
||||
let charts = {
|
||||
request: null,
|
||||
operations: null,
|
||||
language: null,
|
||||
performance: null,
|
||||
errors: null
|
||||
};
|
||||
|
||||
// Load overview statistics
|
||||
function loadOverviewStats() {
|
||||
$.ajax({
|
||||
url: '/admin/api/stats/overview',
|
||||
method: 'GET',
|
||||
success: function(data) {
|
||||
// Update request stats
|
||||
$('#total-requests').text(data.requests.total.toLocaleString());
|
||||
$('#today-requests').text(data.requests.today.toLocaleString());
|
||||
$('#hourly-requests').text(data.requests.hour.toLocaleString());
|
||||
|
||||
// Update operation stats
|
||||
$('#total-translations').text(data.translations.total.toLocaleString());
|
||||
$('#today-translations').text(data.translations.today.toLocaleString());
|
||||
|
||||
$('#total-transcriptions').text(data.transcriptions.total.toLocaleString());
|
||||
$('#today-transcriptions').text(data.transcriptions.today.toLocaleString());
|
||||
|
||||
// Update other metrics
|
||||
$('#active-sessions').text(data.active_sessions.toLocaleString());
|
||||
$('#error-rate').text(data.error_rate.toFixed(2) + '%');
|
||||
$('#cache-hit-rate').text(data.cache_hit_rate.toFixed(2) + '%');
|
||||
},
|
||||
error: function(xhr, status, error) {
|
||||
console.error('Failed to load overview stats:', error);
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
// Load system health status
|
||||
function loadSystemHealth() {
|
||||
$.ajax({
|
||||
url: '/admin/api/health',
|
||||
method: 'GET',
|
||||
success: function(data) {
|
||||
// Update overall status
|
||||
const overallStatus = $('#overall-status');
|
||||
overallStatus.removeClass('text-success text-warning text-danger');
|
||||
|
||||
if (data.status === 'healthy') {
|
||||
overallStatus.addClass('text-success').html('<i class="fas fa-check-circle"></i> All Systems Operational');
|
||||
} else if (data.status === 'degraded') {
|
||||
overallStatus.addClass('text-warning').html('<i class="fas fa-exclamation-triangle"></i> Degraded Performance');
|
||||
} else {
|
||||
overallStatus.addClass('text-danger').html('<i class="fas fa-times-circle"></i> System Issues');
|
||||
}
|
||||
|
||||
// Update component statuses
|
||||
updateComponentStatus('redis', data.components.redis);
|
||||
updateComponentStatus('postgresql', data.components.postgresql);
|
||||
updateComponentStatus('ml', data.components.tts || { status: 'healthy' });
|
||||
},
|
||||
error: function(xhr, status, error) {
|
||||
console.error('Failed to load system health:', error);
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
// Update component status badge
|
||||
function updateComponentStatus(component, data) {
|
||||
const badge = $(`#${component}-status`);
|
||||
badge.removeClass('bg-success bg-warning bg-danger bg-secondary');
|
||||
|
||||
if (data.status === 'healthy') {
|
||||
badge.addClass('bg-success').text('Healthy');
|
||||
} else if (data.status === 'not_configured') {
|
||||
badge.addClass('bg-secondary').text('Not Configured');
|
||||
} else if (data.status === 'unreachable') {
|
||||
badge.addClass('bg-warning').text('Unreachable');
|
||||
} else {
|
||||
badge.addClass('bg-danger').text('Unhealthy');
|
||||
}
|
||||
|
||||
// Update TTS details if applicable
|
||||
if (component === 'ml' && data.status) {
|
||||
const details = $('#tts-details');
|
||||
if (data.status === 'healthy') {
|
||||
details.text('TTS Server Connected');
|
||||
} else if (data.status === 'not_configured') {
|
||||
details.text('No TTS Server');
|
||||
} else if (data.status === 'unreachable') {
|
||||
details.text('Cannot reach TTS server');
|
||||
} else {
|
||||
details.text('TTS Server Error');
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Load detailed TTS status
|
||||
function loadTTSStatus() {
|
||||
$.ajax({
|
||||
url: '/admin/api/tts/status',
|
||||
method: 'GET',
|
||||
success: function(data) {
|
||||
// Configuration status
|
||||
if (data.configured) {
|
||||
$('#tts-config-status').removeClass().addClass('badge bg-success').text('Configured');
|
||||
$('#tts-server-url').text(data.server_url || '-');
|
||||
$('#tts-api-key-status').text(data.api_key_configured ? 'Configured' : 'Not Set');
|
||||
} else {
|
||||
$('#tts-config-status').removeClass().addClass('badge bg-secondary').text('Not Configured');
|
||||
$('#tts-server-url').text('-');
|
||||
$('#tts-api-key-status').text('-');
|
||||
}
|
||||
|
||||
// Health status
|
||||
const healthBadge = $('#tts-health-status');
|
||||
healthBadge.removeClass();
|
||||
|
||||
if (data.status === 'healthy') {
|
||||
healthBadge.addClass('badge bg-success').text('Healthy');
|
||||
$('#tts-error-message').text('-');
|
||||
} else if (data.status === 'unreachable') {
|
||||
healthBadge.addClass('badge bg-warning').text('Unreachable');
|
||||
$('#tts-error-message').text(data.details.error || 'Cannot connect');
|
||||
} else if (data.status === 'not_configured') {
|
||||
healthBadge.addClass('badge bg-secondary').text('Not Configured');
|
||||
$('#tts-error-message').text('-');
|
||||
} else {
|
||||
healthBadge.addClass('badge bg-danger').text('Error');
|
||||
$('#tts-error-message').text(data.details.error || 'Unknown error');
|
||||
}
|
||||
|
||||
// Voice count and list
|
||||
if (data.details && data.details.voice_count !== undefined) {
|
||||
$('#tts-voice-count').text(data.details.voice_count);
|
||||
|
||||
// Show voice list if available
|
||||
if (data.details.available_voices && data.details.available_voices.length > 0) {
|
||||
$('#tts-voices-container').show();
|
||||
const voicesList = $('#tts-voices-list');
|
||||
voicesList.empty();
|
||||
|
||||
data.details.available_voices.forEach(function(voice) {
|
||||
voicesList.append(`<span class="badge bg-primary">${voice}</span>`);
|
||||
});
|
||||
}
|
||||
} else {
|
||||
$('#tts-voice-count').text('-');
|
||||
$('#tts-voices-container').hide();
|
||||
}
|
||||
|
||||
// Usage statistics
|
||||
if (data.usage) {
|
||||
$('#tts-usage-today').text(data.usage.today.toLocaleString());
|
||||
$('#tts-usage-total').text(data.usage.total.toLocaleString());
|
||||
} else {
|
||||
$('#tts-usage-today').text('-');
|
||||
$('#tts-usage-total').text('-');
|
||||
}
|
||||
|
||||
// Performance metrics
|
||||
if (data.performance) {
|
||||
$('#tts-avg-response').text(data.performance.avg_response_time + ' ms');
|
||||
} else {
|
||||
$('#tts-avg-response').text('-');
|
||||
}
|
||||
},
|
||||
error: function(xhr, status, error) {
|
||||
console.error('Failed to load TTS status:', error);
|
||||
$('#tts-config-status').removeClass().addClass('badge bg-danger').text('Error');
|
||||
$('#tts-health-status').removeClass().addClass('badge bg-danger').text('Error');
|
||||
$('#tts-error-message').text('Failed to load status');
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
// Load request chart
|
||||
function loadRequestChart(timeframe) {
|
||||
// Implementation would go here
|
||||
console.log('Loading request chart for timeframe:', timeframe);
|
||||
}
|
||||
|
||||
// Load operation statistics
|
||||
function loadOperationStats() {
|
||||
// Implementation would go here
|
||||
console.log('Loading operation stats');
|
||||
}
|
||||
|
||||
// Load language pairs
|
||||
function loadLanguagePairs() {
|
||||
// Implementation would go here
|
||||
console.log('Loading language pairs');
|
||||
}
|
||||
|
||||
// Load recent errors
|
||||
function loadRecentErrors() {
|
||||
// Implementation would go here
|
||||
console.log('Loading recent errors');
|
||||
}
|
||||
|
||||
// Load active sessions
|
||||
function loadActiveSessions() {
|
||||
// Implementation would go here
|
||||
console.log('Loading active sessions');
|
||||
}
|
||||
|
||||
// Initialize event stream for real-time updates
|
||||
function initializeEventStream() {
|
||||
if (typeof(EventSource) === "undefined") {
|
||||
console.log("Server-sent events not supported");
|
||||
return;
|
||||
}
|
||||
|
||||
const source = new EventSource('/admin/api/stream/updates');
|
||||
|
||||
source.onmessage = function(event) {
|
||||
const data = JSON.parse(event.data);
|
||||
|
||||
// Update real-time metrics
|
||||
if (data.requests_per_minute !== undefined) {
|
||||
$('#realtime-rpm').text(data.requests_per_minute);
|
||||
}
|
||||
if (data.active_sessions !== undefined) {
|
||||
$('#active-sessions').text(data.active_sessions);
|
||||
}
|
||||
if (data.recent_errors !== undefined) {
|
||||
$('#recent-errors-count').text(data.recent_errors);
|
||||
}
|
||||
};
|
||||
|
||||
source.onerror = function(error) {
|
||||
console.error('EventSource error:', error);
|
||||
};
|
||||
}
|
||||
|
||||
// Show toast notification
|
||||
function showToast(message, type = 'info') {
|
||||
// Implementation would go here
|
||||
console.log(`Toast [${type}]: ${message}`);
|
||||
}
|
75
admin/templates/base.html
Normal file
75
admin/templates/base.html
Normal file
@ -0,0 +1,75 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en">
|
||||
<head>
|
||||
<meta charset="UTF-8">
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||||
<title>{% block title %}Talk2Me Admin Dashboard{% endblock %}</title>
|
||||
|
||||
<!-- Bootstrap CSS -->
|
||||
<link href="https://cdn.jsdelivr.net/npm/bootstrap@5.3.0/dist/css/bootstrap.min.css" rel="stylesheet">
|
||||
|
||||
<!-- Font Awesome -->
|
||||
<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.4.0/css/all.min.css">
|
||||
|
||||
<!-- Chart.js -->
|
||||
<script src="https://cdn.jsdelivr.net/npm/chart.js@4.4.0/dist/chart.umd.js"></script>
|
||||
|
||||
<!-- Custom CSS -->
|
||||
<link rel="stylesheet" href="{{ url_for('admin.static', filename='css/admin.css') }}">
|
||||
|
||||
{% block extra_css %}{% endblock %}
|
||||
</head>
|
||||
<body>
|
||||
<!-- Navigation -->
|
||||
<nav class="navbar navbar-expand-lg navbar-dark bg-dark fixed-top">
|
||||
<div class="container-fluid">
|
||||
<a class="navbar-brand" href="{{ url_for('admin.dashboard') }}">
|
||||
<i class="fas fa-language"></i> Talk2Me Admin
|
||||
</a>
|
||||
<button class="navbar-toggler" type="button" data-bs-toggle="collapse" data-bs-target="#navbarNav">
|
||||
<span class="navbar-toggler-icon"></span>
|
||||
</button>
|
||||
<div class="collapse navbar-collapse" id="navbarNav">
|
||||
<ul class="navbar-nav ms-auto">
|
||||
<li class="nav-item">
|
||||
<a class="nav-link" href="{{ url_for('admin.dashboard') }}">
|
||||
<i class="fas fa-tachometer-alt"></i> Dashboard
|
||||
</a>
|
||||
</li>
|
||||
<li class="nav-item">
|
||||
<a class="nav-link" href="{{ url_for('admin.users') }}">
|
||||
<i class="fas fa-users"></i> Users
|
||||
</a>
|
||||
</li>
|
||||
<li class="nav-item">
|
||||
<a class="nav-link" href="#" onclick="exportData('all')">
|
||||
<i class="fas fa-download"></i> Export Data
|
||||
</a>
|
||||
</li>
|
||||
<li class="nav-item">
|
||||
<a class="nav-link" href="{{ url_for('admin.logout') }}">
|
||||
<i class="fas fa-sign-out-alt"></i> Logout
|
||||
</a>
|
||||
</li>
|
||||
</ul>
|
||||
</div>
|
||||
</div>
|
||||
</nav>
|
||||
|
||||
<!-- Main Content -->
|
||||
<main class="container-fluid mt-5 pt-3">
|
||||
{% block content %}{% endblock %}
|
||||
</main>
|
||||
|
||||
<!-- Bootstrap JS -->
|
||||
<script src="https://cdn.jsdelivr.net/npm/bootstrap@5.3.0/dist/js/bootstrap.bundle.min.js"></script>
|
||||
|
||||
<!-- jQuery (for AJAX) -->
|
||||
<script src="https://code.jquery.com/jquery-3.7.0.min.js"></script>
|
||||
|
||||
<!-- Custom JS -->
|
||||
<script src="{{ url_for('admin.static', filename='js/admin.js') }}"></script>
|
||||
|
||||
{% block extra_js %}{% endblock %}
|
||||
</body>
|
||||
</html>
|
326
admin/templates/dashboard.html
Normal file
326
admin/templates/dashboard.html
Normal file
@ -0,0 +1,326 @@
|
||||
{% extends "base.html" %}
|
||||
|
||||
{% block title %}Dashboard - Talk2Me Admin{% endblock %}
|
||||
|
||||
{% block content %}
|
||||
<!-- Quick Actions -->
|
||||
<div class="row mb-4">
|
||||
<div class="col-12">
|
||||
<div class="card bg-light">
|
||||
<div class="card-body">
|
||||
<h5 class="card-title">Quick Actions</h5>
|
||||
<div class="btn-group" role="group">
|
||||
<a href="{{ url_for('admin.users') }}" class="btn btn-primary">
|
||||
<i class="fas fa-users"></i> Manage Users
|
||||
</a>
|
||||
<button onclick="exportData('all')" class="btn btn-secondary">
|
||||
<i class="fas fa-download"></i> Export Data
|
||||
</button>
|
||||
<button onclick="clearCache()" class="btn btn-warning">
|
||||
<i class="fas fa-trash"></i> Clear Cache
|
||||
</button>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Overview Cards -->
|
||||
<div class="row mb-4">
|
||||
<div class="col-md-3 mb-3">
|
||||
<div class="card text-white bg-primary">
|
||||
<div class="card-body">
|
||||
<h5 class="card-title">Total Requests</h5>
|
||||
<h2 class="card-text" id="total-requests">
|
||||
<div class="spinner-border spinner-border-sm" role="status"></div>
|
||||
</h2>
|
||||
<p class="card-text"><small>Today: <span id="today-requests">-</span></small></p>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="col-md-3 mb-3">
|
||||
<div class="card text-white bg-success">
|
||||
<div class="card-body">
|
||||
<h5 class="card-title">Active Sessions</h5>
|
||||
<h2 class="card-text" id="active-sessions">
|
||||
<div class="spinner-border spinner-border-sm" role="status"></div>
|
||||
</h2>
|
||||
<p class="card-text"><small>Live users</small></p>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="col-md-3 mb-3">
|
||||
<div class="card text-white bg-warning">
|
||||
<div class="card-body">
|
||||
<h5 class="card-title">Error Rate</h5>
|
||||
<h2 class="card-text" id="error-rate">
|
||||
<div class="spinner-border spinner-border-sm" role="status"></div>
|
||||
</h2>
|
||||
<p class="card-text"><small>Last 24 hours</small></p>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="col-md-3 mb-3">
|
||||
<div class="card text-white bg-info">
|
||||
<div class="card-body">
|
||||
<h5 class="card-title">Cache Hit Rate</h5>
|
||||
<h2 class="card-text" id="cache-hit-rate">
|
||||
<div class="spinner-border spinner-border-sm" role="status"></div>
|
||||
</h2>
|
||||
<p class="card-text"><small>Performance metric</small></p>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- System Health Status -->
|
||||
<div class="row mb-4">
|
||||
<div class="col-12">
|
||||
<div class="card">
|
||||
<div class="card-header">
|
||||
<h5 class="mb-0"><i class="fas fa-heartbeat"></i> System Health</h5>
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<div class="row">
|
||||
<div class="col-md-4">
|
||||
<div class="d-flex align-items-center">
|
||||
<i class="fas fa-database fa-2x me-3"></i>
|
||||
<div>
|
||||
<h6 class="mb-0">Redis</h6>
|
||||
<span class="badge" id="redis-status">Checking...</span>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
<div class="col-md-4">
|
||||
<div class="d-flex align-items-center">
|
||||
<i class="fas fa-server fa-2x me-3"></i>
|
||||
<div>
|
||||
<h6 class="mb-0">PostgreSQL</h6>
|
||||
<span class="badge" id="postgresql-status">Checking...</span>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
<div class="col-md-4">
|
||||
<div class="d-flex align-items-center">
|
||||
<i class="fas fa-microphone fa-2x me-3"></i>
|
||||
<div>
|
||||
<h6 class="mb-0">Whisper/TTS</h6>
|
||||
<span class="badge" id="ml-status">Checking...</span>
|
||||
<small class="d-block text-muted" id="tts-details"></small>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- TTS Server Status Card -->
|
||||
<div class="row mb-4">
|
||||
<div class="col-md-12">
|
||||
<div class="card">
|
||||
<div class="card-header d-flex justify-content-between align-items-center">
|
||||
<h5 class="mb-0">TTS Server Status</h5>
|
||||
<button class="btn btn-sm btn-outline-primary" onclick="loadTTSStatus()">
|
||||
<i class="fas fa-sync"></i> Refresh
|
||||
</button>
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<div class="row" id="tts-status-container">
|
||||
<div class="col-md-4">
|
||||
<h6>Configuration</h6>
|
||||
<ul class="list-unstyled mb-0">
|
||||
<li><strong>Status:</strong> <span id="tts-config-status" class="badge">Loading...</span></li>
|
||||
<li><strong>Server URL:</strong> <span id="tts-server-url">-</span></li>
|
||||
<li><strong>API Key:</strong> <span id="tts-api-key-status">-</span></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="col-md-4">
|
||||
<h6>Server Health</h6>
|
||||
<ul class="list-unstyled mb-0">
|
||||
<li><strong>Health:</strong> <span id="tts-health-status" class="badge">Loading...</span></li>
|
||||
<li><strong>Available Voices:</strong> <span id="tts-voice-count">-</span></li>
|
||||
<li><strong>Error:</strong> <span id="tts-error-message" class="text-danger">-</span></li>
|
||||
</ul>
|
||||
</div>
|
||||
<div class="col-md-4">
|
||||
<h6>Usage & Performance</h6>
|
||||
<ul class="list-unstyled mb-0">
|
||||
<li><strong>Today's Requests:</strong> <span id="tts-usage-today">-</span></li>
|
||||
<li><strong>Avg Response Time:</strong> <span id="tts-avg-response">-</span></li>
|
||||
<li><strong>Total Requests:</strong> <span id="tts-usage-total">-</span></li>
|
||||
</ul>
|
||||
</div>
|
||||
</div>
|
||||
<div class="row mt-3" id="tts-voices-container" style="display: none;">
|
||||
<div class="col-md-12">
|
||||
<h6>Available Voices</h6>
|
||||
<div id="tts-voices-list" class="d-flex flex-wrap gap-2"></div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Charts Row 1 -->
|
||||
<div class="row mb-4">
|
||||
<div class="col-md-8">
|
||||
<div class="card">
|
||||
<div class="card-header">
|
||||
<h5 class="mb-0">Request Volume</h5>
|
||||
<div class="btn-group btn-group-sm float-end" role="group">
|
||||
<button type="button" class="btn btn-outline-primary active" onclick="updateRequestChart('minute')">Minute</button>
|
||||
<button type="button" class="btn btn-outline-primary" onclick="updateRequestChart('hour')">Hour</button>
|
||||
<button type="button" class="btn btn-outline-primary" onclick="updateRequestChart('day')">Day</button>
|
||||
</div>
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<canvas id="requestChart" height="100"></canvas>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="col-md-4">
|
||||
<div class="card">
|
||||
<div class="card-header">
|
||||
<h5 class="mb-0">Language Pairs</h5>
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<canvas id="languageChart" height="200"></canvas>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Charts Row 2 -->
|
||||
<div class="row mb-4">
|
||||
<div class="col-md-6">
|
||||
<div class="card">
|
||||
<div class="card-header">
|
||||
<h5 class="mb-0">Operations</h5>
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<canvas id="operationsChart" height="120"></canvas>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="col-md-6">
|
||||
<div class="card">
|
||||
<div class="card-header">
|
||||
<h5 class="mb-0">Response Times (ms)</h5>
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<canvas id="responseTimeChart" height="120"></canvas>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Error Analysis -->
|
||||
<div class="row mb-4">
|
||||
<div class="col-md-6">
|
||||
<div class="card">
|
||||
<div class="card-header">
|
||||
<h5 class="mb-0">Error Types</h5>
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<canvas id="errorTypeChart" height="150"></canvas>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="col-md-6">
|
||||
<div class="card">
|
||||
<div class="card-header">
|
||||
<h5 class="mb-0">Recent Errors</h5>
|
||||
</div>
|
||||
<div class="card-body" style="max-height: 300px; overflow-y: auto;">
|
||||
<div id="recent-errors-list">
|
||||
<div class="text-center">
|
||||
<div class="spinner-border" role="status"></div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Performance Metrics -->
|
||||
<div class="row mb-4">
|
||||
<div class="col-12">
|
||||
<div class="card">
|
||||
<div class="card-header">
|
||||
<h5 class="mb-0">Performance Metrics</h5>
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<div class="table-responsive">
|
||||
<table class="table table-striped">
|
||||
<thead>
|
||||
<tr>
|
||||
<th>Operation</th>
|
||||
<th>Average (ms)</th>
|
||||
<th>95th Percentile (ms)</th>
|
||||
<th>99th Percentile (ms)</th>
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody id="performance-table">
|
||||
<tr>
|
||||
<td colspan="4" class="text-center">
|
||||
<div class="spinner-border" role="status"></div>
|
||||
</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Real-time Updates Status -->
|
||||
<div class="position-fixed bottom-0 end-0 p-3" style="z-index: 1000">
|
||||
<div class="toast" id="update-toast" role="alert">
|
||||
<div class="toast-header">
|
||||
<i class="fas fa-sync-alt me-2"></i>
|
||||
<strong class="me-auto">Real-time Updates</strong>
|
||||
<small id="last-update">Just now</small>
|
||||
<button type="button" class="btn-close" data-bs-dismiss="toast"></button>
|
||||
</div>
|
||||
<div class="toast-body">
|
||||
<span id="update-status">Connected</span>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
{% endblock %}
|
||||
|
||||
{% block extra_js %}
|
||||
<script>
|
||||
// Initialize dashboard
|
||||
$(document).ready(function() {
|
||||
initializeDashboard();
|
||||
|
||||
// Start real-time updates
|
||||
startRealtimeUpdates();
|
||||
|
||||
// Load initial data
|
||||
loadOverviewStats();
|
||||
loadRequestChart('minute');
|
||||
loadOperationStats();
|
||||
loadErrorStats();
|
||||
loadPerformanceStats();
|
||||
|
||||
// Refresh data periodically
|
||||
setInterval(loadOverviewStats, 10000); // Every 10 seconds
|
||||
setInterval(function() {
|
||||
loadRequestChart(currentTimeframe);
|
||||
}, 30000); // Every 30 seconds
|
||||
});
|
||||
</script>
|
||||
{% endblock %}
|
120
admin/templates/dashboard_simple.html
Normal file
120
admin/templates/dashboard_simple.html
Normal file
@ -0,0 +1,120 @@
|
||||
{% extends "base.html" %}
|
||||
|
||||
{% block title %}Dashboard - Talk2Me Admin{% endblock %}
|
||||
|
||||
{% block content %}
|
||||
<!-- Simple Mode Notice -->
|
||||
<div class="alert alert-warning" role="alert">
|
||||
<h4 class="alert-heading">Simple Mode Active</h4>
|
||||
<p>The admin dashboard is running in simple mode because Redis and PostgreSQL services are not available.</p>
|
||||
<hr>
|
||||
<p class="mb-0">To enable full analytics and monitoring features, please ensure Redis and PostgreSQL are running.</p>
|
||||
</div>
|
||||
|
||||
<!-- Basic Info Cards -->
|
||||
<div class="row mb-4">
|
||||
<div class="col-md-4 mb-3">
|
||||
<div class="card">
|
||||
<div class="card-body">
|
||||
<h5 class="card-title">System Status</h5>
|
||||
<p class="card-text">
|
||||
<span class="badge badge-success">Online</span>
|
||||
</p>
|
||||
<small class="text-muted">Talk2Me API is running</small>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="col-md-4 mb-3">
|
||||
<div class="card">
|
||||
<div class="card-body">
|
||||
<h5 class="card-title">Admin Access</h5>
|
||||
<p class="card-text">
|
||||
<span class="badge badge-primary">Authenticated</span>
|
||||
</p>
|
||||
<small class="text-muted">You are logged in as admin</small>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="col-md-4 mb-3">
|
||||
<div class="card">
|
||||
<div class="card-body">
|
||||
<h5 class="card-title">Services</h5>
|
||||
<p class="card-text">
|
||||
Redis: <span class="badge badge-secondary">Not configured</span><br>
|
||||
PostgreSQL: <span class="badge badge-secondary">Not configured</span>
|
||||
</p>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Available Actions -->
|
||||
<div class="card">
|
||||
<div class="card-header">
|
||||
<h5 class="mb-0">Available Actions</h5>
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<p>In simple mode, you can:</p>
|
||||
<ul>
|
||||
<li>Access the Talk2Me API with admin privileges</li>
|
||||
<li>View system health status</li>
|
||||
<li>Logout from the admin session</li>
|
||||
</ul>
|
||||
|
||||
<p class="mt-3">To enable full features, set up the following services:</p>
|
||||
<ol>
|
||||
<li><strong>Redis</strong>: For caching, rate limiting, and session management</li>
|
||||
<li><strong>PostgreSQL</strong>: For persistent storage of analytics and user data</li>
|
||||
</ol>
|
||||
|
||||
<div class="mt-4">
|
||||
<a href="/admin/logout" class="btn btn-secondary">Logout</a>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Setup Instructions -->
|
||||
<div class="card mt-4">
|
||||
<div class="card-header">
|
||||
<h5 class="mb-0">Quick Setup Guide</h5>
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<h6>1. Install Redis:</h6>
|
||||
<pre class="bg-light p-2"><code># Ubuntu/Debian
|
||||
sudo apt-get install redis-server
|
||||
sudo systemctl start redis
|
||||
|
||||
# macOS
|
||||
brew install redis
|
||||
brew services start redis</code></pre>
|
||||
|
||||
<h6>2. Install PostgreSQL:</h6>
|
||||
<pre class="bg-light p-2"><code># Ubuntu/Debian
|
||||
sudo apt-get install postgresql
|
||||
sudo systemctl start postgresql
|
||||
|
||||
# macOS
|
||||
brew install postgresql
|
||||
brew services start postgresql</code></pre>
|
||||
|
||||
<h6>3. Configure Environment:</h6>
|
||||
<pre class="bg-light p-2"><code># Add to .env file
|
||||
REDIS_URL=redis://localhost:6379/0
|
||||
DATABASE_URL=postgresql://user:pass@localhost/talk2me</code></pre>
|
||||
|
||||
<h6>4. Initialize Database:</h6>
|
||||
<pre class="bg-light p-2"><code>python init_auth_db.py</code></pre>
|
||||
|
||||
<p class="mt-3">After completing these steps, restart the Talk2Me server to enable full admin features.</p>
|
||||
</div>
|
||||
</div>
|
||||
{% endblock %}
|
||||
|
||||
{% block scripts %}
|
||||
<script>
|
||||
// Simple mode - no need for real-time updates or API calls
|
||||
console.log('Admin dashboard loaded in simple mode');
|
||||
</script>
|
||||
{% endblock %}
|
35
admin/templates/login.html
Normal file
35
admin/templates/login.html
Normal file
@ -0,0 +1,35 @@
|
||||
{% extends "base.html" %}
|
||||
|
||||
{% block title %}Admin Login - Talk2Me{% endblock %}
|
||||
|
||||
{% block content %}
|
||||
<div class="row justify-content-center mt-5">
|
||||
<div class="col-md-6 col-lg-4">
|
||||
<div class="card shadow">
|
||||
<div class="card-body">
|
||||
<h3 class="card-title text-center mb-4">
|
||||
<i class="fas fa-lock"></i> Admin Login
|
||||
</h3>
|
||||
|
||||
{% if error %}
|
||||
<div class="alert alert-danger" role="alert">
|
||||
{{ error }}
|
||||
</div>
|
||||
{% endif %}
|
||||
|
||||
<form method="POST">
|
||||
<div class="mb-3">
|
||||
<label for="token" class="form-label">Admin Token</label>
|
||||
<input type="password" class="form-control" id="token" name="token" required autofocus>
|
||||
<div class="form-text">Enter your admin access token</div>
|
||||
</div>
|
||||
|
||||
<button type="submit" class="btn btn-primary w-100">
|
||||
<i class="fas fa-sign-in-alt"></i> Login
|
||||
</button>
|
||||
</form>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
{% endblock %}
|
47
admin_loader.py
Normal file
47
admin_loader.py
Normal file
@ -0,0 +1,47 @@
|
||||
"""
|
||||
Dynamic admin module loader that chooses between full and simple admin based on service availability
|
||||
"""
|
||||
import os
|
||||
import logging
|
||||
from dotenv import load_dotenv
|
||||
|
||||
# Load environment variables
|
||||
load_dotenv()
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
def load_admin_module():
|
||||
"""
|
||||
Dynamically load admin module based on service availability
|
||||
Returns (admin_bp, init_admin) tuple
|
||||
"""
|
||||
# Check if we should force simple mode
|
||||
if os.environ.get('ADMIN_SIMPLE_MODE', '').lower() in ('1', 'true', 'yes'):
|
||||
logger.info("Simple admin mode forced by environment variable")
|
||||
from admin_simple import admin_bp, init_admin
|
||||
return admin_bp, init_admin
|
||||
|
||||
# Try to import full admin module
|
||||
try:
|
||||
# Quick check for Redis
|
||||
import redis
|
||||
r = redis.Redis.from_url(os.environ.get('REDIS_URL', 'redis://localhost:6379/0'))
|
||||
r.ping()
|
||||
|
||||
# Quick check for PostgreSQL
|
||||
from sqlalchemy import create_engine, text
|
||||
db_url = os.environ.get('DATABASE_URL', 'postgresql://localhost/talk2me')
|
||||
engine = create_engine(db_url, pool_pre_ping=True)
|
||||
with engine.connect() as conn:
|
||||
conn.execute(text("SELECT 1"))
|
||||
|
||||
# If we get here, both services are available
|
||||
from admin import admin_bp, init_admin
|
||||
logger.info("Using full admin module with Redis and PostgreSQL support")
|
||||
return admin_bp, init_admin
|
||||
|
||||
except Exception as e:
|
||||
logger.warning(f"Cannot use full admin module: {e}")
|
||||
logger.info("Falling back to simple admin module")
|
||||
from admin_simple import admin_bp, init_admin
|
||||
return admin_bp, init_admin
|
77
admin_simple.py
Normal file
77
admin_simple.py
Normal file
@ -0,0 +1,77 @@
|
||||
"""
|
||||
Simple admin blueprint that works without Redis/PostgreSQL
|
||||
"""
|
||||
from flask import Blueprint, request, jsonify, render_template, redirect, url_for, session
|
||||
from functools import wraps
|
||||
import os
|
||||
import logging
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
# Create admin blueprint
|
||||
admin_bp = Blueprint('admin', __name__,
|
||||
template_folder='admin/templates',
|
||||
static_folder='admin/static',
|
||||
static_url_path='/admin/static')
|
||||
|
||||
def init_admin(app):
|
||||
"""Initialize admin module with app configuration"""
|
||||
logger.info("Admin dashboard initialized (simple mode)")
|
||||
|
||||
def admin_required(f):
|
||||
"""Decorator to require admin authentication"""
|
||||
@wraps(f)
|
||||
def decorated_function(*args, **kwargs):
|
||||
# Check if user is logged in as admin
|
||||
if not session.get('admin_logged_in'):
|
||||
# Check for admin token in headers (for API access)
|
||||
auth_header = request.headers.get('Authorization', '')
|
||||
if auth_header.startswith('Bearer '):
|
||||
token = auth_header[7:]
|
||||
expected_token = os.environ.get('ADMIN_TOKEN', 'default-admin-token')
|
||||
if token == expected_token:
|
||||
return f(*args, **kwargs)
|
||||
|
||||
# Redirect to login for web access
|
||||
return redirect(url_for('admin.login', next=request.url))
|
||||
return f(*args, **kwargs)
|
||||
return decorated_function
|
||||
|
||||
@admin_bp.route('/')
|
||||
@admin_required
|
||||
def dashboard():
|
||||
"""Main admin dashboard"""
|
||||
# Use simple dashboard template
|
||||
return render_template('dashboard_simple.html')
|
||||
|
||||
@admin_bp.route('/login', methods=['GET', 'POST'])
|
||||
def login():
|
||||
"""Admin login page"""
|
||||
if request.method == 'POST':
|
||||
token = request.form.get('token', '')
|
||||
expected_token = os.environ.get('ADMIN_TOKEN', 'default-admin-token')
|
||||
|
||||
if token == expected_token:
|
||||
session['admin_logged_in'] = True
|
||||
next_page = request.args.get('next', url_for('admin.dashboard'))
|
||||
return redirect(next_page)
|
||||
else:
|
||||
return render_template('login.html', error='Invalid admin token')
|
||||
|
||||
return render_template('login.html')
|
||||
|
||||
@admin_bp.route('/logout')
|
||||
def logout():
|
||||
"""Admin logout"""
|
||||
session.pop('admin_logged_in', None)
|
||||
return redirect(url_for('admin.login'))
|
||||
|
||||
@admin_bp.route('/health')
|
||||
def health():
|
||||
"""Check admin dashboard health"""
|
||||
return jsonify({
|
||||
'status': 'ok',
|
||||
'mode': 'simple',
|
||||
'redis': 'not configured',
|
||||
'postgresql': 'not configured'
|
||||
})
|
426
analytics_middleware.py
Normal file
426
analytics_middleware.py
Normal file
@ -0,0 +1,426 @@
|
||||
"""Analytics middleware for tracking requests and operations"""
|
||||
|
||||
import time
|
||||
import json
|
||||
import logging
|
||||
from datetime import datetime
|
||||
from flask import request, g
|
||||
import redis
|
||||
import psycopg2
|
||||
from psycopg2.extras import RealDictCursor
|
||||
import threading
|
||||
from queue import Queue
|
||||
from functools import wraps
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
class AnalyticsTracker:
|
||||
"""Track and store analytics data"""
|
||||
|
||||
def __init__(self, app=None):
|
||||
self.app = app
|
||||
self.redis_client = None
|
||||
self.pg_conn = None
|
||||
self.write_queue = Queue()
|
||||
self.writer_thread = None
|
||||
|
||||
if app:
|
||||
self.init_app(app)
|
||||
|
||||
def init_app(self, app):
|
||||
"""Initialize analytics with Flask app"""
|
||||
self.app = app
|
||||
|
||||
# Initialize Redis connection
|
||||
try:
|
||||
self.redis_client = redis.from_url(
|
||||
app.config.get('REDIS_URL', 'redis://localhost:6379/0'),
|
||||
decode_responses=True
|
||||
)
|
||||
self.redis_client.ping()
|
||||
logger.info("Analytics Redis connection established")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to connect to Redis for analytics: {e}")
|
||||
self.redis_client = None
|
||||
|
||||
# Initialize PostgreSQL connection
|
||||
try:
|
||||
self.pg_conn = psycopg2.connect(
|
||||
app.config.get('DATABASE_URL', 'postgresql://localhost/talk2me')
|
||||
)
|
||||
self.pg_conn.autocommit = True
|
||||
logger.info("Analytics PostgreSQL connection established")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to connect to PostgreSQL for analytics: {e}")
|
||||
self.pg_conn = None
|
||||
|
||||
# Start background writer thread
|
||||
self.writer_thread = threading.Thread(target=self._write_worker, daemon=True)
|
||||
self.writer_thread.start()
|
||||
|
||||
# Register before/after request handlers
|
||||
app.before_request(self.before_request)
|
||||
app.after_request(self.after_request)
|
||||
|
||||
def before_request(self):
|
||||
"""Track request start time"""
|
||||
g.start_time = time.time()
|
||||
g.request_size = request.content_length or 0
|
||||
|
||||
def after_request(self, response):
|
||||
"""Track request completion and metrics"""
|
||||
try:
|
||||
# Skip if analytics is disabled
|
||||
if not self.enabled:
|
||||
return response
|
||||
|
||||
# Calculate response time
|
||||
response_time = int((time.time() - g.start_time) * 1000) # in ms
|
||||
|
||||
# Track in Redis for real-time stats
|
||||
if self.redis_client:
|
||||
self._track_redis_stats(request, response, response_time)
|
||||
|
||||
# Queue for PostgreSQL logging
|
||||
if self.pg_conn and request.endpoint not in ['static', 'admin.static']:
|
||||
self._queue_request_log(request, response, response_time)
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error in analytics after_request: {e}")
|
||||
|
||||
return response
|
||||
|
||||
def _track_redis_stats(self, request, response, response_time):
|
||||
"""Track statistics in Redis"""
|
||||
try:
|
||||
now = datetime.now()
|
||||
|
||||
# Increment request counters
|
||||
pipe = self.redis_client.pipeline()
|
||||
|
||||
# Total requests
|
||||
pipe.incr('stats:requests:total')
|
||||
|
||||
# Time-based counters
|
||||
pipe.incr(f'stats:requests:minute:{now.strftime("%Y-%m-%d-%H-%M")}')
|
||||
pipe.expire(f'stats:requests:minute:{now.strftime("%Y-%m-%d-%H-%M")}', 3600) # 1 hour
|
||||
|
||||
pipe.incr(f'stats:requests:hourly:{now.strftime("%Y-%m-%d-%H")}')
|
||||
pipe.expire(f'stats:requests:hourly:{now.strftime("%Y-%m-%d-%H")}', 86400) # 24 hours
|
||||
|
||||
pipe.incr(f'stats:requests:daily:{now.strftime("%Y-%m-%d")}')
|
||||
pipe.expire(f'stats:requests:daily:{now.strftime("%Y-%m-%d")}', 604800) # 7 days
|
||||
|
||||
# Track errors
|
||||
if response.status_code >= 400:
|
||||
pipe.incr(f'stats:errors:daily:{now.strftime("%Y-%m-%d")}')
|
||||
pipe.incr(f'stats:errors:hourly:{now.strftime("%Y-%m-%d-%H")}')
|
||||
pipe.expire(f'stats:errors:hourly:{now.strftime("%Y-%m-%d-%H")}', 86400)
|
||||
|
||||
# Track response times
|
||||
endpoint_key = request.endpoint or 'unknown'
|
||||
pipe.lpush(f'stats:response_times:{endpoint_key}', response_time)
|
||||
pipe.ltrim(f'stats:response_times:{endpoint_key}', 0, 999) # Keep last 1000
|
||||
|
||||
# Track slow requests
|
||||
if response_time > 1000: # Over 1 second
|
||||
slow_request = {
|
||||
'endpoint': request.endpoint,
|
||||
'method': request.method,
|
||||
'response_time': response_time,
|
||||
'timestamp': now.isoformat()
|
||||
}
|
||||
pipe.lpush('stats:slow_requests', json.dumps(slow_request))
|
||||
pipe.ltrim('stats:slow_requests', 0, 99) # Keep last 100
|
||||
|
||||
pipe.execute()
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error tracking Redis stats: {e}")
|
||||
|
||||
def _queue_request_log(self, request, response, response_time):
|
||||
"""Queue request log for PostgreSQL"""
|
||||
try:
|
||||
log_entry = {
|
||||
'endpoint': request.endpoint,
|
||||
'method': request.method,
|
||||
'status_code': response.status_code,
|
||||
'response_time_ms': response_time,
|
||||
'ip_address': request.remote_addr,
|
||||
'user_agent': request.headers.get('User-Agent', '')[:500],
|
||||
'request_size_bytes': g.get('request_size', 0),
|
||||
'response_size_bytes': len(response.get_data()),
|
||||
'session_id': g.get('session_id'),
|
||||
'created_at': datetime.now()
|
||||
}
|
||||
|
||||
self.write_queue.put(('request_log', log_entry))
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error queuing request log: {e}")
|
||||
|
||||
def track_operation(self, operation_type, **kwargs):
|
||||
"""Track specific operations (translation, transcription, etc.)"""
|
||||
def decorator(f):
|
||||
@wraps(f)
|
||||
def wrapped(*args, **inner_kwargs):
|
||||
start_time = time.time()
|
||||
success = True
|
||||
error_message = None
|
||||
result = None
|
||||
|
||||
try:
|
||||
result = f(*args, **inner_kwargs)
|
||||
return result
|
||||
except Exception as e:
|
||||
success = False
|
||||
error_message = str(e)
|
||||
raise
|
||||
finally:
|
||||
# Track operation
|
||||
response_time = int((time.time() - start_time) * 1000)
|
||||
self._track_operation_complete(
|
||||
operation_type, response_time, success,
|
||||
error_message, kwargs, result
|
||||
)
|
||||
|
||||
return wrapped
|
||||
return decorator
|
||||
|
||||
def _track_operation_complete(self, operation_type, response_time, success,
|
||||
error_message, metadata, result):
|
||||
"""Track operation completion"""
|
||||
try:
|
||||
now = datetime.now()
|
||||
|
||||
# Update Redis counters
|
||||
if self.redis_client:
|
||||
pipe = self.redis_client.pipeline()
|
||||
|
||||
# Operation counters
|
||||
pipe.incr(f'stats:{operation_type}:total')
|
||||
pipe.incr(f'stats:{operation_type}:daily:{now.strftime("%Y-%m-%d")}')
|
||||
pipe.expire(f'stats:{operation_type}:daily:{now.strftime("%Y-%m-%d")}', 604800)
|
||||
|
||||
# Response times
|
||||
pipe.lpush(f'stats:response_times:{operation_type}', response_time)
|
||||
pipe.ltrim(f'stats:response_times:{operation_type}', 0, 999)
|
||||
|
||||
# Language pairs for translations
|
||||
if operation_type == 'translations' and 'source_lang' in metadata:
|
||||
lang_pair = f"{metadata.get('source_lang')} -> {metadata.get('target_lang')}"
|
||||
pipe.hincrby('stats:language_pairs', lang_pair, 1)
|
||||
|
||||
# Error tracking
|
||||
if not success:
|
||||
pipe.hincrby('stats:error_types', error_message[:100], 1)
|
||||
|
||||
pipe.execute()
|
||||
|
||||
# Queue for PostgreSQL
|
||||
if self.pg_conn:
|
||||
log_entry = {
|
||||
'operation_type': operation_type,
|
||||
'response_time_ms': response_time,
|
||||
'success': success,
|
||||
'error_message': error_message,
|
||||
'metadata': metadata,
|
||||
'result': result,
|
||||
'session_id': g.get('session_id'),
|
||||
'created_at': now
|
||||
}
|
||||
|
||||
self.write_queue.put((operation_type, log_entry))
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error tracking operation: {e}")
|
||||
|
||||
def _write_worker(self):
|
||||
"""Background worker to write logs to PostgreSQL"""
|
||||
while True:
|
||||
try:
|
||||
# Get items from queue (blocking)
|
||||
operation_type, log_entry = self.write_queue.get()
|
||||
|
||||
if operation_type == 'request_log':
|
||||
self._write_request_log(log_entry)
|
||||
elif operation_type == 'translations':
|
||||
self._write_translation_log(log_entry)
|
||||
elif operation_type == 'transcriptions':
|
||||
self._write_transcription_log(log_entry)
|
||||
elif operation_type == 'tts':
|
||||
self._write_tts_log(log_entry)
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error in analytics write worker: {e}")
|
||||
|
||||
def _write_request_log(self, log_entry):
|
||||
"""Write request log to PostgreSQL"""
|
||||
try:
|
||||
with self.pg_conn.cursor() as cursor:
|
||||
cursor.execute("""
|
||||
INSERT INTO request_logs
|
||||
(endpoint, method, status_code, response_time_ms,
|
||||
ip_address, user_agent, request_size_bytes,
|
||||
response_size_bytes, session_id, created_at)
|
||||
VALUES (%(endpoint)s, %(method)s, %(status_code)s,
|
||||
%(response_time_ms)s, %(ip_address)s, %(user_agent)s,
|
||||
%(request_size_bytes)s, %(response_size_bytes)s,
|
||||
%(session_id)s, %(created_at)s)
|
||||
""", log_entry)
|
||||
except Exception as e:
|
||||
error_msg = str(e)
|
||||
if 'relation "request_logs" does not exist' in error_msg:
|
||||
logger.warning("Analytics tables not found. Run init_analytics_db.py to create them.")
|
||||
# Disable analytics to prevent repeated errors
|
||||
self.enabled = False
|
||||
else:
|
||||
logger.error(f"Error writing request log: {e}")
|
||||
|
||||
def _write_translation_log(self, log_entry):
|
||||
"""Write translation log to PostgreSQL"""
|
||||
try:
|
||||
metadata = log_entry.get('metadata', {})
|
||||
|
||||
with self.pg_conn.cursor() as cursor:
|
||||
cursor.execute("""
|
||||
INSERT INTO translation_logs
|
||||
(source_language, target_language, text_length,
|
||||
response_time_ms, success, error_message,
|
||||
session_id, created_at)
|
||||
VALUES (%(source_language)s, %(target_language)s,
|
||||
%(text_length)s, %(response_time_ms)s,
|
||||
%(success)s, %(error_message)s,
|
||||
%(session_id)s, %(created_at)s)
|
||||
""", {
|
||||
'source_language': metadata.get('source_lang'),
|
||||
'target_language': metadata.get('target_lang'),
|
||||
'text_length': metadata.get('text_length', 0),
|
||||
'response_time_ms': log_entry['response_time_ms'],
|
||||
'success': log_entry['success'],
|
||||
'error_message': log_entry['error_message'],
|
||||
'session_id': log_entry['session_id'],
|
||||
'created_at': log_entry['created_at']
|
||||
})
|
||||
except Exception as e:
|
||||
logger.error(f"Error writing translation log: {e}")
|
||||
|
||||
def _write_transcription_log(self, log_entry):
|
||||
"""Write transcription log to PostgreSQL"""
|
||||
try:
|
||||
metadata = log_entry.get('metadata', {})
|
||||
result = log_entry.get('result', {})
|
||||
|
||||
with self.pg_conn.cursor() as cursor:
|
||||
cursor.execute("""
|
||||
INSERT INTO transcription_logs
|
||||
(detected_language, audio_duration_seconds,
|
||||
file_size_bytes, response_time_ms, success,
|
||||
error_message, session_id, created_at)
|
||||
VALUES (%(detected_language)s, %(audio_duration_seconds)s,
|
||||
%(file_size_bytes)s, %(response_time_ms)s,
|
||||
%(success)s, %(error_message)s,
|
||||
%(session_id)s, %(created_at)s)
|
||||
""", {
|
||||
'detected_language': result.get('detected_language') if isinstance(result, dict) else None,
|
||||
'audio_duration_seconds': metadata.get('audio_duration', 0),
|
||||
'file_size_bytes': metadata.get('file_size', 0),
|
||||
'response_time_ms': log_entry['response_time_ms'],
|
||||
'success': log_entry['success'],
|
||||
'error_message': log_entry['error_message'],
|
||||
'session_id': log_entry['session_id'],
|
||||
'created_at': log_entry['created_at']
|
||||
})
|
||||
except Exception as e:
|
||||
logger.error(f"Error writing transcription log: {e}")
|
||||
|
||||
def _write_tts_log(self, log_entry):
|
||||
"""Write TTS log to PostgreSQL"""
|
||||
try:
|
||||
metadata = log_entry.get('metadata', {})
|
||||
|
||||
with self.pg_conn.cursor() as cursor:
|
||||
cursor.execute("""
|
||||
INSERT INTO tts_logs
|
||||
(language, text_length, voice, response_time_ms,
|
||||
success, error_message, session_id, created_at)
|
||||
VALUES (%(language)s, %(text_length)s, %(voice)s,
|
||||
%(response_time_ms)s, %(success)s,
|
||||
%(error_message)s, %(session_id)s, %(created_at)s)
|
||||
""", {
|
||||
'language': metadata.get('language'),
|
||||
'text_length': metadata.get('text_length', 0),
|
||||
'voice': metadata.get('voice'),
|
||||
'response_time_ms': log_entry['response_time_ms'],
|
||||
'success': log_entry['success'],
|
||||
'error_message': log_entry['error_message'],
|
||||
'session_id': log_entry['session_id'],
|
||||
'created_at': log_entry['created_at']
|
||||
})
|
||||
except Exception as e:
|
||||
logger.error(f"Error writing TTS log: {e}")
|
||||
|
||||
def log_error(self, error_type, error_message, **kwargs):
|
||||
"""Log error to analytics"""
|
||||
try:
|
||||
# Track in Redis
|
||||
if self.redis_client:
|
||||
pipe = self.redis_client.pipeline()
|
||||
pipe.hincrby('stats:error_types', error_type, 1)
|
||||
pipe.incr(f'stats:errors:daily:{datetime.now().strftime("%Y-%m-%d")}')
|
||||
pipe.execute()
|
||||
|
||||
# Log to PostgreSQL
|
||||
if self.pg_conn:
|
||||
with self.pg_conn.cursor() as cursor:
|
||||
cursor.execute("""
|
||||
INSERT INTO error_logs
|
||||
(error_type, error_message, endpoint, method,
|
||||
status_code, ip_address, user_agent, request_id,
|
||||
stack_trace, created_at)
|
||||
VALUES (%(error_type)s, %(error_message)s,
|
||||
%(endpoint)s, %(method)s, %(status_code)s,
|
||||
%(ip_address)s, %(user_agent)s,
|
||||
%(request_id)s, %(stack_trace)s,
|
||||
%(created_at)s)
|
||||
""", {
|
||||
'error_type': error_type,
|
||||
'error_message': error_message[:1000],
|
||||
'endpoint': kwargs.get('endpoint'),
|
||||
'method': kwargs.get('method'),
|
||||
'status_code': kwargs.get('status_code'),
|
||||
'ip_address': kwargs.get('ip_address'),
|
||||
'user_agent': kwargs.get('user_agent', '')[:500],
|
||||
'request_id': kwargs.get('request_id'),
|
||||
'stack_trace': kwargs.get('stack_trace', '')[:5000],
|
||||
'created_at': datetime.now()
|
||||
})
|
||||
except Exception as e:
|
||||
logger.error(f"Error logging analytics error: {e}")
|
||||
|
||||
def update_cache_stats(self, hit=True):
|
||||
"""Update cache hit/miss statistics"""
|
||||
try:
|
||||
if self.redis_client:
|
||||
if hit:
|
||||
self.redis_client.incr('stats:cache:hits')
|
||||
else:
|
||||
self.redis_client.incr('stats:cache:misses')
|
||||
except Exception as e:
|
||||
logger.error(f"Error updating cache stats: {e}")
|
||||
|
||||
# Create global instance
|
||||
analytics_tracker = AnalyticsTracker()
|
||||
|
||||
# Convenience decorators
|
||||
def track_translation(**kwargs):
|
||||
"""Decorator to track translation operations"""
|
||||
return analytics_tracker.track_operation('translations', **kwargs)
|
||||
|
||||
def track_transcription(**kwargs):
|
||||
"""Decorator to track transcription operations"""
|
||||
return analytics_tracker.track_operation('transcriptions', **kwargs)
|
||||
|
||||
def track_tts(**kwargs):
|
||||
"""Decorator to track TTS operations"""
|
||||
return analytics_tracker.track_operation('tts', **kwargs)
|
746
app_with_db.py
Normal file
746
app_with_db.py
Normal file
@ -0,0 +1,746 @@
|
||||
# This is the updated app.py with Redis and PostgreSQL integration
|
||||
# To use this, rename it to app.py after backing up the original
|
||||
|
||||
import os
|
||||
import time
|
||||
import tempfile
|
||||
import requests
|
||||
import json
|
||||
import logging
|
||||
from dotenv import load_dotenv
|
||||
from flask import Flask, render_template, request, jsonify, Response, send_file, send_from_directory, stream_with_context, g
|
||||
from flask_cors import CORS, cross_origin
|
||||
import whisper
|
||||
import torch
|
||||
import ollama
|
||||
from whisper_config import MODEL_SIZE, GPU_OPTIMIZATIONS, TRANSCRIBE_OPTIONS
|
||||
from pywebpush import webpush, WebPushException
|
||||
import base64
|
||||
from cryptography.hazmat.primitives.asymmetric import ec
|
||||
from cryptography.hazmat.primitives import serialization
|
||||
from cryptography.hazmat.backends import default_backend
|
||||
import gc
|
||||
from functools import wraps
|
||||
import traceback
|
||||
from validators import Validators
|
||||
import atexit
|
||||
import threading
|
||||
from datetime import datetime, timedelta
|
||||
|
||||
# Import new database and Redis components
|
||||
from database import db, init_db, Translation, Transcription, UserPreferences, UsageAnalytics
|
||||
from redis_manager import RedisManager, redis_cache
|
||||
from redis_rate_limiter import RedisRateLimiter, rate_limit
|
||||
from redis_session_manager import RedisSessionManager, init_app as init_redis_sessions
|
||||
|
||||
# Load environment variables
|
||||
load_dotenv()
|
||||
|
||||
# Initialize logging
|
||||
logging.basicConfig(level=logging.INFO)
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
# Import other components
|
||||
from werkzeug.middleware.proxy_fix import ProxyFix
|
||||
from config import init_app as init_config
|
||||
from secrets_manager import init_app as init_secrets
|
||||
from request_size_limiter import RequestSizeLimiter, limit_request_size
|
||||
from error_logger import ErrorLogger, log_errors, log_performance, log_exception, get_logger
|
||||
from memory_manager import MemoryManager, AudioProcessingContext, with_memory_management
|
||||
|
||||
# Error boundary decorator
|
||||
def with_error_boundary(func):
|
||||
@wraps(func)
|
||||
def wrapper(*args, **kwargs):
|
||||
try:
|
||||
return func(*args, **kwargs)
|
||||
except Exception as e:
|
||||
log_exception(
|
||||
e,
|
||||
message=f"Error in {func.__name__}",
|
||||
endpoint=request.endpoint,
|
||||
method=request.method,
|
||||
path=request.path,
|
||||
ip=request.remote_addr,
|
||||
function=func.__name__,
|
||||
module=func.__module__
|
||||
)
|
||||
|
||||
if any(keyword in str(e).lower() for keyword in ['inject', 'attack', 'malicious', 'unauthorized']):
|
||||
app.error_logger.log_security(
|
||||
'suspicious_error',
|
||||
severity='warning',
|
||||
error_type=type(e).__name__,
|
||||
error_message=str(e),
|
||||
endpoint=request.endpoint,
|
||||
ip=request.remote_addr
|
||||
)
|
||||
|
||||
error_message = str(e) if app.debug else "An internal error occurred"
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': error_message,
|
||||
'component': func.__name__,
|
||||
'request_id': getattr(g, 'request_id', None)
|
||||
}), 500
|
||||
return wrapper
|
||||
|
||||
app = Flask(__name__)
|
||||
|
||||
# Apply ProxyFix middleware
|
||||
app.wsgi_app = ProxyFix(
|
||||
app.wsgi_app,
|
||||
x_for=1,
|
||||
x_proto=1,
|
||||
x_host=1,
|
||||
x_prefix=1
|
||||
)
|
||||
|
||||
# Initialize configuration and secrets
|
||||
init_config(app)
|
||||
init_secrets(app)
|
||||
|
||||
# Initialize database
|
||||
init_db(app)
|
||||
|
||||
# Initialize Redis
|
||||
redis_manager = RedisManager(app)
|
||||
app.redis = redis_manager
|
||||
|
||||
# Initialize Redis-based rate limiter
|
||||
redis_rate_limiter = RedisRateLimiter(redis_manager)
|
||||
app.redis_rate_limiter = redis_rate_limiter
|
||||
|
||||
# Initialize Redis-based session management
|
||||
init_redis_sessions(app)
|
||||
|
||||
# Configure CORS
|
||||
cors_config = {
|
||||
"origins": app.config.get('CORS_ORIGINS', ['*']),
|
||||
"methods": ["GET", "POST", "OPTIONS"],
|
||||
"allow_headers": ["Content-Type", "Authorization", "X-Requested-With", "X-Admin-Token"],
|
||||
"expose_headers": ["Content-Range", "X-Content-Range"],
|
||||
"supports_credentials": True,
|
||||
"max_age": 3600
|
||||
}
|
||||
|
||||
CORS(app, resources={
|
||||
r"/api/*": cors_config,
|
||||
r"/transcribe": cors_config,
|
||||
r"/translate": cors_config,
|
||||
r"/translate/stream": cors_config,
|
||||
r"/speak": cors_config,
|
||||
r"/get_audio/*": cors_config,
|
||||
r"/check_tts_server": cors_config,
|
||||
r"/update_tts_config": cors_config,
|
||||
r"/health/*": cors_config,
|
||||
r"/admin/*": {
|
||||
**cors_config,
|
||||
"origins": app.config.get('ADMIN_CORS_ORIGINS', ['http://localhost:*'])
|
||||
}
|
||||
})
|
||||
|
||||
# Configure upload folder
|
||||
upload_folder = app.config.get('UPLOAD_FOLDER')
|
||||
if not upload_folder:
|
||||
upload_folder = os.path.join(tempfile.gettempdir(), 'talk2me_uploads')
|
||||
|
||||
try:
|
||||
os.makedirs(upload_folder, mode=0o755, exist_ok=True)
|
||||
logger.info(f"Using upload folder: {upload_folder}")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to create upload folder {upload_folder}: {str(e)}")
|
||||
upload_folder = tempfile.mkdtemp(prefix='talk2me_')
|
||||
logger.warning(f"Falling back to temporary folder: {upload_folder}")
|
||||
|
||||
app.config['UPLOAD_FOLDER'] = upload_folder
|
||||
|
||||
# Initialize request size limiter
|
||||
request_size_limiter = RequestSizeLimiter(app, {
|
||||
'max_content_length': app.config.get('MAX_CONTENT_LENGTH', 50 * 1024 * 1024),
|
||||
'max_audio_size': app.config.get('MAX_AUDIO_SIZE', 25 * 1024 * 1024),
|
||||
'max_json_size': app.config.get('MAX_JSON_SIZE', 1 * 1024 * 1024),
|
||||
'max_image_size': app.config.get('MAX_IMAGE_SIZE', 10 * 1024 * 1024),
|
||||
})
|
||||
|
||||
# Initialize error logging
|
||||
error_logger = ErrorLogger(app, {
|
||||
'log_level': app.config.get('LOG_LEVEL', 'INFO'),
|
||||
'log_file': app.config.get('LOG_FILE', 'logs/talk2me.log'),
|
||||
'error_log_file': app.config.get('ERROR_LOG_FILE', 'logs/errors.log'),
|
||||
'max_bytes': app.config.get('LOG_MAX_BYTES', 50 * 1024 * 1024),
|
||||
'backup_count': app.config.get('LOG_BACKUP_COUNT', 10)
|
||||
})
|
||||
|
||||
logger = get_logger(__name__)
|
||||
|
||||
# Initialize memory management
|
||||
memory_manager = MemoryManager(app, {
|
||||
'memory_threshold_mb': app.config.get('MEMORY_THRESHOLD_MB', 4096),
|
||||
'gpu_memory_threshold_mb': app.config.get('GPU_MEMORY_THRESHOLD_MB', 2048),
|
||||
'cleanup_interval': app.config.get('MEMORY_CLEANUP_INTERVAL', 30)
|
||||
})
|
||||
|
||||
# Initialize Whisper model
|
||||
logger.info("Initializing Whisper model with GPU optimization...")
|
||||
|
||||
if torch.cuda.is_available():
|
||||
device = torch.device("cuda")
|
||||
try:
|
||||
gpu_name = torch.cuda.get_device_name(0)
|
||||
if 'AMD' in gpu_name or 'Radeon' in gpu_name:
|
||||
logger.info(f"AMD GPU detected via ROCm: {gpu_name}")
|
||||
else:
|
||||
logger.info(f"NVIDIA GPU detected: {gpu_name}")
|
||||
except:
|
||||
logger.info("GPU detected - using CUDA/ROCm acceleration")
|
||||
elif hasattr(torch.backends, 'mps') and torch.backends.mps.is_available():
|
||||
device = torch.device("mps")
|
||||
logger.info("Apple Silicon detected - using Metal Performance Shaders")
|
||||
else:
|
||||
device = torch.device("cpu")
|
||||
logger.info("No GPU acceleration available - using CPU")
|
||||
|
||||
logger.info(f"Using device: {device}")
|
||||
|
||||
whisper_model = whisper.load_model(MODEL_SIZE, device=device)
|
||||
|
||||
# Enable GPU optimizations
|
||||
if device.type == 'cuda':
|
||||
try:
|
||||
torch.backends.cuda.matmul.allow_tf32 = True
|
||||
torch.backends.cudnn.allow_tf32 = True
|
||||
torch.backends.cudnn.benchmark = True
|
||||
whisper_model.eval()
|
||||
whisper_model = whisper_model.half()
|
||||
torch.cuda.empty_cache()
|
||||
|
||||
logger.info("Warming up GPU with dummy inference...")
|
||||
with torch.no_grad():
|
||||
dummy_audio = torch.randn(1, 16000 * 30).to(device).half()
|
||||
_ = whisper_model.encode(whisper.pad_or_trim(dummy_audio))
|
||||
|
||||
logger.info(f"GPU memory allocated: {torch.cuda.memory_allocated() / 1024**2:.2f} MB")
|
||||
logger.info("Whisper model loaded and optimized for GPU")
|
||||
except Exception as e:
|
||||
logger.warning(f"Some GPU optimizations failed: {e}")
|
||||
elif device.type == 'mps':
|
||||
whisper_model.eval()
|
||||
logger.info("Whisper model loaded and optimized for Apple Silicon")
|
||||
else:
|
||||
whisper_model.eval()
|
||||
logger.info("Whisper model loaded (CPU mode)")
|
||||
|
||||
memory_manager.set_whisper_model(whisper_model)
|
||||
app.whisper_model = whisper_model
|
||||
|
||||
# Supported languages
|
||||
SUPPORTED_LANGUAGES = {
|
||||
"ar": "Arabic",
|
||||
"hy": "Armenian",
|
||||
"az": "Azerbaijani",
|
||||
"en": "English",
|
||||
"fr": "French",
|
||||
"ka": "Georgian",
|
||||
"kk": "Kazakh",
|
||||
"zh": "Mandarin",
|
||||
"fa": "Farsi",
|
||||
"pt": "Portuguese",
|
||||
"ru": "Russian",
|
||||
"es": "Spanish",
|
||||
"tr": "Turkish",
|
||||
"uz": "Uzbek"
|
||||
}
|
||||
|
||||
LANGUAGE_TO_CODE = {v: k for k, v in SUPPORTED_LANGUAGES.items()}
|
||||
|
||||
LANGUAGE_TO_VOICE = {
|
||||
"Arabic": "ar-EG-ShakirNeural",
|
||||
"Armenian": "echo",
|
||||
"Azerbaijani": "az-AZ-BanuNeural",
|
||||
"English": "en-GB-RyanNeural",
|
||||
"French": "fr-FR-DeniseNeural",
|
||||
"Georgian": "ka-GE-GiorgiNeural",
|
||||
"Kazakh": "kk-KZ-DauletNeural",
|
||||
"Mandarin": "zh-CN-YunjianNeural",
|
||||
"Farsi": "fa-IR-FaridNeural",
|
||||
"Portuguese": "pt-BR-ThalitaNeural",
|
||||
"Russian": "ru-RU-SvetlanaNeural",
|
||||
"Spanish": "es-CR-MariaNeural",
|
||||
"Turkish": "tr-TR-EmelNeural",
|
||||
"Uzbek": "uz-UZ-SardorNeural"
|
||||
}
|
||||
|
||||
# Generate VAPID keys for push notifications
|
||||
if not os.path.exists('vapid_private.pem'):
|
||||
private_key = ec.generate_private_key(ec.SECP256R1(), default_backend())
|
||||
public_key = private_key.public_key()
|
||||
|
||||
with open('vapid_private.pem', 'wb') as f:
|
||||
f.write(private_key.private_bytes(
|
||||
encoding=serialization.Encoding.PEM,
|
||||
format=serialization.PrivateFormat.PKCS8,
|
||||
encryption_algorithm=serialization.NoEncryption()
|
||||
))
|
||||
|
||||
with open('vapid_public.pem', 'wb') as f:
|
||||
f.write(public_key.public_bytes(
|
||||
encoding=serialization.Encoding.PEM,
|
||||
format=serialization.PublicFormat.SubjectPublicKeyInfo
|
||||
))
|
||||
|
||||
with open('vapid_private.pem', 'rb') as f:
|
||||
vapid_private_key = f.read()
|
||||
with open('vapid_public.pem', 'rb') as f:
|
||||
vapid_public_pem = f.read()
|
||||
vapid_public_key = serialization.load_pem_public_key(
|
||||
vapid_public_pem,
|
||||
backend=default_backend()
|
||||
)
|
||||
|
||||
public_numbers = vapid_public_key.public_numbers()
|
||||
x = public_numbers.x.to_bytes(32, byteorder='big')
|
||||
y = public_numbers.y.to_bytes(32, byteorder='big')
|
||||
vapid_public_key_base64 = base64.urlsafe_b64encode(b'\x04' + x + y).decode('utf-8').rstrip('=')
|
||||
|
||||
# Store push subscriptions in Redis instead of memory
|
||||
# push_subscriptions = [] # Removed - now using Redis
|
||||
|
||||
# Temporary file cleanup
|
||||
TEMP_FILE_MAX_AGE = 300
|
||||
CLEANUP_INTERVAL = 60
|
||||
|
||||
def cleanup_temp_files():
|
||||
"""Clean up old temporary files"""
|
||||
try:
|
||||
current_time = datetime.now()
|
||||
|
||||
# Clean files from upload folder
|
||||
if os.path.exists(app.config['UPLOAD_FOLDER']):
|
||||
for filename in os.listdir(app.config['UPLOAD_FOLDER']):
|
||||
filepath = os.path.join(app.config['UPLOAD_FOLDER'], filename)
|
||||
if os.path.isfile(filepath):
|
||||
file_age = current_time - datetime.fromtimestamp(os.path.getmtime(filepath))
|
||||
if file_age > timedelta(seconds=TEMP_FILE_MAX_AGE):
|
||||
try:
|
||||
os.remove(filepath)
|
||||
logger.info(f"Cleaned up file: {filepath}")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to remove file {filepath}: {str(e)}")
|
||||
|
||||
logger.debug("Cleanup completed")
|
||||
except Exception as e:
|
||||
logger.error(f"Error during temp file cleanup: {str(e)}")
|
||||
|
||||
def run_cleanup_loop():
|
||||
"""Run cleanup in a separate thread"""
|
||||
while True:
|
||||
time.sleep(CLEANUP_INTERVAL)
|
||||
cleanup_temp_files()
|
||||
|
||||
# Start cleanup thread
|
||||
cleanup_thread = threading.Thread(target=run_cleanup_loop, daemon=True)
|
||||
cleanup_thread.start()
|
||||
|
||||
# Analytics collection helper
|
||||
def collect_analytics(service: str, duration_ms: int = None, metadata: dict = None):
|
||||
"""Collect usage analytics to database"""
|
||||
try:
|
||||
from sqlalchemy import func
|
||||
|
||||
today = datetime.utcnow().date()
|
||||
hour = datetime.utcnow().hour
|
||||
|
||||
# Get or create analytics record
|
||||
analytics = UsageAnalytics.query.filter_by(date=today, hour=hour).first()
|
||||
if not analytics:
|
||||
analytics = UsageAnalytics(date=today, hour=hour)
|
||||
db.session.add(analytics)
|
||||
|
||||
# Update counters
|
||||
analytics.total_requests += 1
|
||||
|
||||
if service == 'transcription':
|
||||
analytics.transcriptions += 1
|
||||
if duration_ms:
|
||||
if analytics.avg_transcription_time_ms:
|
||||
analytics.avg_transcription_time_ms = (
|
||||
(analytics.avg_transcription_time_ms * (analytics.transcriptions - 1) + duration_ms)
|
||||
/ analytics.transcriptions
|
||||
)
|
||||
else:
|
||||
analytics.avg_transcription_time_ms = duration_ms
|
||||
|
||||
elif service == 'translation':
|
||||
analytics.translations += 1
|
||||
if duration_ms:
|
||||
if analytics.avg_translation_time_ms:
|
||||
analytics.avg_translation_time_ms = (
|
||||
(analytics.avg_translation_time_ms * (analytics.translations - 1) + duration_ms)
|
||||
/ analytics.translations
|
||||
)
|
||||
else:
|
||||
analytics.avg_translation_time_ms = duration_ms
|
||||
|
||||
elif service == 'tts':
|
||||
analytics.tts_requests += 1
|
||||
if duration_ms:
|
||||
if analytics.avg_tts_time_ms:
|
||||
analytics.avg_tts_time_ms = (
|
||||
(analytics.avg_tts_time_ms * (analytics.tts_requests - 1) + duration_ms)
|
||||
/ analytics.tts_requests
|
||||
)
|
||||
else:
|
||||
analytics.avg_tts_time_ms = duration_ms
|
||||
|
||||
db.session.commit()
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to collect analytics: {e}")
|
||||
db.session.rollback()
|
||||
|
||||
# Routes
|
||||
@app.route('/')
|
||||
def index():
|
||||
return render_template('index.html', languages=sorted(SUPPORTED_LANGUAGES.values()))
|
||||
|
||||
@app.route('/transcribe', methods=['POST'])
|
||||
@rate_limit(requests_per_minute=10, requests_per_hour=100, check_size=True)
|
||||
@limit_request_size(max_audio_size=25 * 1024 * 1024)
|
||||
@with_error_boundary
|
||||
@log_performance('transcribe_audio')
|
||||
@with_memory_management
|
||||
def transcribe():
|
||||
start_time = time.time()
|
||||
|
||||
with AudioProcessingContext(app.memory_manager, name='transcribe') as ctx:
|
||||
if 'audio' not in request.files:
|
||||
return jsonify({'error': 'No audio file provided'}), 400
|
||||
|
||||
audio_file = request.files['audio']
|
||||
|
||||
valid, error_msg = Validators.validate_audio_file(audio_file)
|
||||
if not valid:
|
||||
return jsonify({'error': error_msg}), 400
|
||||
|
||||
source_lang = request.form.get('source_lang', '')
|
||||
allowed_languages = set(SUPPORTED_LANGUAGES.values())
|
||||
source_lang = Validators.validate_language_code(source_lang, allowed_languages) or ''
|
||||
|
||||
temp_filename = f'input_audio_{int(time.time() * 1000)}.wav'
|
||||
temp_path = os.path.join(app.config['UPLOAD_FOLDER'], temp_filename)
|
||||
|
||||
with open(temp_path, 'wb') as f:
|
||||
audio_file.save(f)
|
||||
|
||||
ctx.add_temp_file(temp_path)
|
||||
|
||||
# Add to Redis session
|
||||
if hasattr(g, 'session_manager') and hasattr(g, 'user_session'):
|
||||
file_size = os.path.getsize(temp_path)
|
||||
g.session_manager.add_resource(
|
||||
session_id=g.user_session.session_id,
|
||||
resource_type='audio_file',
|
||||
resource_id=temp_filename,
|
||||
path=temp_path,
|
||||
size_bytes=file_size,
|
||||
metadata={'filename': temp_filename, 'purpose': 'transcription'}
|
||||
)
|
||||
|
||||
try:
|
||||
auto_detect = source_lang == 'auto' or source_lang == ''
|
||||
|
||||
transcribe_options = {
|
||||
"task": "transcribe",
|
||||
"temperature": 0,
|
||||
"best_of": 1,
|
||||
"beam_size": 1,
|
||||
"fp16": device.type == 'cuda',
|
||||
"condition_on_previous_text": False,
|
||||
"compression_ratio_threshold": 2.4,
|
||||
"logprob_threshold": -1.0,
|
||||
"no_speech_threshold": 0.6
|
||||
}
|
||||
|
||||
if not auto_detect:
|
||||
transcribe_options["language"] = LANGUAGE_TO_CODE.get(source_lang, None)
|
||||
|
||||
if device.type == 'cuda':
|
||||
torch.cuda.empty_cache()
|
||||
|
||||
with torch.no_grad():
|
||||
result = whisper_model.transcribe(
|
||||
temp_path,
|
||||
**transcribe_options
|
||||
)
|
||||
|
||||
transcribed_text = result["text"]
|
||||
|
||||
detected_language = None
|
||||
if auto_detect and 'language' in result:
|
||||
detected_code = result['language']
|
||||
for lang_name, lang_code in LANGUAGE_TO_CODE.items():
|
||||
if lang_code == detected_code:
|
||||
detected_language = lang_name
|
||||
break
|
||||
|
||||
logger.info(f"Auto-detected language: {detected_language} ({detected_code})")
|
||||
|
||||
# Calculate duration
|
||||
duration_ms = int((time.time() - start_time) * 1000)
|
||||
|
||||
# Save to database
|
||||
try:
|
||||
transcription = Transcription(
|
||||
session_id=g.user_session.session_id if hasattr(g, 'user_session') else None,
|
||||
user_id=g.user_session.user_id if hasattr(g, 'user_session') else None,
|
||||
transcribed_text=transcribed_text,
|
||||
detected_language=detected_language or source_lang,
|
||||
transcription_time_ms=duration_ms,
|
||||
model_used=MODEL_SIZE,
|
||||
audio_file_size=os.path.getsize(temp_path),
|
||||
ip_address=request.remote_addr,
|
||||
user_agent=request.headers.get('User-Agent')
|
||||
)
|
||||
db.session.add(transcription)
|
||||
db.session.commit()
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to save transcription to database: {e}")
|
||||
db.session.rollback()
|
||||
|
||||
# Collect analytics
|
||||
collect_analytics('transcription', duration_ms)
|
||||
|
||||
# Send notification if push is enabled
|
||||
push_count = redis_manager.scard('push_subscriptions')
|
||||
if push_count > 0:
|
||||
send_push_notification(
|
||||
title="Transcription Complete",
|
||||
body=f"Successfully transcribed: {transcribed_text[:50]}...",
|
||||
tag="transcription-complete"
|
||||
)
|
||||
|
||||
response = {
|
||||
'success': True,
|
||||
'text': transcribed_text
|
||||
}
|
||||
|
||||
if detected_language:
|
||||
response['detected_language'] = detected_language
|
||||
|
||||
return jsonify(response)
|
||||
except Exception as e:
|
||||
logger.error(f"Transcription error: {str(e)}")
|
||||
return jsonify({'error': f'Transcription failed: {str(e)}'}), 500
|
||||
finally:
|
||||
try:
|
||||
if 'temp_path' in locals() and os.path.exists(temp_path):
|
||||
os.remove(temp_path)
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to clean up temp file: {e}")
|
||||
|
||||
if device.type == 'cuda':
|
||||
torch.cuda.empty_cache()
|
||||
torch.cuda.synchronize()
|
||||
gc.collect()
|
||||
|
||||
@app.route('/translate', methods=['POST'])
|
||||
@rate_limit(requests_per_minute=20, requests_per_hour=300, check_size=True)
|
||||
@limit_request_size(max_size=1 * 1024 * 1024)
|
||||
@with_error_boundary
|
||||
@log_performance('translate_text')
|
||||
def translate():
|
||||
start_time = time.time()
|
||||
|
||||
try:
|
||||
if not Validators.validate_json_size(request.json, max_size_kb=100):
|
||||
return jsonify({'error': 'Request too large'}), 413
|
||||
|
||||
data = request.json
|
||||
|
||||
text = data.get('text', '')
|
||||
text = Validators.sanitize_text(text)
|
||||
if not text:
|
||||
return jsonify({'error': 'No text provided'}), 400
|
||||
|
||||
allowed_languages = set(SUPPORTED_LANGUAGES.values())
|
||||
source_lang = Validators.validate_language_code(
|
||||
data.get('source_lang', ''), allowed_languages
|
||||
) or 'auto'
|
||||
target_lang = Validators.validate_language_code(
|
||||
data.get('target_lang', ''), allowed_languages
|
||||
)
|
||||
|
||||
if not target_lang:
|
||||
return jsonify({'error': 'Invalid target language'}), 400
|
||||
|
||||
# Check cache first
|
||||
cached_translation = redis_manager.get_cached_translation(
|
||||
text, source_lang, target_lang
|
||||
)
|
||||
if cached_translation:
|
||||
logger.info("Translation served from cache")
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'translation': cached_translation,
|
||||
'cached': True
|
||||
})
|
||||
|
||||
# Create prompt for translation
|
||||
prompt = f"""
|
||||
Translate the following text from {source_lang} to {target_lang}:
|
||||
|
||||
"{text}"
|
||||
|
||||
Provide only the translation without any additional text.
|
||||
"""
|
||||
|
||||
response = ollama.chat(
|
||||
model="gemma3:27b",
|
||||
messages=[
|
||||
{
|
||||
"role": "user",
|
||||
"content": prompt
|
||||
}
|
||||
]
|
||||
)
|
||||
|
||||
translated_text = response['message']['content'].strip()
|
||||
|
||||
# Calculate duration
|
||||
duration_ms = int((time.time() - start_time) * 1000)
|
||||
|
||||
# Cache the translation
|
||||
redis_manager.cache_translation(
|
||||
text, source_lang, target_lang, translated_text
|
||||
)
|
||||
|
||||
# Save to database
|
||||
try:
|
||||
translation = Translation(
|
||||
session_id=g.user_session.session_id if hasattr(g, 'user_session') else None,
|
||||
user_id=g.user_session.user_id if hasattr(g, 'user_session') else None,
|
||||
source_text=text,
|
||||
source_language=source_lang,
|
||||
target_text=translated_text,
|
||||
target_language=target_lang,
|
||||
translation_time_ms=duration_ms,
|
||||
model_used="gemma3:27b",
|
||||
ip_address=request.remote_addr,
|
||||
user_agent=request.headers.get('User-Agent')
|
||||
)
|
||||
db.session.add(translation)
|
||||
db.session.commit()
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to save translation to database: {e}")
|
||||
db.session.rollback()
|
||||
|
||||
# Collect analytics
|
||||
collect_analytics('translation', duration_ms)
|
||||
|
||||
# Send notification
|
||||
push_count = redis_manager.scard('push_subscriptions')
|
||||
if push_count > 0:
|
||||
send_push_notification(
|
||||
title="Translation Complete",
|
||||
body=f"Translated from {source_lang} to {target_lang}",
|
||||
tag="translation-complete",
|
||||
data={'translation': translated_text[:100]}
|
||||
)
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'translation': translated_text
|
||||
})
|
||||
except Exception as e:
|
||||
logger.error(f"Translation error: {str(e)}")
|
||||
return jsonify({'error': f'Translation failed: {str(e)}'}), 500
|
||||
|
||||
@app.route('/api/push-subscribe', methods=['POST'])
|
||||
@rate_limit(requests_per_minute=10, requests_per_hour=50)
|
||||
def push_subscribe():
|
||||
try:
|
||||
subscription = request.json
|
||||
# Store subscription in Redis
|
||||
subscription_id = f"sub_{int(time.time() * 1000)}"
|
||||
redis_manager.set(f"push_subscription:{subscription_id}", subscription, expire=86400 * 30) # 30 days
|
||||
redis_manager.sadd('push_subscriptions', subscription_id)
|
||||
|
||||
logger.info(f"New push subscription registered: {subscription_id}")
|
||||
return jsonify({'success': True})
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to register push subscription: {str(e)}")
|
||||
return jsonify({'success': False, 'error': str(e)}), 500
|
||||
|
||||
def send_push_notification(title, body, icon='/static/icons/icon-192x192.png',
|
||||
badge='/static/icons/icon-192x192.png', tag=None, data=None):
|
||||
"""Send push notification to all subscribed clients"""
|
||||
claims = {
|
||||
"sub": "mailto:admin@talk2me.app",
|
||||
"exp": int(time.time()) + 86400
|
||||
}
|
||||
|
||||
notification_sent = 0
|
||||
|
||||
# Get all subscription IDs from Redis
|
||||
subscription_ids = redis_manager.smembers('push_subscriptions')
|
||||
|
||||
for sub_id in subscription_ids:
|
||||
subscription = redis_manager.get(f"push_subscription:{sub_id}")
|
||||
if not subscription:
|
||||
continue
|
||||
|
||||
try:
|
||||
webpush(
|
||||
subscription_info=subscription,
|
||||
data=json.dumps({
|
||||
'title': title,
|
||||
'body': body,
|
||||
'icon': icon,
|
||||
'badge': badge,
|
||||
'tag': tag or 'talk2me-notification',
|
||||
'data': data or {}
|
||||
}),
|
||||
vapid_private_key=vapid_private_key,
|
||||
vapid_claims=claims
|
||||
)
|
||||
notification_sent += 1
|
||||
except WebPushException as e:
|
||||
logger.error(f"Failed to send push notification: {str(e)}")
|
||||
if e.response and e.response.status_code == 410:
|
||||
# Remove invalid subscription
|
||||
redis_manager.delete(f"push_subscription:{sub_id}")
|
||||
redis_manager.srem('push_subscriptions', sub_id)
|
||||
|
||||
logger.info(f"Sent {notification_sent} push notifications")
|
||||
return notification_sent
|
||||
|
||||
# Initialize app
|
||||
app.start_time = time.time()
|
||||
app.request_count = 0
|
||||
|
||||
@app.before_request
|
||||
def before_request():
|
||||
app.request_count = getattr(app, 'request_count', 0) + 1
|
||||
|
||||
# Error handlers
|
||||
@app.errorhandler(404)
|
||||
def not_found_error(error):
|
||||
logger.warning(f"404 error: {request.url}")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Resource not found',
|
||||
'status': 404
|
||||
}), 404
|
||||
|
||||
@app.errorhandler(500)
|
||||
def internal_error(error):
|
||||
logger.error(f"500 error: {str(error)}")
|
||||
logger.error(traceback.format_exc())
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Internal server error',
|
||||
'status': 500
|
||||
}), 500
|
||||
|
||||
if __name__ == '__main__':
|
||||
app.run(host='0.0.0.0', port=5005, debug=True)
|
476
auth.py
Normal file
476
auth.py
Normal file
@ -0,0 +1,476 @@
|
||||
"""Authentication and authorization utilities for Talk2Me"""
|
||||
import os
|
||||
import uuid
|
||||
import functools
|
||||
from datetime import datetime, timedelta, timezone
|
||||
from typing import Optional, Dict, Any, Callable, Union, List
|
||||
from flask import request, jsonify, g, current_app
|
||||
from flask_jwt_extended import (
|
||||
JWTManager, create_access_token, create_refresh_token,
|
||||
get_jwt_identity, jwt_required, get_jwt, verify_jwt_in_request
|
||||
)
|
||||
from werkzeug.exceptions import Unauthorized
|
||||
from sqlalchemy.exc import IntegrityError
|
||||
|
||||
from database import db
|
||||
from auth_models import User, LoginHistory, UserSession, RevokedToken, bcrypt
|
||||
from error_logger import log_exception
|
||||
|
||||
# Initialize JWT Manager
|
||||
jwt = JWTManager()
|
||||
|
||||
|
||||
def init_auth(app):
|
||||
"""Initialize authentication system with app"""
|
||||
# Configure JWT
|
||||
app.config['JWT_SECRET_KEY'] = app.config.get('JWT_SECRET_KEY', os.environ.get('JWT_SECRET_KEY', 'your-secret-key-change-in-production'))
|
||||
app.config['JWT_ACCESS_TOKEN_EXPIRES'] = timedelta(hours=1)
|
||||
app.config['JWT_REFRESH_TOKEN_EXPIRES'] = timedelta(days=30)
|
||||
app.config['JWT_ALGORITHM'] = 'HS256'
|
||||
app.config['JWT_BLACKLIST_ENABLED'] = True
|
||||
app.config['JWT_BLACKLIST_TOKEN_CHECKS'] = ['access', 'refresh']
|
||||
|
||||
# Initialize JWT manager
|
||||
jwt.init_app(app)
|
||||
|
||||
# Initialize bcrypt
|
||||
bcrypt.init_app(app)
|
||||
|
||||
# Register JWT callbacks
|
||||
@jwt.token_in_blocklist_loader
|
||||
def check_if_token_revoked(jwt_header, jwt_payload):
|
||||
jti = jwt_payload["jti"]
|
||||
return RevokedToken.is_token_revoked(jti)
|
||||
|
||||
@jwt.expired_token_loader
|
||||
def expired_token_callback(jwt_header, jwt_payload):
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Token has expired',
|
||||
'code': 'token_expired'
|
||||
}), 401
|
||||
|
||||
@jwt.invalid_token_loader
|
||||
def invalid_token_callback(error):
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Invalid token',
|
||||
'code': 'invalid_token'
|
||||
}), 401
|
||||
|
||||
@jwt.unauthorized_loader
|
||||
def missing_token_callback(error):
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Authorization required',
|
||||
'code': 'authorization_required'
|
||||
}), 401
|
||||
|
||||
@jwt.revoked_token_loader
|
||||
def revoked_token_callback(jwt_header, jwt_payload):
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Token has been revoked',
|
||||
'code': 'token_revoked'
|
||||
}), 401
|
||||
|
||||
|
||||
def create_user(email: str, username: str, password: str, full_name: Optional[str] = None,
|
||||
role: str = 'user', is_verified: bool = False) -> tuple[Optional[User], Optional[str]]:
|
||||
"""Create a new user account"""
|
||||
try:
|
||||
# Check if user already exists
|
||||
if User.query.filter((User.email == email) | (User.username == username)).first():
|
||||
return None, "User with this email or username already exists"
|
||||
|
||||
# Create user
|
||||
user = User(
|
||||
email=email,
|
||||
username=username,
|
||||
full_name=full_name,
|
||||
role=role,
|
||||
is_verified=is_verified
|
||||
)
|
||||
user.set_password(password)
|
||||
|
||||
db.session.add(user)
|
||||
db.session.commit()
|
||||
|
||||
return user, None
|
||||
except IntegrityError:
|
||||
db.session.rollback()
|
||||
return None, "User with this email or username already exists"
|
||||
except Exception as e:
|
||||
db.session.rollback()
|
||||
log_exception(e, "Failed to create user")
|
||||
return None, "Failed to create user account"
|
||||
|
||||
|
||||
def authenticate_user(username_or_email: str, password: str) -> tuple[Optional[User], Optional[str]]:
|
||||
"""Authenticate user with username/email and password"""
|
||||
# Find user by username or email
|
||||
user = User.query.filter(
|
||||
(User.username == username_or_email) | (User.email == username_or_email)
|
||||
).first()
|
||||
|
||||
if not user:
|
||||
return None, "Invalid credentials"
|
||||
|
||||
# Check if user can login
|
||||
can_login, reason = user.can_login()
|
||||
if not can_login:
|
||||
user.record_login_attempt(False)
|
||||
db.session.commit()
|
||||
return None, reason
|
||||
|
||||
# Verify password
|
||||
if not user.check_password(password):
|
||||
user.record_login_attempt(False)
|
||||
db.session.commit()
|
||||
return None, "Invalid credentials"
|
||||
|
||||
# Success
|
||||
user.record_login_attempt(True)
|
||||
db.session.commit()
|
||||
|
||||
return user, None
|
||||
|
||||
|
||||
def authenticate_api_key(api_key: str) -> tuple[Optional[User], Optional[str]]:
|
||||
"""Authenticate user with API key"""
|
||||
user = User.query.filter_by(api_key=api_key).first()
|
||||
|
||||
if not user:
|
||||
return None, "Invalid API key"
|
||||
|
||||
# Check if user can login
|
||||
can_login, reason = user.can_login()
|
||||
if not can_login:
|
||||
return None, reason
|
||||
|
||||
# Update last active
|
||||
user.last_active_at = datetime.utcnow()
|
||||
db.session.commit()
|
||||
|
||||
return user, None
|
||||
|
||||
|
||||
def create_tokens(user: User, session_id: Optional[str] = None) -> Dict[str, Any]:
|
||||
"""Create JWT tokens for user"""
|
||||
# Generate JTIs
|
||||
access_jti = str(uuid.uuid4())
|
||||
refresh_jti = str(uuid.uuid4())
|
||||
|
||||
# Create tokens with custom claims
|
||||
identity = str(user.id)
|
||||
additional_claims = {
|
||||
'username': user.username,
|
||||
'role': user.role,
|
||||
'permissions': user.permissions or [],
|
||||
'session_id': session_id
|
||||
}
|
||||
|
||||
access_token = create_access_token(
|
||||
identity=identity,
|
||||
additional_claims=additional_claims,
|
||||
fresh=True
|
||||
)
|
||||
|
||||
refresh_token = create_refresh_token(
|
||||
identity=identity,
|
||||
additional_claims={'session_id': session_id}
|
||||
)
|
||||
|
||||
return {
|
||||
'access_token': access_token,
|
||||
'refresh_token': refresh_token,
|
||||
'token_type': 'Bearer',
|
||||
'expires_in': current_app.config['JWT_ACCESS_TOKEN_EXPIRES'].total_seconds()
|
||||
}
|
||||
|
||||
|
||||
def create_user_session(user: User, request_info: Dict[str, Any]) -> UserSession:
|
||||
"""Create a new user session"""
|
||||
session = UserSession(
|
||||
session_id=str(uuid.uuid4()),
|
||||
user_id=user.id,
|
||||
ip_address=request_info.get('ip_address'),
|
||||
user_agent=request_info.get('user_agent'),
|
||||
expires_at=datetime.utcnow() + timedelta(days=30)
|
||||
)
|
||||
|
||||
db.session.add(session)
|
||||
db.session.commit()
|
||||
|
||||
return session
|
||||
|
||||
|
||||
def log_login_attempt(user_id: Optional[uuid.UUID], success: bool, method: str,
|
||||
failure_reason: Optional[str] = None, session_id: Optional[str] = None,
|
||||
jwt_jti: Optional[str] = None) -> LoginHistory:
|
||||
"""Log a login attempt"""
|
||||
login_record = LoginHistory(
|
||||
user_id=user_id,
|
||||
login_method=method,
|
||||
success=success,
|
||||
failure_reason=failure_reason,
|
||||
session_id=session_id,
|
||||
jwt_jti=jwt_jti,
|
||||
ip_address=request.remote_addr,
|
||||
user_agent=request.headers.get('User-Agent')
|
||||
)
|
||||
|
||||
db.session.add(login_record)
|
||||
db.session.commit()
|
||||
|
||||
return login_record
|
||||
|
||||
|
||||
def revoke_token(jti: str, token_type: str, user_id: Optional[uuid.UUID] = None,
|
||||
reason: Optional[str] = None, expires_at: Optional[datetime] = None):
|
||||
"""Revoke a JWT token"""
|
||||
if not expires_at:
|
||||
# Default expiration based on token type
|
||||
if token_type == 'access':
|
||||
expires_at = datetime.utcnow() + current_app.config['JWT_ACCESS_TOKEN_EXPIRES']
|
||||
else:
|
||||
expires_at = datetime.utcnow() + current_app.config['JWT_REFRESH_TOKEN_EXPIRES']
|
||||
|
||||
revoked = RevokedToken(
|
||||
jti=jti,
|
||||
token_type=token_type,
|
||||
user_id=user_id,
|
||||
reason=reason,
|
||||
expires_at=expires_at
|
||||
)
|
||||
|
||||
db.session.add(revoked)
|
||||
db.session.commit()
|
||||
|
||||
|
||||
def get_current_user() -> Optional[User]:
|
||||
"""Get current authenticated user from JWT, API key, or session"""
|
||||
# Try JWT first
|
||||
try:
|
||||
verify_jwt_in_request(optional=True)
|
||||
user_id = get_jwt_identity()
|
||||
if user_id:
|
||||
user = User.query.get(user_id)
|
||||
if user and user.is_active and not user.is_suspended_now:
|
||||
# Update last active
|
||||
user.last_active_at = datetime.utcnow()
|
||||
db.session.commit()
|
||||
return user
|
||||
except:
|
||||
pass
|
||||
|
||||
# Try API key from header
|
||||
api_key = request.headers.get('X-API-Key')
|
||||
if api_key:
|
||||
user, _ = authenticate_api_key(api_key)
|
||||
if user:
|
||||
return user
|
||||
|
||||
# Try API key from query parameter
|
||||
api_key = request.args.get('api_key')
|
||||
if api_key:
|
||||
user, _ = authenticate_api_key(api_key)
|
||||
if user:
|
||||
return user
|
||||
|
||||
# Try session-based authentication (for admin panel)
|
||||
from flask import session
|
||||
if session.get('logged_in') and session.get('user_id'):
|
||||
# Check if it's the admin token user
|
||||
if session.get('user_id') == 'admin-token-user' and session.get('user_role') == 'admin':
|
||||
# Create a pseudo-admin user for session-based admin access
|
||||
admin_user = User.query.filter_by(role='admin').first()
|
||||
if admin_user:
|
||||
return admin_user
|
||||
else:
|
||||
# Create a temporary admin user object (not saved to DB)
|
||||
admin_user = User(
|
||||
id=uuid.uuid4(),
|
||||
username='admin',
|
||||
email='admin@talk2me.local',
|
||||
role='admin',
|
||||
is_active=True,
|
||||
is_verified=True,
|
||||
is_suspended=False,
|
||||
total_requests=0,
|
||||
total_translations=0,
|
||||
total_transcriptions=0,
|
||||
total_tts_requests=0
|
||||
)
|
||||
# Don't add to session, just return for authorization
|
||||
return admin_user
|
||||
else:
|
||||
# Regular user session
|
||||
user = User.query.get(session.get('user_id'))
|
||||
if user and user.is_active and not user.is_suspended_now:
|
||||
# Update last active
|
||||
user.last_active_at = datetime.utcnow()
|
||||
db.session.commit()
|
||||
return user
|
||||
|
||||
return None
|
||||
|
||||
|
||||
def require_auth(f: Callable) -> Callable:
|
||||
"""Decorator to require authentication (JWT, API key, or session)"""
|
||||
@functools.wraps(f)
|
||||
def decorated_function(*args, **kwargs):
|
||||
user = get_current_user()
|
||||
if not user:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Authentication required',
|
||||
'code': 'auth_required'
|
||||
}), 401
|
||||
|
||||
# Store user in g for access in route
|
||||
g.current_user = user
|
||||
|
||||
# Track usage only for database-backed users
|
||||
try:
|
||||
if hasattr(user, 'id') and db.session.query(User).filter_by(id=user.id).first():
|
||||
user.total_requests += 1
|
||||
db.session.commit()
|
||||
except Exception as e:
|
||||
# Ignore tracking errors for temporary users
|
||||
pass
|
||||
|
||||
return f(*args, **kwargs)
|
||||
|
||||
return decorated_function
|
||||
|
||||
|
||||
def require_admin(f: Callable) -> Callable:
|
||||
"""Decorator to require admin role"""
|
||||
@functools.wraps(f)
|
||||
@require_auth
|
||||
def decorated_function(*args, **kwargs):
|
||||
if not g.current_user.is_admin:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Admin access required',
|
||||
'code': 'admin_required'
|
||||
}), 403
|
||||
|
||||
return f(*args, **kwargs)
|
||||
|
||||
return decorated_function
|
||||
|
||||
|
||||
def require_permission(permission: str) -> Callable:
|
||||
"""Decorator to require specific permission"""
|
||||
def decorator(f: Callable) -> Callable:
|
||||
@functools.wraps(f)
|
||||
@require_auth
|
||||
def decorated_function(*args, **kwargs):
|
||||
if not g.current_user.has_permission(permission):
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': f'Permission required: {permission}',
|
||||
'code': 'permission_denied'
|
||||
}), 403
|
||||
|
||||
return f(*args, **kwargs)
|
||||
|
||||
return decorated_function
|
||||
|
||||
return decorator
|
||||
|
||||
|
||||
def require_verified(f: Callable) -> Callable:
|
||||
"""Decorator to require verified email"""
|
||||
@functools.wraps(f)
|
||||
@require_auth
|
||||
def decorated_function(*args, **kwargs):
|
||||
if not g.current_user.is_verified:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Email verification required',
|
||||
'code': 'verification_required'
|
||||
}), 403
|
||||
|
||||
return f(*args, **kwargs)
|
||||
|
||||
return decorated_function
|
||||
|
||||
|
||||
def get_user_rate_limits(user: User) -> Dict[str, int]:
|
||||
"""Get user-specific rate limits"""
|
||||
return {
|
||||
'per_minute': user.rate_limit_per_minute,
|
||||
'per_hour': user.rate_limit_per_hour,
|
||||
'per_day': user.rate_limit_per_day
|
||||
}
|
||||
|
||||
|
||||
def check_user_rate_limit(user: User, endpoint: str) -> tuple[bool, Optional[str]]:
|
||||
"""Check if user has exceeded rate limits"""
|
||||
# This would integrate with the existing rate limiter
|
||||
# For now, return True to allow requests
|
||||
return True, None
|
||||
|
||||
|
||||
def update_user_usage_stats(user: User, operation: str) -> None:
|
||||
"""Update user usage statistics"""
|
||||
user.total_requests += 1
|
||||
|
||||
if operation == 'translation':
|
||||
user.total_translations += 1
|
||||
elif operation == 'transcription':
|
||||
user.total_transcriptions += 1
|
||||
elif operation == 'tts':
|
||||
user.total_tts_requests += 1
|
||||
|
||||
user.last_active_at = datetime.utcnow()
|
||||
db.session.commit()
|
||||
|
||||
|
||||
def cleanup_expired_sessions() -> int:
|
||||
"""Clean up expired user sessions"""
|
||||
deleted = UserSession.query.filter(
|
||||
UserSession.expires_at < datetime.utcnow()
|
||||
).delete()
|
||||
db.session.commit()
|
||||
return deleted
|
||||
|
||||
|
||||
def cleanup_expired_tokens() -> int:
|
||||
"""Clean up expired revoked tokens"""
|
||||
return RevokedToken.cleanup_expired()
|
||||
|
||||
|
||||
def get_user_sessions(user_id: Union[str, uuid.UUID]) -> List[UserSession]:
|
||||
"""Get all active sessions for a user"""
|
||||
return UserSession.query.filter_by(
|
||||
user_id=user_id
|
||||
).filter(
|
||||
UserSession.expires_at > datetime.utcnow()
|
||||
).order_by(UserSession.last_active_at.desc()).all()
|
||||
|
||||
|
||||
def revoke_user_sessions(user_id: Union[str, uuid.UUID], except_session: Optional[str] = None) -> int:
|
||||
"""Revoke all sessions for a user"""
|
||||
sessions = UserSession.query.filter_by(user_id=user_id)
|
||||
|
||||
if except_session:
|
||||
sessions = sessions.filter(UserSession.session_id != except_session)
|
||||
|
||||
count = 0
|
||||
for session in sessions:
|
||||
# Revoke associated tokens
|
||||
if session.access_token_jti:
|
||||
revoke_token(session.access_token_jti, 'access', user_id, 'Session revoked')
|
||||
if session.refresh_token_jti:
|
||||
revoke_token(session.refresh_token_jti, 'refresh', user_id, 'Session revoked')
|
||||
count += 1
|
||||
|
||||
# Delete sessions
|
||||
sessions.delete()
|
||||
db.session.commit()
|
||||
|
||||
return count
|
366
auth_models.py
Normal file
366
auth_models.py
Normal file
@ -0,0 +1,366 @@
|
||||
"""Authentication models for Talk2Me application"""
|
||||
import uuid
|
||||
import secrets
|
||||
from datetime import datetime, timedelta
|
||||
from typing import Optional, Dict, Any, List
|
||||
from flask_sqlalchemy import SQLAlchemy
|
||||
from flask_bcrypt import Bcrypt
|
||||
from sqlalchemy import Index, text, func
|
||||
from sqlalchemy.dialects.postgresql import UUID, JSONB, ENUM
|
||||
from sqlalchemy.ext.hybrid import hybrid_property
|
||||
from sqlalchemy.orm import relationship
|
||||
|
||||
from database import db
|
||||
|
||||
bcrypt = Bcrypt()
|
||||
|
||||
|
||||
class User(db.Model):
|
||||
"""User account model with authentication and authorization"""
|
||||
__tablename__ = 'users'
|
||||
|
||||
id = db.Column(UUID(as_uuid=True), primary_key=True, default=uuid.uuid4)
|
||||
email = db.Column(db.String(255), unique=True, nullable=False, index=True)
|
||||
username = db.Column(db.String(100), unique=True, nullable=False, index=True)
|
||||
password_hash = db.Column(db.String(255), nullable=False)
|
||||
|
||||
# User profile
|
||||
full_name = db.Column(db.String(255), nullable=True)
|
||||
avatar_url = db.Column(db.String(500), nullable=True)
|
||||
|
||||
# API Key - unique per user
|
||||
api_key = db.Column(db.String(64), unique=True, nullable=False, index=True)
|
||||
api_key_created_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow)
|
||||
|
||||
# Account status
|
||||
is_active = db.Column(db.Boolean, default=True, nullable=False)
|
||||
is_verified = db.Column(db.Boolean, default=False, nullable=False)
|
||||
is_suspended = db.Column(db.Boolean, default=False, nullable=False)
|
||||
suspension_reason = db.Column(db.Text, nullable=True)
|
||||
suspended_at = db.Column(db.DateTime, nullable=True)
|
||||
suspended_until = db.Column(db.DateTime, nullable=True)
|
||||
|
||||
# Role and permissions
|
||||
role = db.Column(db.String(20), nullable=False, default='user') # admin, user
|
||||
permissions = db.Column(JSONB, default=[], nullable=False) # Additional granular permissions
|
||||
|
||||
# Usage limits (per user)
|
||||
rate_limit_per_minute = db.Column(db.Integer, default=30, nullable=False)
|
||||
rate_limit_per_hour = db.Column(db.Integer, default=500, nullable=False)
|
||||
rate_limit_per_day = db.Column(db.Integer, default=5000, nullable=False)
|
||||
|
||||
# Usage tracking
|
||||
total_requests = db.Column(db.Integer, default=0, nullable=False)
|
||||
total_translations = db.Column(db.Integer, default=0, nullable=False)
|
||||
total_transcriptions = db.Column(db.Integer, default=0, nullable=False)
|
||||
total_tts_requests = db.Column(db.Integer, default=0, nullable=False)
|
||||
|
||||
# Timestamps
|
||||
created_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow)
|
||||
updated_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow, onupdate=datetime.utcnow)
|
||||
last_login_at = db.Column(db.DateTime, nullable=True)
|
||||
last_active_at = db.Column(db.DateTime, nullable=True)
|
||||
|
||||
# Security
|
||||
password_changed_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow)
|
||||
failed_login_attempts = db.Column(db.Integer, default=0, nullable=False)
|
||||
locked_until = db.Column(db.DateTime, nullable=True)
|
||||
|
||||
# Settings
|
||||
settings = db.Column(JSONB, default={}, nullable=False)
|
||||
|
||||
# Relationships
|
||||
login_history = relationship('LoginHistory', back_populates='user', cascade='all, delete-orphan')
|
||||
sessions = relationship('UserSession', back_populates='user', cascade='all, delete-orphan')
|
||||
|
||||
__table_args__ = (
|
||||
Index('idx_users_email_active', 'email', 'is_active'),
|
||||
Index('idx_users_role_active', 'role', 'is_active'),
|
||||
Index('idx_users_created_at', 'created_at'),
|
||||
)
|
||||
|
||||
def __init__(self, **kwargs):
|
||||
super(User, self).__init__(**kwargs)
|
||||
if not self.api_key:
|
||||
self.api_key = self.generate_api_key()
|
||||
|
||||
@staticmethod
|
||||
def generate_api_key() -> str:
|
||||
"""Generate a secure API key"""
|
||||
return f"tk_{secrets.token_urlsafe(32)}"
|
||||
|
||||
def regenerate_api_key(self) -> str:
|
||||
"""Regenerate user's API key"""
|
||||
self.api_key = self.generate_api_key()
|
||||
self.api_key_created_at = datetime.utcnow()
|
||||
return self.api_key
|
||||
|
||||
def set_password(self, password: str) -> None:
|
||||
"""Hash and set user password"""
|
||||
self.password_hash = bcrypt.generate_password_hash(password).decode('utf-8')
|
||||
self.password_changed_at = datetime.utcnow()
|
||||
|
||||
def check_password(self, password: str) -> bool:
|
||||
"""Check if provided password matches hash"""
|
||||
return bcrypt.check_password_hash(self.password_hash, password)
|
||||
|
||||
@hybrid_property
|
||||
def is_admin(self) -> bool:
|
||||
"""Check if user has admin role"""
|
||||
return self.role == 'admin'
|
||||
|
||||
@hybrid_property
|
||||
def is_locked(self) -> bool:
|
||||
"""Check if account is locked due to failed login attempts"""
|
||||
if self.locked_until is None:
|
||||
return False
|
||||
return datetime.utcnow() < self.locked_until
|
||||
|
||||
@hybrid_property
|
||||
def is_suspended_now(self) -> bool:
|
||||
"""Check if account is currently suspended"""
|
||||
if not self.is_suspended:
|
||||
return False
|
||||
if self.suspended_until is None:
|
||||
return True # Indefinite suspension
|
||||
return datetime.utcnow() < self.suspended_until
|
||||
|
||||
def can_login(self) -> tuple[bool, Optional[str]]:
|
||||
"""Check if user can login"""
|
||||
if not self.is_active:
|
||||
return False, "Account is deactivated"
|
||||
if self.is_locked:
|
||||
return False, "Account is locked due to failed login attempts"
|
||||
if self.is_suspended_now:
|
||||
return False, f"Account is suspended: {self.suspension_reason or 'Policy violation'}"
|
||||
return True, None
|
||||
|
||||
def record_login_attempt(self, success: bool) -> None:
|
||||
"""Record login attempt and handle lockout"""
|
||||
if success:
|
||||
self.failed_login_attempts = 0
|
||||
self.locked_until = None
|
||||
self.last_login_at = datetime.utcnow()
|
||||
else:
|
||||
self.failed_login_attempts += 1
|
||||
# Lock account after 5 failed attempts
|
||||
if self.failed_login_attempts >= 5:
|
||||
self.locked_until = datetime.utcnow() + timedelta(minutes=30)
|
||||
|
||||
def has_permission(self, permission: str) -> bool:
|
||||
"""Check if user has specific permission"""
|
||||
if self.is_admin:
|
||||
return True # Admins have all permissions
|
||||
return permission in (self.permissions or [])
|
||||
|
||||
def add_permission(self, permission: str) -> None:
|
||||
"""Add permission to user"""
|
||||
if self.permissions is None:
|
||||
self.permissions = []
|
||||
if permission not in self.permissions:
|
||||
self.permissions = self.permissions + [permission]
|
||||
|
||||
def remove_permission(self, permission: str) -> None:
|
||||
"""Remove permission from user"""
|
||||
if self.permissions and permission in self.permissions:
|
||||
self.permissions = [p for p in self.permissions if p != permission]
|
||||
|
||||
def suspend(self, reason: str, until: Optional[datetime] = None) -> None:
|
||||
"""Suspend user account"""
|
||||
self.is_suspended = True
|
||||
self.suspension_reason = reason
|
||||
self.suspended_at = datetime.utcnow()
|
||||
self.suspended_until = until
|
||||
|
||||
def unsuspend(self) -> None:
|
||||
"""Unsuspend user account"""
|
||||
self.is_suspended = False
|
||||
self.suspension_reason = None
|
||||
self.suspended_at = None
|
||||
self.suspended_until = None
|
||||
|
||||
def to_dict(self, include_sensitive: bool = False) -> Dict[str, Any]:
|
||||
"""Convert user to dictionary"""
|
||||
data = {
|
||||
'id': str(self.id),
|
||||
'email': self.email,
|
||||
'username': self.username,
|
||||
'full_name': self.full_name,
|
||||
'avatar_url': self.avatar_url,
|
||||
'role': self.role,
|
||||
'is_active': self.is_active,
|
||||
'is_verified': self.is_verified,
|
||||
'is_suspended': self.is_suspended_now,
|
||||
'created_at': self.created_at.isoformat(),
|
||||
'last_login_at': self.last_login_at.isoformat() if self.last_login_at else None,
|
||||
'last_active_at': self.last_active_at.isoformat() if self.last_active_at else None,
|
||||
'total_requests': self.total_requests,
|
||||
'total_translations': self.total_translations,
|
||||
'total_transcriptions': self.total_transcriptions,
|
||||
'total_tts_requests': self.total_tts_requests,
|
||||
'settings': self.settings or {}
|
||||
}
|
||||
|
||||
if include_sensitive:
|
||||
data.update({
|
||||
'api_key': self.api_key,
|
||||
'api_key_created_at': self.api_key_created_at.isoformat(),
|
||||
'permissions': self.permissions or [],
|
||||
'rate_limit_per_minute': self.rate_limit_per_minute,
|
||||
'rate_limit_per_hour': self.rate_limit_per_hour,
|
||||
'rate_limit_per_day': self.rate_limit_per_day,
|
||||
'suspension_reason': self.suspension_reason,
|
||||
'suspended_until': self.suspended_until.isoformat() if self.suspended_until else None,
|
||||
'failed_login_attempts': self.failed_login_attempts,
|
||||
'is_locked': self.is_locked
|
||||
})
|
||||
|
||||
return data
|
||||
|
||||
|
||||
class LoginHistory(db.Model):
|
||||
"""Track user login history for security auditing"""
|
||||
__tablename__ = 'login_history'
|
||||
|
||||
id = db.Column(UUID(as_uuid=True), primary_key=True, default=uuid.uuid4)
|
||||
user_id = db.Column(UUID(as_uuid=True), db.ForeignKey('users.id'), nullable=False, index=True)
|
||||
|
||||
# Login details
|
||||
login_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow)
|
||||
logout_at = db.Column(db.DateTime, nullable=True)
|
||||
login_method = db.Column(db.String(20), nullable=False) # password, api_key, jwt
|
||||
success = db.Column(db.Boolean, nullable=False)
|
||||
failure_reason = db.Column(db.String(255), nullable=True)
|
||||
|
||||
# Session info
|
||||
session_id = db.Column(db.String(255), nullable=True, index=True)
|
||||
jwt_jti = db.Column(db.String(255), nullable=True, index=True) # JWT ID for revocation
|
||||
|
||||
# Client info
|
||||
ip_address = db.Column(db.String(45), nullable=False)
|
||||
user_agent = db.Column(db.String(500), nullable=True)
|
||||
device_info = db.Column(JSONB, nullable=True) # Parsed user agent info
|
||||
|
||||
# Location info (if available)
|
||||
country = db.Column(db.String(2), nullable=True)
|
||||
city = db.Column(db.String(100), nullable=True)
|
||||
|
||||
# Security flags
|
||||
is_suspicious = db.Column(db.Boolean, default=False, nullable=False)
|
||||
security_notes = db.Column(db.Text, nullable=True)
|
||||
|
||||
# Relationship
|
||||
user = relationship('User', back_populates='login_history')
|
||||
|
||||
__table_args__ = (
|
||||
Index('idx_login_history_user_time', 'user_id', 'login_at'),
|
||||
Index('idx_login_history_session', 'session_id'),
|
||||
Index('idx_login_history_ip', 'ip_address'),
|
||||
)
|
||||
|
||||
def to_dict(self) -> Dict[str, Any]:
|
||||
"""Convert login history to dictionary"""
|
||||
return {
|
||||
'id': str(self.id),
|
||||
'user_id': str(self.user_id),
|
||||
'login_at': self.login_at.isoformat(),
|
||||
'logout_at': self.logout_at.isoformat() if self.logout_at else None,
|
||||
'login_method': self.login_method,
|
||||
'success': self.success,
|
||||
'failure_reason': self.failure_reason,
|
||||
'session_id': self.session_id,
|
||||
'ip_address': self.ip_address,
|
||||
'user_agent': self.user_agent,
|
||||
'device_info': self.device_info,
|
||||
'country': self.country,
|
||||
'city': self.city,
|
||||
'is_suspicious': self.is_suspicious,
|
||||
'security_notes': self.security_notes
|
||||
}
|
||||
|
||||
|
||||
class UserSession(db.Model):
|
||||
"""Active user sessions for session management"""
|
||||
__tablename__ = 'user_sessions'
|
||||
|
||||
id = db.Column(UUID(as_uuid=True), primary_key=True, default=uuid.uuid4)
|
||||
session_id = db.Column(db.String(255), unique=True, nullable=False, index=True)
|
||||
user_id = db.Column(UUID(as_uuid=True), db.ForeignKey('users.id'), nullable=False, index=True)
|
||||
|
||||
# JWT tokens
|
||||
access_token_jti = db.Column(db.String(255), nullable=True, index=True)
|
||||
refresh_token_jti = db.Column(db.String(255), nullable=True, index=True)
|
||||
|
||||
# Session info
|
||||
created_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow)
|
||||
last_active_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow)
|
||||
expires_at = db.Column(db.DateTime, nullable=False)
|
||||
|
||||
# Client info
|
||||
ip_address = db.Column(db.String(45), nullable=False)
|
||||
user_agent = db.Column(db.String(500), nullable=True)
|
||||
|
||||
# Session data
|
||||
data = db.Column(JSONB, default={}, nullable=False)
|
||||
|
||||
# Relationship
|
||||
user = relationship('User', back_populates='sessions')
|
||||
|
||||
__table_args__ = (
|
||||
Index('idx_user_sessions_user_active', 'user_id', 'expires_at'),
|
||||
Index('idx_user_sessions_token', 'access_token_jti'),
|
||||
)
|
||||
|
||||
@hybrid_property
|
||||
def is_expired(self) -> bool:
|
||||
"""Check if session is expired"""
|
||||
return datetime.utcnow() > self.expires_at
|
||||
|
||||
def refresh(self, duration_hours: int = 24) -> None:
|
||||
"""Refresh session expiration"""
|
||||
self.last_active_at = datetime.utcnow()
|
||||
self.expires_at = datetime.utcnow() + timedelta(hours=duration_hours)
|
||||
|
||||
def to_dict(self) -> Dict[str, Any]:
|
||||
"""Convert session to dictionary"""
|
||||
return {
|
||||
'id': str(self.id),
|
||||
'session_id': self.session_id,
|
||||
'user_id': str(self.user_id),
|
||||
'created_at': self.created_at.isoformat(),
|
||||
'last_active_at': self.last_active_at.isoformat(),
|
||||
'expires_at': self.expires_at.isoformat(),
|
||||
'is_expired': self.is_expired,
|
||||
'ip_address': self.ip_address,
|
||||
'user_agent': self.user_agent,
|
||||
'data': self.data or {}
|
||||
}
|
||||
|
||||
|
||||
class RevokedToken(db.Model):
|
||||
"""Store revoked JWT tokens for blacklisting"""
|
||||
__tablename__ = 'revoked_tokens'
|
||||
|
||||
id = db.Column(UUID(as_uuid=True), primary_key=True, default=uuid.uuid4)
|
||||
jti = db.Column(db.String(255), unique=True, nullable=False, index=True)
|
||||
token_type = db.Column(db.String(20), nullable=False) # access, refresh
|
||||
user_id = db.Column(UUID(as_uuid=True), nullable=True, index=True)
|
||||
revoked_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow)
|
||||
expires_at = db.Column(db.DateTime, nullable=False) # When token would have expired
|
||||
reason = db.Column(db.String(255), nullable=True)
|
||||
|
||||
__table_args__ = (
|
||||
Index('idx_revoked_tokens_expires', 'expires_at'),
|
||||
)
|
||||
|
||||
@classmethod
|
||||
def is_token_revoked(cls, jti: str) -> bool:
|
||||
"""Check if a token JTI is revoked"""
|
||||
return cls.query.filter_by(jti=jti).first() is not None
|
||||
|
||||
@classmethod
|
||||
def cleanup_expired(cls) -> int:
|
||||
"""Remove revoked tokens that have expired anyway"""
|
||||
deleted = cls.query.filter(cls.expires_at < datetime.utcnow()).delete()
|
||||
db.session.commit()
|
||||
return deleted
|
930
auth_routes.py
Normal file
930
auth_routes.py
Normal file
@ -0,0 +1,930 @@
|
||||
"""Authentication and user management routes"""
|
||||
import os
|
||||
import logging
|
||||
from datetime import datetime, timedelta
|
||||
from flask import Blueprint, request, jsonify, g
|
||||
from flask_jwt_extended import jwt_required, get_jwt_identity, get_jwt
|
||||
from sqlalchemy import or_, func
|
||||
from werkzeug.exceptions import BadRequest
|
||||
|
||||
from database import db
|
||||
from auth_models import User, LoginHistory, UserSession
|
||||
from auth import (
|
||||
create_user, authenticate_user, create_tokens, create_user_session,
|
||||
revoke_token, get_current_user, require_admin,
|
||||
require_auth, revoke_user_sessions, update_user_usage_stats
|
||||
)
|
||||
from rate_limiter import rate_limit
|
||||
from validators import Validators
|
||||
from error_logger import log_exception
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
auth_bp = Blueprint('auth', __name__)
|
||||
|
||||
|
||||
@auth_bp.route('/login', methods=['POST'])
|
||||
@rate_limit(requests_per_minute=5, requests_per_hour=30)
|
||||
def login():
|
||||
"""User login endpoint"""
|
||||
try:
|
||||
data = request.get_json()
|
||||
|
||||
# Validate input
|
||||
username_or_email = data.get('username') or data.get('email')
|
||||
password = data.get('password')
|
||||
|
||||
if not username_or_email or not password:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Username/email and password required'
|
||||
}), 400
|
||||
|
||||
# Authenticate user
|
||||
user, error = authenticate_user(username_or_email, password)
|
||||
if error:
|
||||
# Log failed attempt
|
||||
login_record = LoginHistory(
|
||||
user_id=None,
|
||||
login_method='password',
|
||||
success=False,
|
||||
failure_reason=error,
|
||||
ip_address=request.remote_addr,
|
||||
user_agent=request.headers.get('User-Agent')
|
||||
)
|
||||
db.session.add(login_record)
|
||||
db.session.commit()
|
||||
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': error
|
||||
}), 401
|
||||
|
||||
# Create session
|
||||
session = create_user_session(user, {
|
||||
'ip_address': request.remote_addr,
|
||||
'user_agent': request.headers.get('User-Agent')
|
||||
})
|
||||
|
||||
# Create tokens
|
||||
tokens = create_tokens(user, session.session_id)
|
||||
|
||||
# Note: We can't get JWT payload here since we haven't set the JWT context yet
|
||||
# The session JTI will be updated on the next authenticated request
|
||||
db.session.commit()
|
||||
|
||||
# Log successful login with request info
|
||||
login_record = LoginHistory(
|
||||
user_id=user.id,
|
||||
login_method='password',
|
||||
success=True,
|
||||
session_id=session.session_id,
|
||||
ip_address=request.remote_addr,
|
||||
user_agent=request.headers.get('User-Agent')
|
||||
)
|
||||
db.session.add(login_record)
|
||||
db.session.commit()
|
||||
|
||||
# Store user info in Flask session for web access
|
||||
from flask import session as flask_session
|
||||
flask_session['user_id'] = str(user.id)
|
||||
flask_session['username'] = user.username
|
||||
flask_session['user_role'] = user.role
|
||||
flask_session['logged_in'] = True
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'user': user.to_dict(),
|
||||
'tokens': tokens,
|
||||
'session_id': session.session_id
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Login error")
|
||||
# In development, show the actual error
|
||||
import os
|
||||
if os.environ.get('FLASK_ENV') == 'development':
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': f'Login failed: {str(e)}'
|
||||
}), 500
|
||||
else:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Login failed'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/logout', methods=['POST'])
|
||||
@jwt_required()
|
||||
def logout():
|
||||
"""User logout endpoint"""
|
||||
try:
|
||||
jti = get_jwt()["jti"]
|
||||
user_id = get_jwt_identity()
|
||||
|
||||
# Revoke the access token
|
||||
revoke_token(jti, 'access', user_id, 'User logout')
|
||||
|
||||
# Update login history
|
||||
session_id = get_jwt().get('session_id')
|
||||
if session_id:
|
||||
login_record = LoginHistory.query.filter_by(
|
||||
session_id=session_id,
|
||||
logout_at=None
|
||||
).first()
|
||||
if login_record:
|
||||
login_record.logout_at = datetime.utcnow()
|
||||
db.session.commit()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'message': 'Successfully logged out'
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Logout error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Logout failed'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/refresh', methods=['POST'])
|
||||
@jwt_required(refresh=True)
|
||||
def refresh_token():
|
||||
"""Refresh access token"""
|
||||
try:
|
||||
user_id = get_jwt_identity()
|
||||
user = User.query.get(user_id)
|
||||
|
||||
if not user or not user.is_active:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Invalid user'
|
||||
}), 401
|
||||
|
||||
# Check if user can login
|
||||
can_login, reason = user.can_login()
|
||||
if not can_login:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': reason
|
||||
}), 401
|
||||
|
||||
# Create new access token
|
||||
session_id = get_jwt().get('session_id')
|
||||
tokens = create_tokens(user, session_id)
|
||||
|
||||
# Update session if exists
|
||||
if session_id:
|
||||
session = UserSession.query.filter_by(session_id=session_id).first()
|
||||
if session:
|
||||
session.refresh()
|
||||
db.session.commit()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'access_token': tokens['access_token'],
|
||||
'expires_in': tokens['expires_in']
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Token refresh error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Token refresh failed'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/profile', methods=['GET'])
|
||||
@require_auth
|
||||
def get_profile():
|
||||
"""Get current user profile"""
|
||||
try:
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'user': g.current_user.to_dict(include_sensitive=True)
|
||||
})
|
||||
except Exception as e:
|
||||
log_exception(e, "Profile fetch error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to fetch profile'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/profile', methods=['PUT'])
|
||||
@require_auth
|
||||
def update_profile():
|
||||
"""Update user profile"""
|
||||
try:
|
||||
data = request.get_json()
|
||||
user = g.current_user
|
||||
|
||||
# Update allowed fields
|
||||
if 'full_name' in data:
|
||||
user.full_name = Validators.sanitize_text(data['full_name'], max_length=255)
|
||||
|
||||
if 'avatar_url' in data:
|
||||
validated_url = Validators.validate_url(data['avatar_url'])
|
||||
if validated_url:
|
||||
user.avatar_url = validated_url
|
||||
|
||||
if 'settings' in data and isinstance(data['settings'], dict):
|
||||
user.settings = {**user.settings, **data['settings']}
|
||||
|
||||
db.session.commit()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'user': user.to_dict(include_sensitive=True)
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Profile update error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to update profile'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/change-password', methods=['POST'])
|
||||
@require_auth
|
||||
def change_password():
|
||||
"""Change user password"""
|
||||
try:
|
||||
data = request.get_json()
|
||||
user = g.current_user
|
||||
|
||||
current_password = data.get('current_password')
|
||||
new_password = data.get('new_password')
|
||||
|
||||
if not current_password or not new_password:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Current and new passwords required'
|
||||
}), 400
|
||||
|
||||
# Verify current password
|
||||
if not user.check_password(current_password):
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Invalid current password'
|
||||
}), 401
|
||||
|
||||
# Validate new password
|
||||
if len(new_password) < 8:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Password must be at least 8 characters'
|
||||
}), 400
|
||||
|
||||
# Update password
|
||||
user.set_password(new_password)
|
||||
db.session.commit()
|
||||
|
||||
# Revoke all sessions except current
|
||||
session_id = get_jwt().get('session_id') if hasattr(g, 'jwt_payload') else None
|
||||
revoked_count = revoke_user_sessions(user.id, except_session=session_id)
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'message': 'Password changed successfully',
|
||||
'revoked_sessions': revoked_count
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Password change error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to change password'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/regenerate-api-key', methods=['POST'])
|
||||
@require_auth
|
||||
def regenerate_api_key():
|
||||
"""Regenerate user's API key"""
|
||||
try:
|
||||
user = g.current_user
|
||||
new_key = user.regenerate_api_key()
|
||||
db.session.commit()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'api_key': new_key,
|
||||
'created_at': user.api_key_created_at.isoformat()
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "API key regeneration error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to regenerate API key'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/sessions', methods=['GET'])
|
||||
@require_auth
|
||||
def get_user_sessions():
|
||||
"""Get user's active sessions"""
|
||||
try:
|
||||
sessions = UserSession.query.filter_by(
|
||||
user_id=g.current_user.id
|
||||
).filter(
|
||||
UserSession.expires_at > datetime.utcnow()
|
||||
).order_by(UserSession.last_active_at.desc()).all()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'sessions': [s.to_dict() for s in sessions]
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Sessions fetch error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to fetch sessions'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/sessions/<session_id>', methods=['DELETE'])
|
||||
@require_auth
|
||||
def revoke_session(session_id):
|
||||
"""Revoke a specific session"""
|
||||
try:
|
||||
session = UserSession.query.filter_by(
|
||||
session_id=session_id,
|
||||
user_id=g.current_user.id
|
||||
).first()
|
||||
|
||||
if not session:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Session not found'
|
||||
}), 404
|
||||
|
||||
# Revoke tokens
|
||||
if session.access_token_jti:
|
||||
revoke_token(session.access_token_jti, 'access', g.current_user.id, 'Session revoked by user')
|
||||
if session.refresh_token_jti:
|
||||
revoke_token(session.refresh_token_jti, 'refresh', g.current_user.id, 'Session revoked by user')
|
||||
|
||||
# Delete session
|
||||
db.session.delete(session)
|
||||
db.session.commit()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'message': 'Session revoked successfully'
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Session revocation error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to revoke session'
|
||||
}), 500
|
||||
|
||||
|
||||
# Admin endpoints for user management
|
||||
|
||||
@auth_bp.route('/admin/users', methods=['GET'])
|
||||
@require_admin
|
||||
def admin_list_users():
|
||||
"""List all users (admin only)"""
|
||||
try:
|
||||
# Get query parameters
|
||||
page = request.args.get('page', 1, type=int)
|
||||
per_page = request.args.get('per_page', 20, type=int)
|
||||
search = request.args.get('search', '')
|
||||
role = request.args.get('role')
|
||||
status = request.args.get('status')
|
||||
sort_by = request.args.get('sort_by', 'created_at')
|
||||
sort_order = request.args.get('sort_order', 'desc')
|
||||
|
||||
# Build query
|
||||
query = User.query
|
||||
|
||||
# Debug logging
|
||||
logger.info(f"Admin user list query parameters: page={page}, per_page={per_page}, search={search}, role={role}, status={status}, sort_by={sort_by}, sort_order={sort_order}")
|
||||
|
||||
# Search filter
|
||||
if search:
|
||||
search_term = f'%{search}%'
|
||||
query = query.filter(or_(
|
||||
User.email.ilike(search_term),
|
||||
User.username.ilike(search_term),
|
||||
User.full_name.ilike(search_term)
|
||||
))
|
||||
logger.info(f"Applied search filter: {search_term}")
|
||||
|
||||
# Role filter
|
||||
if role:
|
||||
query = query.filter(User.role == role)
|
||||
logger.info(f"Applied role filter: {role}")
|
||||
|
||||
# Status filter
|
||||
if status == 'active':
|
||||
query = query.filter(User.is_active == True, User.is_suspended == False)
|
||||
logger.info(f"Applied status filter: active")
|
||||
elif status == 'suspended':
|
||||
query = query.filter(User.is_suspended == True)
|
||||
logger.info(f"Applied status filter: suspended")
|
||||
elif status == 'inactive':
|
||||
query = query.filter(User.is_active == False)
|
||||
logger.info(f"Applied status filter: inactive")
|
||||
|
||||
# Sorting
|
||||
order_column = getattr(User, sort_by, User.created_at)
|
||||
if sort_order == 'desc':
|
||||
query = query.order_by(order_column.desc())
|
||||
else:
|
||||
query = query.order_by(order_column.asc())
|
||||
logger.info(f"Applied sorting: {sort_by} {sort_order}")
|
||||
|
||||
# Log the SQL query being generated
|
||||
try:
|
||||
sql = str(query.statement.compile(compile_kwargs={"literal_binds": True}))
|
||||
logger.info(f"Generated SQL query: {sql}")
|
||||
except Exception as e:
|
||||
logger.warning(f"Could not log SQL query: {e}")
|
||||
|
||||
# Count total results before pagination
|
||||
total_count = query.count()
|
||||
logger.info(f"Total users matching query (before pagination): {total_count}")
|
||||
|
||||
# Get all users without pagination for debugging
|
||||
all_matching_users = query.all()
|
||||
logger.info(f"All matching users: {[u.username for u in all_matching_users[:10]]}") # Log first 10 usernames
|
||||
|
||||
# Paginate
|
||||
pagination = query.paginate(page=page, per_page=per_page, error_out=False)
|
||||
|
||||
# Debug logging for results
|
||||
logger.info(f"Query returned {pagination.total} total users, showing {len(pagination.items)} on page {pagination.page}")
|
||||
logger.info(f"Pagination items: {[u.username for u in pagination.items]}")
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'users': [u.to_dict(include_sensitive=True) for u in pagination.items],
|
||||
'pagination': {
|
||||
'page': pagination.page,
|
||||
'per_page': pagination.per_page,
|
||||
'total': pagination.total,
|
||||
'pages': pagination.pages
|
||||
}
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Admin user list error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to fetch users'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/admin/users', methods=['POST'])
|
||||
@require_admin
|
||||
def admin_create_user():
|
||||
"""Create a new user (admin only)"""
|
||||
try:
|
||||
data = request.get_json()
|
||||
|
||||
# Validate required fields
|
||||
email = data.get('email')
|
||||
username = data.get('username')
|
||||
password = data.get('password')
|
||||
|
||||
if not email or not username or not password:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Email, username, and password are required'
|
||||
}), 400
|
||||
|
||||
# Validate email
|
||||
if not Validators.validate_email(email):
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Invalid email address'
|
||||
}), 400
|
||||
|
||||
# Create user
|
||||
user, error = create_user(
|
||||
email=email,
|
||||
username=username,
|
||||
password=password,
|
||||
full_name=data.get('full_name'),
|
||||
role=data.get('role', 'user'),
|
||||
is_verified=data.get('is_verified', False)
|
||||
)
|
||||
|
||||
if error:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': error
|
||||
}), 400
|
||||
|
||||
# Set additional properties
|
||||
if 'rate_limit_per_minute' in data:
|
||||
user.rate_limit_per_minute = data['rate_limit_per_minute']
|
||||
if 'rate_limit_per_hour' in data:
|
||||
user.rate_limit_per_hour = data['rate_limit_per_hour']
|
||||
if 'rate_limit_per_day' in data:
|
||||
user.rate_limit_per_day = data['rate_limit_per_day']
|
||||
if 'permissions' in data:
|
||||
user.permissions = data['permissions']
|
||||
|
||||
db.session.commit()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'user': user.to_dict(include_sensitive=True)
|
||||
}), 201
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Admin user creation error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to create user'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/admin/users/<user_id>', methods=['GET'])
|
||||
@require_admin
|
||||
def admin_get_user(user_id):
|
||||
"""Get user details (admin only)"""
|
||||
try:
|
||||
user = User.query.get(user_id)
|
||||
if not user:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'User not found'
|
||||
}), 404
|
||||
|
||||
# Get additional info
|
||||
login_history = LoginHistory.query.filter_by(
|
||||
user_id=user.id
|
||||
).order_by(LoginHistory.login_at.desc()).limit(10).all()
|
||||
|
||||
active_sessions = UserSession.query.filter_by(
|
||||
user_id=user.id
|
||||
).filter(
|
||||
UserSession.expires_at > datetime.utcnow()
|
||||
).all()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'user': user.to_dict(include_sensitive=True),
|
||||
'login_history': [l.to_dict() for l in login_history],
|
||||
'active_sessions': [s.to_dict() for s in active_sessions]
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Admin user fetch error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to fetch user'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/admin/users/<user_id>', methods=['PUT'])
|
||||
@require_admin
|
||||
def admin_update_user(user_id):
|
||||
"""Update user (admin only)"""
|
||||
try:
|
||||
user = User.query.get(user_id)
|
||||
if not user:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'User not found'
|
||||
}), 404
|
||||
|
||||
data = request.get_json()
|
||||
|
||||
# Update allowed fields
|
||||
if 'email' in data:
|
||||
if Validators.validate_email(data['email']):
|
||||
user.email = data['email']
|
||||
|
||||
if 'username' in data:
|
||||
user.username = data['username']
|
||||
|
||||
if 'full_name' in data:
|
||||
user.full_name = data['full_name']
|
||||
|
||||
if 'role' in data and data['role'] in ['admin', 'user']:
|
||||
user.role = data['role']
|
||||
|
||||
if 'is_active' in data:
|
||||
user.is_active = data['is_active']
|
||||
|
||||
if 'is_verified' in data:
|
||||
user.is_verified = data['is_verified']
|
||||
|
||||
if 'permissions' in data:
|
||||
user.permissions = data['permissions']
|
||||
|
||||
if 'rate_limit_per_minute' in data:
|
||||
user.rate_limit_per_minute = data['rate_limit_per_minute']
|
||||
|
||||
if 'rate_limit_per_hour' in data:
|
||||
user.rate_limit_per_hour = data['rate_limit_per_hour']
|
||||
|
||||
if 'rate_limit_per_day' in data:
|
||||
user.rate_limit_per_day = data['rate_limit_per_day']
|
||||
|
||||
db.session.commit()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'user': user.to_dict(include_sensitive=True)
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Admin user update error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to update user'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/admin/users/<user_id>', methods=['DELETE'])
|
||||
@require_admin
|
||||
def admin_delete_user(user_id):
|
||||
"""Delete user (admin only)"""
|
||||
try:
|
||||
user = User.query.get(user_id)
|
||||
if not user:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'User not found'
|
||||
}), 404
|
||||
|
||||
# Don't allow deleting admin users
|
||||
if user.is_admin:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Cannot delete admin users'
|
||||
}), 403
|
||||
|
||||
# Revoke all sessions
|
||||
revoke_user_sessions(user.id)
|
||||
|
||||
# Delete user (cascades to related records)
|
||||
db.session.delete(user)
|
||||
db.session.commit()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'message': 'User deleted successfully'
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Admin user deletion error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to delete user'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/admin/users/<user_id>/suspend', methods=['POST'])
|
||||
@require_admin
|
||||
def admin_suspend_user(user_id):
|
||||
"""Suspend user account (admin only)"""
|
||||
try:
|
||||
user = User.query.get(user_id)
|
||||
if not user:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'User not found'
|
||||
}), 404
|
||||
|
||||
data = request.get_json()
|
||||
reason = data.get('reason', 'Policy violation')
|
||||
until = data.get('until') # ISO datetime string or None for indefinite
|
||||
|
||||
# Parse until date if provided
|
||||
suspend_until = None
|
||||
if until:
|
||||
try:
|
||||
suspend_until = datetime.fromisoformat(until.replace('Z', '+00:00'))
|
||||
except:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Invalid date format for until'
|
||||
}), 400
|
||||
|
||||
# Suspend user
|
||||
user.suspend(reason, suspend_until)
|
||||
|
||||
# Revoke all sessions
|
||||
revoked_count = revoke_user_sessions(user.id)
|
||||
|
||||
db.session.commit()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'message': 'User suspended successfully',
|
||||
'revoked_sessions': revoked_count,
|
||||
'suspended_until': suspend_until.isoformat() if suspend_until else None
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Admin user suspension error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to suspend user'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/admin/users/<user_id>/unsuspend', methods=['POST'])
|
||||
@require_admin
|
||||
def admin_unsuspend_user(user_id):
|
||||
"""Unsuspend user account (admin only)"""
|
||||
try:
|
||||
user = User.query.get(user_id)
|
||||
if not user:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'User not found'
|
||||
}), 404
|
||||
|
||||
user.unsuspend()
|
||||
db.session.commit()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'message': 'User unsuspended successfully'
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Admin user unsuspension error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to unsuspend user'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/admin/users/<user_id>/reset-password', methods=['POST'])
|
||||
@require_admin
|
||||
def admin_reset_password(user_id):
|
||||
"""Reset user password (admin only)"""
|
||||
try:
|
||||
user = User.query.get(user_id)
|
||||
if not user:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'User not found'
|
||||
}), 404
|
||||
|
||||
data = request.get_json()
|
||||
new_password = data.get('password')
|
||||
|
||||
if not new_password or len(new_password) < 8:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Password must be at least 8 characters'
|
||||
}), 400
|
||||
|
||||
# Reset password
|
||||
user.set_password(new_password)
|
||||
user.failed_login_attempts = 0
|
||||
user.locked_until = None
|
||||
|
||||
# Revoke all sessions
|
||||
revoked_count = revoke_user_sessions(user.id)
|
||||
|
||||
db.session.commit()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'message': 'Password reset successfully',
|
||||
'revoked_sessions': revoked_count
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Admin password reset error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to reset password'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/admin/users/bulk', methods=['POST'])
|
||||
@require_admin
|
||||
def admin_bulk_operation():
|
||||
"""Perform bulk operations on users (admin only)"""
|
||||
try:
|
||||
data = request.get_json()
|
||||
user_ids = data.get('user_ids', [])
|
||||
operation = data.get('operation')
|
||||
|
||||
if not user_ids or not operation:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'User IDs and operation required'
|
||||
}), 400
|
||||
|
||||
# Get users
|
||||
users = User.query.filter(User.id.in_(user_ids)).all()
|
||||
|
||||
if not users:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'No users found'
|
||||
}), 404
|
||||
|
||||
results = {
|
||||
'success': 0,
|
||||
'failed': 0,
|
||||
'errors': []
|
||||
}
|
||||
|
||||
for user in users:
|
||||
try:
|
||||
if operation == 'suspend':
|
||||
user.suspend(data.get('reason', 'Bulk suspension'))
|
||||
revoke_user_sessions(user.id)
|
||||
elif operation == 'unsuspend':
|
||||
user.unsuspend()
|
||||
elif operation == 'activate':
|
||||
user.is_active = True
|
||||
elif operation == 'deactivate':
|
||||
user.is_active = False
|
||||
revoke_user_sessions(user.id)
|
||||
elif operation == 'verify':
|
||||
user.is_verified = True
|
||||
elif operation == 'unverify':
|
||||
user.is_verified = False
|
||||
elif operation == 'delete':
|
||||
if not user.is_admin:
|
||||
revoke_user_sessions(user.id)
|
||||
db.session.delete(user)
|
||||
else:
|
||||
results['errors'].append(f"Cannot delete admin user {user.username}")
|
||||
results['failed'] += 1
|
||||
continue
|
||||
else:
|
||||
results['errors'].append(f"Unknown operation for user {user.username}")
|
||||
results['failed'] += 1
|
||||
continue
|
||||
|
||||
results['success'] += 1
|
||||
except Exception as e:
|
||||
results['errors'].append(f"Failed for user {user.username}: {str(e)}")
|
||||
results['failed'] += 1
|
||||
|
||||
db.session.commit()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'results': results
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Admin bulk operation error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to perform bulk operation'
|
||||
}), 500
|
||||
|
||||
|
||||
@auth_bp.route('/admin/stats/users', methods=['GET'])
|
||||
@require_admin
|
||||
def admin_user_stats():
|
||||
"""Get user statistics (admin only)"""
|
||||
try:
|
||||
stats = {
|
||||
'total_users': User.query.count(),
|
||||
'active_users': User.query.filter(
|
||||
User.is_active == True,
|
||||
User.is_suspended == False
|
||||
).count(),
|
||||
'suspended_users': User.query.filter(User.is_suspended == True).count(),
|
||||
'verified_users': User.query.filter(User.is_verified == True).count(),
|
||||
'admin_users': User.query.filter(User.role == 'admin').count(),
|
||||
'users_by_role': dict(
|
||||
db.session.query(User.role, func.count(User.id))
|
||||
.group_by(User.role).all()
|
||||
),
|
||||
'recent_registrations': User.query.filter(
|
||||
User.created_at >= datetime.utcnow() - timedelta(days=7)
|
||||
).count(),
|
||||
'active_sessions': UserSession.query.filter(
|
||||
UserSession.expires_at > datetime.utcnow()
|
||||
).count()
|
||||
}
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'stats': stats
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
log_exception(e, "Admin stats error")
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Failed to fetch statistics'
|
||||
}), 500
|
168
check-pwa-status.html
Normal file
168
check-pwa-status.html
Normal file
@ -0,0 +1,168 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en">
|
||||
<head>
|
||||
<meta charset="UTF-8">
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||||
<title>PWA Installation Checker</title>
|
||||
<style>
|
||||
body {
|
||||
font-family: Arial, sans-serif;
|
||||
padding: 20px;
|
||||
max-width: 800px;
|
||||
margin: 0 auto;
|
||||
}
|
||||
.check {
|
||||
margin: 10px 0;
|
||||
padding: 10px;
|
||||
border-radius: 5px;
|
||||
}
|
||||
.pass {
|
||||
background-color: #d4edda;
|
||||
color: #155724;
|
||||
}
|
||||
.fail {
|
||||
background-color: #f8d7da;
|
||||
color: #721c24;
|
||||
}
|
||||
.info {
|
||||
background-color: #d1ecf1;
|
||||
color: #0c5460;
|
||||
}
|
||||
pre {
|
||||
background: #f4f4f4;
|
||||
padding: 10px;
|
||||
border-radius: 5px;
|
||||
overflow-x: auto;
|
||||
}
|
||||
</style>
|
||||
</head>
|
||||
<body>
|
||||
<h1>PWA Installation Status Checker</h1>
|
||||
<div id="results"></div>
|
||||
|
||||
<script>
|
||||
const results = document.getElementById('results');
|
||||
|
||||
function addResult(message, status = 'info') {
|
||||
const div = document.createElement('div');
|
||||
div.className = `check ${status}`;
|
||||
div.innerHTML = message;
|
||||
results.appendChild(div);
|
||||
}
|
||||
|
||||
// Check HTTPS
|
||||
if (location.protocol === 'https:' || location.hostname === 'localhost') {
|
||||
addResult('✅ HTTPS/Localhost: ' + location.protocol + '//' + location.hostname, 'pass');
|
||||
} else {
|
||||
addResult('❌ Not HTTPS: PWAs require HTTPS (or localhost)', 'fail');
|
||||
}
|
||||
|
||||
// Check Service Worker support
|
||||
if ('serviceWorker' in navigator) {
|
||||
addResult('✅ Service Worker API supported', 'pass');
|
||||
|
||||
// Check registration
|
||||
navigator.serviceWorker.getRegistration().then(reg => {
|
||||
if (reg) {
|
||||
addResult('✅ Service Worker registered: ' + reg.scope, 'pass');
|
||||
addResult('Service Worker state: ' + (reg.active ? 'active' : 'not active'), 'info');
|
||||
} else {
|
||||
addResult('❌ No Service Worker registered', 'fail');
|
||||
}
|
||||
});
|
||||
} else {
|
||||
addResult('❌ Service Worker API not supported', 'fail');
|
||||
}
|
||||
|
||||
// Check manifest
|
||||
const manifestLink = document.querySelector('link[rel="manifest"]');
|
||||
if (manifestLink) {
|
||||
addResult('✅ Manifest link found: ' + manifestLink.href, 'pass');
|
||||
|
||||
// Fetch and validate manifest
|
||||
fetch(manifestLink.href)
|
||||
.then(response => response.json())
|
||||
.then(manifest => {
|
||||
addResult('Manifest loaded successfully', 'info');
|
||||
|
||||
// Check required fields
|
||||
const required = ['name', 'short_name', 'start_url', 'display', 'icons'];
|
||||
required.forEach(field => {
|
||||
if (manifest[field]) {
|
||||
addResult(`✅ Manifest has ${field}: ${JSON.stringify(manifest[field])}`, 'pass');
|
||||
} else {
|
||||
addResult(`❌ Manifest missing ${field}`, 'fail');
|
||||
}
|
||||
});
|
||||
|
||||
// Check icons
|
||||
if (manifest.icons && manifest.icons.length > 0) {
|
||||
const has192 = manifest.icons.some(icon => icon.sizes && icon.sizes.includes('192'));
|
||||
const has512 = manifest.icons.some(icon => icon.sizes && icon.sizes.includes('512'));
|
||||
|
||||
if (has192) addResult('✅ Has 192x192 icon', 'pass');
|
||||
else addResult('❌ Missing 192x192 icon', 'fail');
|
||||
|
||||
if (has512) addResult('✅ Has 512x512 icon', 'pass');
|
||||
else addResult('⚠️ Missing 512x512 icon (recommended)', 'info');
|
||||
|
||||
// Check icon purposes
|
||||
manifest.icons.forEach((icon, i) => {
|
||||
addResult(`Icon ${i + 1}: ${icon.sizes} - purpose: ${icon.purpose || 'not specified'}`, 'info');
|
||||
});
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
addResult('❌ Failed to load manifest: ' + error.message, 'fail');
|
||||
});
|
||||
} else {
|
||||
addResult('❌ No manifest link found in HTML', 'fail');
|
||||
}
|
||||
|
||||
// Check installability
|
||||
window.addEventListener('beforeinstallprompt', (e) => {
|
||||
e.preventDefault();
|
||||
addResult('✅ Browser considers app installable (beforeinstallprompt fired)', 'pass');
|
||||
|
||||
// Show install criteria met
|
||||
const criteria = [
|
||||
'HTTPS or localhost',
|
||||
'Valid manifest with required fields',
|
||||
'Service Worker with fetch handler',
|
||||
'Icons (192x192 minimum)',
|
||||
'Not already installed'
|
||||
];
|
||||
|
||||
addResult('<strong>Installation criteria met:</strong><br>' + criteria.join('<br>'), 'info');
|
||||
});
|
||||
|
||||
// Check if already installed
|
||||
if (window.matchMedia('(display-mode: standalone)').matches) {
|
||||
addResult('✅ App is already installed (running in standalone mode)', 'pass');
|
||||
}
|
||||
|
||||
// Additional Chrome-specific checks
|
||||
if (navigator.userAgent.includes('Chrome')) {
|
||||
addResult('Chrome browser detected - checking Chrome-specific requirements', 'info');
|
||||
|
||||
setTimeout(() => {
|
||||
// If no beforeinstallprompt event fired after 3 seconds
|
||||
if (!window.installPromptFired) {
|
||||
addResult('⚠️ beforeinstallprompt event not fired after 3 seconds', 'info');
|
||||
addResult('Possible reasons:<br>' +
|
||||
'- App already installed<br>' +
|
||||
'- User dismissed install prompt recently<br>' +
|
||||
'- Missing PWA criteria<br>' +
|
||||
'- Chrome needs a user gesture to show prompt', 'info');
|
||||
}
|
||||
}, 3000);
|
||||
}
|
||||
|
||||
// Log all checks completed
|
||||
setTimeout(() => {
|
||||
addResult('<br><strong>All checks completed</strong>', 'info');
|
||||
console.log('PWA Status Check Complete');
|
||||
}, 4000);
|
||||
</script>
|
||||
</body>
|
||||
</html>
|
48
check_services.py
Normal file
48
check_services.py
Normal file
@ -0,0 +1,48 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
Check if Redis and PostgreSQL are available
|
||||
"""
|
||||
import os
|
||||
import sys
|
||||
from dotenv import load_dotenv
|
||||
|
||||
# Load environment variables
|
||||
load_dotenv()
|
||||
|
||||
def check_redis():
|
||||
"""Check if Redis is available"""
|
||||
try:
|
||||
import redis
|
||||
r = redis.Redis.from_url(os.environ.get('REDIS_URL', 'redis://localhost:6379/0'))
|
||||
r.ping()
|
||||
return True
|
||||
except:
|
||||
return False
|
||||
|
||||
def check_postgresql():
|
||||
"""Check if PostgreSQL is available"""
|
||||
try:
|
||||
from sqlalchemy import create_engine, text
|
||||
db_url = os.environ.get('DATABASE_URL', 'postgresql://localhost/talk2me')
|
||||
engine = create_engine(db_url, pool_pre_ping=True)
|
||||
with engine.connect() as conn:
|
||||
result = conn.execute(text("SELECT 1"))
|
||||
result.fetchone()
|
||||
return True
|
||||
except Exception as e:
|
||||
print(f"PostgreSQL connection error: {e}")
|
||||
return False
|
||||
|
||||
if __name__ == '__main__':
|
||||
redis_ok = check_redis()
|
||||
postgres_ok = check_postgresql()
|
||||
|
||||
print(f"Redis: {'✓ Available' if redis_ok else '✗ Not available'}")
|
||||
print(f"PostgreSQL: {'✓ Available' if postgres_ok else '✗ Not available'}")
|
||||
|
||||
if redis_ok and postgres_ok:
|
||||
print("\nAll services available - use full admin module")
|
||||
sys.exit(0)
|
||||
else:
|
||||
print("\nSome services missing - use simple admin module")
|
||||
sys.exit(1)
|
213
config.py
Normal file
213
config.py
Normal file
@ -0,0 +1,213 @@
|
||||
# Configuration management with secrets integration
|
||||
import os
|
||||
import logging
|
||||
from datetime import timedelta
|
||||
from secrets_manager import get_secret, get_secrets_manager
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
class Config:
|
||||
"""Base configuration with secrets management"""
|
||||
|
||||
def __init__(self):
|
||||
self.secrets_manager = get_secrets_manager()
|
||||
self._load_config()
|
||||
|
||||
def _load_config(self):
|
||||
"""Load configuration from environment and secrets"""
|
||||
# Flask configuration
|
||||
self.SECRET_KEY = self._get_secret('FLASK_SECRET_KEY',
|
||||
os.environ.get('SECRET_KEY', 'dev-key-change-this'))
|
||||
|
||||
# Security
|
||||
self.SESSION_COOKIE_SECURE = self._get_bool('SESSION_COOKIE_SECURE', True)
|
||||
self.SESSION_COOKIE_HTTPONLY = True
|
||||
self.SESSION_COOKIE_SAMESITE = 'Lax'
|
||||
self.PERMANENT_SESSION_LIFETIME = timedelta(hours=24)
|
||||
|
||||
# TTS Configuration
|
||||
self.TTS_SERVER_URL = os.environ.get('TTS_SERVER_URL', 'http://localhost:5050/v1/audio/speech')
|
||||
self.TTS_API_KEY = self._get_secret('TTS_API_KEY', os.environ.get('TTS_API_KEY', ''))
|
||||
|
||||
# Upload configuration
|
||||
self.UPLOAD_FOLDER = os.environ.get('UPLOAD_FOLDER', None)
|
||||
|
||||
# Request size limits (in bytes)
|
||||
self.MAX_CONTENT_LENGTH = int(os.environ.get('MAX_CONTENT_LENGTH', 50 * 1024 * 1024)) # 50MB
|
||||
self.MAX_AUDIO_SIZE = int(os.environ.get('MAX_AUDIO_SIZE', 25 * 1024 * 1024)) # 25MB
|
||||
self.MAX_JSON_SIZE = int(os.environ.get('MAX_JSON_SIZE', 1 * 1024 * 1024)) # 1MB
|
||||
self.MAX_IMAGE_SIZE = int(os.environ.get('MAX_IMAGE_SIZE', 10 * 1024 * 1024)) # 10MB
|
||||
|
||||
# CORS configuration
|
||||
self.CORS_ORIGINS = os.environ.get('CORS_ORIGINS', '*').split(',')
|
||||
self.ADMIN_CORS_ORIGINS = os.environ.get('ADMIN_CORS_ORIGINS', 'http://localhost:*').split(',')
|
||||
|
||||
# Admin configuration
|
||||
self.ADMIN_TOKEN = self._get_secret('ADMIN_TOKEN',
|
||||
os.environ.get('ADMIN_TOKEN', 'default-admin-token'))
|
||||
|
||||
# Database configuration
|
||||
self.DATABASE_URL = self._get_secret('DATABASE_URL',
|
||||
os.environ.get('DATABASE_URL', 'postgresql://localhost/talk2me'))
|
||||
self.SQLALCHEMY_DATABASE_URI = self.DATABASE_URL
|
||||
self.SQLALCHEMY_TRACK_MODIFICATIONS = False
|
||||
self.SQLALCHEMY_ENGINE_OPTIONS = {
|
||||
'pool_size': 10,
|
||||
'pool_recycle': 3600,
|
||||
'pool_pre_ping': True
|
||||
}
|
||||
|
||||
# Redis configuration
|
||||
self.REDIS_URL = self._get_secret('REDIS_URL',
|
||||
os.environ.get('REDIS_URL', 'redis://localhost:6379/0'))
|
||||
self.REDIS_DECODE_RESPONSES = False
|
||||
self.REDIS_MAX_CONNECTIONS = int(os.environ.get('REDIS_MAX_CONNECTIONS', 50))
|
||||
self.REDIS_SOCKET_TIMEOUT = int(os.environ.get('REDIS_SOCKET_TIMEOUT', 5))
|
||||
|
||||
# Whisper configuration
|
||||
self.WHISPER_MODEL_SIZE = os.environ.get('WHISPER_MODEL_SIZE', 'base')
|
||||
self.WHISPER_DEVICE = os.environ.get('WHISPER_DEVICE', 'auto')
|
||||
|
||||
# Ollama configuration
|
||||
self.OLLAMA_HOST = os.environ.get('OLLAMA_HOST', 'http://localhost:11434')
|
||||
self.OLLAMA_MODEL = os.environ.get('OLLAMA_MODEL', 'gemma3:27b')
|
||||
|
||||
# Rate limiting configuration
|
||||
self.RATE_LIMIT_ENABLED = self._get_bool('RATE_LIMIT_ENABLED', True)
|
||||
self.RATE_LIMIT_STORAGE_URL = self._get_secret('RATE_LIMIT_STORAGE_URL',
|
||||
os.environ.get('RATE_LIMIT_STORAGE_URL', 'memory://'))
|
||||
|
||||
# Logging configuration
|
||||
self.LOG_LEVEL = os.environ.get('LOG_LEVEL', 'INFO')
|
||||
self.LOG_FILE = os.environ.get('LOG_FILE', 'talk2me.log')
|
||||
|
||||
# Feature flags
|
||||
self.ENABLE_PUSH_NOTIFICATIONS = self._get_bool('ENABLE_PUSH_NOTIFICATIONS', True)
|
||||
self.ENABLE_OFFLINE_MODE = self._get_bool('ENABLE_OFFLINE_MODE', True)
|
||||
self.ENABLE_STREAMING = self._get_bool('ENABLE_STREAMING', True)
|
||||
self.ENABLE_MULTI_SPEAKER = self._get_bool('ENABLE_MULTI_SPEAKER', True)
|
||||
|
||||
# Performance tuning
|
||||
self.WORKER_CONNECTIONS = int(os.environ.get('WORKER_CONNECTIONS', '1000'))
|
||||
self.WORKER_TIMEOUT = int(os.environ.get('WORKER_TIMEOUT', '120'))
|
||||
|
||||
# Validate configuration
|
||||
self._validate_config()
|
||||
|
||||
def _get_secret(self, key: str, default: str = None) -> str:
|
||||
"""Get secret from secrets manager or environment"""
|
||||
value = self.secrets_manager.get(key)
|
||||
if value is None:
|
||||
value = default
|
||||
|
||||
if value is None:
|
||||
logger.warning(f"Configuration {key} not set")
|
||||
|
||||
return value
|
||||
|
||||
def _get_bool(self, key: str, default: bool = False) -> bool:
|
||||
"""Get boolean configuration value"""
|
||||
value = os.environ.get(key, '').lower()
|
||||
if value in ('true', '1', 'yes', 'on'):
|
||||
return True
|
||||
elif value in ('false', '0', 'no', 'off'):
|
||||
return False
|
||||
return default
|
||||
|
||||
def _validate_config(self):
|
||||
"""Validate configuration values"""
|
||||
# Check required secrets
|
||||
if not self.SECRET_KEY or self.SECRET_KEY == 'dev-key-change-this':
|
||||
logger.warning("Using default SECRET_KEY - this is insecure for production!")
|
||||
|
||||
if not self.TTS_API_KEY:
|
||||
logger.warning("TTS_API_KEY not configured - TTS functionality may not work")
|
||||
|
||||
if self.ADMIN_TOKEN == 'default-admin-token':
|
||||
logger.warning("Using default ADMIN_TOKEN - this is insecure for production!")
|
||||
|
||||
# Validate URLs
|
||||
if not self._is_valid_url(self.TTS_SERVER_URL):
|
||||
logger.error(f"Invalid TTS_SERVER_URL: {self.TTS_SERVER_URL}")
|
||||
|
||||
# Check file permissions
|
||||
if self.UPLOAD_FOLDER and not os.access(self.UPLOAD_FOLDER, os.W_OK):
|
||||
logger.warning(f"Upload folder {self.UPLOAD_FOLDER} is not writable")
|
||||
|
||||
def _is_valid_url(self, url: str) -> bool:
|
||||
"""Check if URL is valid"""
|
||||
return url.startswith(('http://', 'https://'))
|
||||
|
||||
def to_dict(self) -> dict:
|
||||
"""Export configuration as dictionary (excluding secrets)"""
|
||||
config = {}
|
||||
for key in dir(self):
|
||||
if key.isupper() and not key.startswith('_'):
|
||||
value = getattr(self, key)
|
||||
# Mask sensitive values
|
||||
if any(sensitive in key for sensitive in ['KEY', 'TOKEN', 'PASSWORD', 'SECRET']):
|
||||
config[key] = '***MASKED***'
|
||||
else:
|
||||
config[key] = value
|
||||
return config
|
||||
|
||||
class DevelopmentConfig(Config):
|
||||
"""Development configuration"""
|
||||
def _load_config(self):
|
||||
super()._load_config()
|
||||
self.DEBUG = True
|
||||
self.TESTING = False
|
||||
self.SESSION_COOKIE_SECURE = False # Allow HTTP in development
|
||||
|
||||
class ProductionConfig(Config):
|
||||
"""Production configuration"""
|
||||
def _load_config(self):
|
||||
super()._load_config()
|
||||
self.DEBUG = False
|
||||
self.TESTING = False
|
||||
|
||||
# Enforce security in production
|
||||
if not self.SECRET_KEY or self.SECRET_KEY == 'dev-key-change-this':
|
||||
raise ValueError("SECRET_KEY must be set in production")
|
||||
|
||||
if self.ADMIN_TOKEN == 'default-admin-token':
|
||||
raise ValueError("ADMIN_TOKEN must be changed in production")
|
||||
|
||||
class TestingConfig(Config):
|
||||
"""Testing configuration"""
|
||||
def _load_config(self):
|
||||
super()._load_config()
|
||||
self.DEBUG = True
|
||||
self.TESTING = True
|
||||
self.WTF_CSRF_ENABLED = False
|
||||
self.RATE_LIMIT_ENABLED = False
|
||||
|
||||
# Configuration factory
|
||||
def get_config(env: str = None) -> Config:
|
||||
"""Get configuration based on environment"""
|
||||
if env is None:
|
||||
env = os.environ.get('FLASK_ENV', 'development')
|
||||
|
||||
configs = {
|
||||
'development': DevelopmentConfig,
|
||||
'production': ProductionConfig,
|
||||
'testing': TestingConfig
|
||||
}
|
||||
|
||||
config_class = configs.get(env, DevelopmentConfig)
|
||||
return config_class()
|
||||
|
||||
# Convenience function for Flask app
|
||||
def init_app(app):
|
||||
"""Initialize Flask app with configuration"""
|
||||
config = get_config()
|
||||
|
||||
# Apply configuration to app
|
||||
for key in dir(config):
|
||||
if key.isupper():
|
||||
app.config[key] = getattr(config, key)
|
||||
|
||||
# Store config object
|
||||
app.app_config = config
|
||||
|
||||
logger.info(f"Configuration loaded for environment: {os.environ.get('FLASK_ENV', 'development')}")
|
273
database.py
Normal file
273
database.py
Normal file
@ -0,0 +1,273 @@
|
||||
# Database models and configuration for Talk2Me application
|
||||
import os
|
||||
from datetime import datetime
|
||||
from typing import Optional, Dict, Any
|
||||
from flask_sqlalchemy import SQLAlchemy
|
||||
from sqlalchemy import Index, text
|
||||
from sqlalchemy.dialects.postgresql import UUID, JSONB
|
||||
from sqlalchemy.ext.hybrid import hybrid_property
|
||||
import uuid
|
||||
|
||||
db = SQLAlchemy()
|
||||
|
||||
class Translation(db.Model):
|
||||
"""Store translation history for analytics and caching"""
|
||||
__tablename__ = 'translations'
|
||||
|
||||
id = db.Column(UUID(as_uuid=True), primary_key=True, default=uuid.uuid4)
|
||||
session_id = db.Column(db.String(255), nullable=False, index=True)
|
||||
user_id = db.Column(db.String(255), nullable=True, index=True)
|
||||
|
||||
# Translation data
|
||||
source_text = db.Column(db.Text, nullable=False)
|
||||
source_language = db.Column(db.String(10), nullable=False)
|
||||
target_text = db.Column(db.Text, nullable=False)
|
||||
target_language = db.Column(db.String(10), nullable=False)
|
||||
|
||||
# Metadata
|
||||
translation_time_ms = db.Column(db.Integer, nullable=True)
|
||||
model_used = db.Column(db.String(50), default='gemma3:27b')
|
||||
confidence_score = db.Column(db.Float, nullable=True)
|
||||
|
||||
# Timestamps
|
||||
created_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow)
|
||||
accessed_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow)
|
||||
access_count = db.Column(db.Integer, default=1)
|
||||
|
||||
# Client info
|
||||
ip_address = db.Column(db.String(45), nullable=True)
|
||||
user_agent = db.Column(db.String(500), nullable=True)
|
||||
|
||||
# Create indexes for better query performance
|
||||
__table_args__ = (
|
||||
Index('idx_translations_languages', 'source_language', 'target_language'),
|
||||
Index('idx_translations_created_at', 'created_at'),
|
||||
Index('idx_translations_session_user', 'session_id', 'user_id'),
|
||||
)
|
||||
|
||||
def to_dict(self) -> Dict[str, Any]:
|
||||
"""Convert translation to dictionary"""
|
||||
return {
|
||||
'id': str(self.id),
|
||||
'session_id': self.session_id,
|
||||
'user_id': self.user_id,
|
||||
'source_text': self.source_text,
|
||||
'source_language': self.source_language,
|
||||
'target_text': self.target_text,
|
||||
'target_language': self.target_language,
|
||||
'translation_time_ms': self.translation_time_ms,
|
||||
'model_used': self.model_used,
|
||||
'confidence_score': self.confidence_score,
|
||||
'created_at': self.created_at.isoformat(),
|
||||
'accessed_at': self.accessed_at.isoformat(),
|
||||
'access_count': self.access_count
|
||||
}
|
||||
|
||||
|
||||
class Transcription(db.Model):
|
||||
"""Store transcription history"""
|
||||
__tablename__ = 'transcriptions'
|
||||
|
||||
id = db.Column(UUID(as_uuid=True), primary_key=True, default=uuid.uuid4)
|
||||
session_id = db.Column(db.String(255), nullable=False, index=True)
|
||||
user_id = db.Column(db.String(255), nullable=True, index=True)
|
||||
|
||||
# Transcription data
|
||||
transcribed_text = db.Column(db.Text, nullable=False)
|
||||
detected_language = db.Column(db.String(10), nullable=True)
|
||||
audio_duration_seconds = db.Column(db.Float, nullable=True)
|
||||
|
||||
# Metadata
|
||||
transcription_time_ms = db.Column(db.Integer, nullable=True)
|
||||
model_used = db.Column(db.String(50), default='whisper-base')
|
||||
confidence_score = db.Column(db.Float, nullable=True)
|
||||
|
||||
# File info
|
||||
audio_file_size = db.Column(db.Integer, nullable=True)
|
||||
audio_format = db.Column(db.String(10), nullable=True)
|
||||
|
||||
# Timestamps
|
||||
created_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow)
|
||||
|
||||
# Client info
|
||||
ip_address = db.Column(db.String(45), nullable=True)
|
||||
user_agent = db.Column(db.String(500), nullable=True)
|
||||
|
||||
__table_args__ = (
|
||||
Index('idx_transcriptions_created_at', 'created_at'),
|
||||
Index('idx_transcriptions_session_user', 'session_id', 'user_id'),
|
||||
)
|
||||
|
||||
def to_dict(self) -> Dict[str, Any]:
|
||||
"""Convert transcription to dictionary"""
|
||||
return {
|
||||
'id': str(self.id),
|
||||
'session_id': self.session_id,
|
||||
'user_id': self.user_id,
|
||||
'transcribed_text': self.transcribed_text,
|
||||
'detected_language': self.detected_language,
|
||||
'audio_duration_seconds': self.audio_duration_seconds,
|
||||
'transcription_time_ms': self.transcription_time_ms,
|
||||
'model_used': self.model_used,
|
||||
'confidence_score': self.confidence_score,
|
||||
'audio_file_size': self.audio_file_size,
|
||||
'audio_format': self.audio_format,
|
||||
'created_at': self.created_at.isoformat()
|
||||
}
|
||||
|
||||
|
||||
class UserPreferences(db.Model):
|
||||
"""Store user preferences and settings"""
|
||||
__tablename__ = 'user_preferences'
|
||||
|
||||
id = db.Column(UUID(as_uuid=True), primary_key=True, default=uuid.uuid4)
|
||||
user_id = db.Column(db.String(255), nullable=False, unique=True, index=True)
|
||||
session_id = db.Column(db.String(255), nullable=True)
|
||||
|
||||
# Preferences
|
||||
preferred_source_language = db.Column(db.String(10), nullable=True)
|
||||
preferred_target_language = db.Column(db.String(10), nullable=True)
|
||||
preferred_voice = db.Column(db.String(50), nullable=True)
|
||||
speech_speed = db.Column(db.Float, default=1.0)
|
||||
|
||||
# Settings stored as JSONB for flexibility
|
||||
settings = db.Column(JSONB, default={})
|
||||
|
||||
# Usage stats
|
||||
total_translations = db.Column(db.Integer, default=0)
|
||||
total_transcriptions = db.Column(db.Integer, default=0)
|
||||
total_tts_requests = db.Column(db.Integer, default=0)
|
||||
|
||||
# Timestamps
|
||||
created_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow)
|
||||
updated_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow, onupdate=datetime.utcnow)
|
||||
last_active_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow)
|
||||
|
||||
def to_dict(self) -> Dict[str, Any]:
|
||||
"""Convert preferences to dictionary"""
|
||||
return {
|
||||
'id': str(self.id),
|
||||
'user_id': self.user_id,
|
||||
'preferred_source_language': self.preferred_source_language,
|
||||
'preferred_target_language': self.preferred_target_language,
|
||||
'preferred_voice': self.preferred_voice,
|
||||
'speech_speed': self.speech_speed,
|
||||
'settings': self.settings or {},
|
||||
'total_translations': self.total_translations,
|
||||
'total_transcriptions': self.total_transcriptions,
|
||||
'total_tts_requests': self.total_tts_requests,
|
||||
'created_at': self.created_at.isoformat(),
|
||||
'updated_at': self.updated_at.isoformat(),
|
||||
'last_active_at': self.last_active_at.isoformat()
|
||||
}
|
||||
|
||||
|
||||
class UsageAnalytics(db.Model):
|
||||
"""Store aggregated usage analytics"""
|
||||
__tablename__ = 'usage_analytics'
|
||||
|
||||
id = db.Column(UUID(as_uuid=True), primary_key=True, default=uuid.uuid4)
|
||||
|
||||
# Time period
|
||||
date = db.Column(db.Date, nullable=False, index=True)
|
||||
hour = db.Column(db.Integer, nullable=True) # 0-23, null for daily aggregates
|
||||
|
||||
# Metrics
|
||||
total_requests = db.Column(db.Integer, default=0)
|
||||
unique_sessions = db.Column(db.Integer, default=0)
|
||||
unique_users = db.Column(db.Integer, default=0)
|
||||
|
||||
# Service breakdown
|
||||
transcriptions = db.Column(db.Integer, default=0)
|
||||
translations = db.Column(db.Integer, default=0)
|
||||
tts_requests = db.Column(db.Integer, default=0)
|
||||
|
||||
# Performance metrics
|
||||
avg_transcription_time_ms = db.Column(db.Float, nullable=True)
|
||||
avg_translation_time_ms = db.Column(db.Float, nullable=True)
|
||||
avg_tts_time_ms = db.Column(db.Float, nullable=True)
|
||||
|
||||
# Language stats (stored as JSONB)
|
||||
language_pairs = db.Column(JSONB, default={}) # {"en-es": 100, "fr-en": 50}
|
||||
detected_languages = db.Column(JSONB, default={}) # {"en": 150, "es": 100}
|
||||
|
||||
# Error stats
|
||||
error_count = db.Column(db.Integer, default=0)
|
||||
error_details = db.Column(JSONB, default={})
|
||||
|
||||
__table_args__ = (
|
||||
Index('idx_analytics_date_hour', 'date', 'hour'),
|
||||
db.UniqueConstraint('date', 'hour', name='uq_analytics_date_hour'),
|
||||
)
|
||||
|
||||
|
||||
class ApiKey(db.Model):
|
||||
"""Store API keys for authenticated access"""
|
||||
__tablename__ = 'api_keys'
|
||||
|
||||
id = db.Column(UUID(as_uuid=True), primary_key=True, default=uuid.uuid4)
|
||||
key_hash = db.Column(db.String(255), nullable=False, unique=True, index=True)
|
||||
name = db.Column(db.String(100), nullable=False)
|
||||
user_id = db.Column(db.String(255), nullable=True)
|
||||
|
||||
# Permissions and limits
|
||||
is_active = db.Column(db.Boolean, default=True)
|
||||
rate_limit_per_minute = db.Column(db.Integer, default=60)
|
||||
rate_limit_per_hour = db.Column(db.Integer, default=1000)
|
||||
allowed_endpoints = db.Column(JSONB, default=[]) # Empty = all endpoints
|
||||
|
||||
# Usage tracking
|
||||
total_requests = db.Column(db.Integer, default=0)
|
||||
last_used_at = db.Column(db.DateTime, nullable=True)
|
||||
|
||||
# Timestamps
|
||||
created_at = db.Column(db.DateTime, nullable=False, default=datetime.utcnow)
|
||||
expires_at = db.Column(db.DateTime, nullable=True)
|
||||
|
||||
@hybrid_property
|
||||
def is_expired(self):
|
||||
"""Check if API key is expired"""
|
||||
if self.expires_at is None:
|
||||
return False
|
||||
return datetime.utcnow() > self.expires_at
|
||||
|
||||
|
||||
def init_db(app):
|
||||
"""Initialize database with app"""
|
||||
db.init_app(app)
|
||||
|
||||
with app.app_context():
|
||||
# Create tables if they don't exist
|
||||
db.create_all()
|
||||
|
||||
# Create any custom indexes or functions
|
||||
try:
|
||||
# Create a function for updating updated_at timestamp
|
||||
db.session.execute(text("""
|
||||
CREATE OR REPLACE FUNCTION update_updated_at_column()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
NEW.updated_at = NOW();
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ language 'plpgsql';
|
||||
"""))
|
||||
|
||||
# Drop existing trigger if it exists and recreate it
|
||||
db.session.execute(text("""
|
||||
DROP TRIGGER IF EXISTS update_user_preferences_updated_at ON user_preferences;
|
||||
"""))
|
||||
|
||||
# Create trigger for user_preferences
|
||||
db.session.execute(text("""
|
||||
CREATE TRIGGER update_user_preferences_updated_at
|
||||
BEFORE UPDATE ON user_preferences
|
||||
FOR EACH ROW
|
||||
EXECUTE FUNCTION update_updated_at_column();
|
||||
"""))
|
||||
|
||||
db.session.commit()
|
||||
except Exception as e:
|
||||
# Log error but don't fail - database might not support triggers
|
||||
db.session.rollback()
|
||||
app.logger.debug(f"Database initialization note: {e}")
|
135
database_init.py
Normal file
135
database_init.py
Normal file
@ -0,0 +1,135 @@
|
||||
#!/usr/bin/env python3
|
||||
# Database initialization script
|
||||
import os
|
||||
import sys
|
||||
import logging
|
||||
from sqlalchemy import create_engine, text
|
||||
from config import get_config
|
||||
|
||||
logging.basicConfig(level=logging.INFO)
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
def create_database():
|
||||
"""Create the database if it doesn't exist"""
|
||||
config = get_config()
|
||||
db_url = config.DATABASE_URL
|
||||
|
||||
if db_url.startswith('postgresql'):
|
||||
# Parse database name from URL
|
||||
parts = db_url.split('/')
|
||||
db_name = parts[-1].split('?')[0]
|
||||
base_url = '/'.join(parts[:-1])
|
||||
|
||||
# Connect to postgres database to create our database
|
||||
engine = create_engine(f"{base_url}/postgres")
|
||||
|
||||
try:
|
||||
with engine.connect() as conn:
|
||||
# Check if database exists
|
||||
result = conn.execute(
|
||||
text("SELECT 1 FROM pg_database WHERE datname = :dbname"),
|
||||
{"dbname": db_name}
|
||||
)
|
||||
exists = result.fetchone() is not None
|
||||
|
||||
if not exists:
|
||||
# Create database
|
||||
conn.execute(text(f"CREATE DATABASE {db_name}"))
|
||||
logger.info(f"Database '{db_name}' created successfully")
|
||||
else:
|
||||
logger.info(f"Database '{db_name}' already exists")
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error creating database: {e}")
|
||||
return False
|
||||
finally:
|
||||
engine.dispose()
|
||||
|
||||
return True
|
||||
|
||||
def check_redis():
|
||||
"""Check Redis connectivity"""
|
||||
config = get_config()
|
||||
|
||||
try:
|
||||
import redis
|
||||
r = redis.from_url(config.REDIS_URL)
|
||||
r.ping()
|
||||
logger.info("Redis connection successful")
|
||||
return True
|
||||
except Exception as e:
|
||||
logger.error(f"Redis connection failed: {e}")
|
||||
return False
|
||||
|
||||
def init_database_extensions():
|
||||
"""Initialize PostgreSQL extensions"""
|
||||
config = get_config()
|
||||
|
||||
if not config.DATABASE_URL.startswith('postgresql'):
|
||||
return True
|
||||
|
||||
engine = create_engine(config.DATABASE_URL)
|
||||
|
||||
try:
|
||||
with engine.connect() as conn:
|
||||
# Enable UUID extension
|
||||
conn.execute(text("CREATE EXTENSION IF NOT EXISTS \"uuid-ossp\""))
|
||||
logger.info("PostgreSQL extensions initialized")
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error initializing extensions: {e}")
|
||||
return False
|
||||
finally:
|
||||
engine.dispose()
|
||||
|
||||
return True
|
||||
|
||||
def main():
|
||||
"""Main initialization function"""
|
||||
logger.info("Starting database initialization...")
|
||||
|
||||
# Create database
|
||||
if not create_database():
|
||||
logger.error("Failed to create database")
|
||||
sys.exit(1)
|
||||
|
||||
# Initialize extensions
|
||||
if not init_database_extensions():
|
||||
logger.error("Failed to initialize database extensions")
|
||||
sys.exit(1)
|
||||
|
||||
# Check Redis
|
||||
if not check_redis():
|
||||
logger.warning("Redis not available - caching will be disabled")
|
||||
|
||||
logger.info("Database initialization completed successfully")
|
||||
|
||||
# Create all tables using SQLAlchemy models
|
||||
logger.info("Creating database tables...")
|
||||
try:
|
||||
from flask import Flask
|
||||
from database import db, init_db
|
||||
from config import get_config
|
||||
|
||||
# Import all models to ensure they're registered
|
||||
from auth_models import User, LoginHistory, UserSession, RevokedToken
|
||||
|
||||
# Create Flask app context
|
||||
app = Flask(__name__)
|
||||
config = get_config()
|
||||
app.config.from_mapping(config.__dict__)
|
||||
|
||||
# Initialize database
|
||||
init_db(app)
|
||||
|
||||
with app.app_context():
|
||||
# Create all tables
|
||||
db.create_all()
|
||||
logger.info("Database tables created successfully")
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to create database tables: {e}")
|
||||
sys.exit(1)
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
208
deploy.sh
Executable file
208
deploy.sh
Executable file
@ -0,0 +1,208 @@
|
||||
#!/bin/bash
|
||||
# Production deployment script for Talk2Me
|
||||
|
||||
set -e # Exit on error
|
||||
|
||||
# Colors for output
|
||||
RED='\033[0;31m'
|
||||
GREEN='\033[0;32m'
|
||||
YELLOW='\033[1;33m'
|
||||
NC='\033[0m' # No Color
|
||||
|
||||
# Configuration
|
||||
APP_NAME="talk2me"
|
||||
APP_USER="talk2me"
|
||||
APP_DIR="/opt/talk2me"
|
||||
VENV_DIR="$APP_DIR/venv"
|
||||
LOG_DIR="/var/log/talk2me"
|
||||
PID_FILE="/var/run/talk2me.pid"
|
||||
WORKERS=${WORKERS:-4}
|
||||
|
||||
# Functions
|
||||
print_status() {
|
||||
echo -e "${GREEN}[INFO]${NC} $1"
|
||||
}
|
||||
|
||||
print_error() {
|
||||
echo -e "${RED}[ERROR]${NC} $1"
|
||||
}
|
||||
|
||||
print_warning() {
|
||||
echo -e "${YELLOW}[WARNING]${NC} $1"
|
||||
}
|
||||
|
||||
# Check if running as root
|
||||
if [[ $EUID -ne 0 ]]; then
|
||||
print_error "This script must be run as root"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Create application user if doesn't exist
|
||||
if ! id "$APP_USER" &>/dev/null; then
|
||||
print_status "Creating application user: $APP_USER"
|
||||
useradd -m -s /bin/bash $APP_USER
|
||||
fi
|
||||
|
||||
# Create directories
|
||||
print_status "Creating application directories"
|
||||
mkdir -p $APP_DIR $LOG_DIR
|
||||
chown -R $APP_USER:$APP_USER $APP_DIR $LOG_DIR
|
||||
|
||||
# Copy application files
|
||||
print_status "Copying application files"
|
||||
rsync -av --exclude='venv' --exclude='__pycache__' --exclude='*.pyc' \
|
||||
--exclude='logs' --exclude='.git' --exclude='node_modules' \
|
||||
./ $APP_DIR/
|
||||
|
||||
# Create virtual environment
|
||||
print_status "Setting up Python virtual environment"
|
||||
su - $APP_USER -c "cd $APP_DIR && python3 -m venv $VENV_DIR"
|
||||
|
||||
# Install dependencies
|
||||
print_status "Installing Python dependencies"
|
||||
su - $APP_USER -c "cd $APP_DIR && $VENV_DIR/bin/pip install --upgrade pip"
|
||||
su - $APP_USER -c "cd $APP_DIR && $VENV_DIR/bin/pip install -r requirements-prod.txt"
|
||||
|
||||
# Install Whisper model
|
||||
print_status "Downloading Whisper model (this may take a while)"
|
||||
su - $APP_USER -c "cd $APP_DIR && $VENV_DIR/bin/python -c 'import whisper; whisper.load_model(\"base\")'"
|
||||
|
||||
# Build frontend assets
|
||||
if [ -f "package.json" ]; then
|
||||
print_status "Building frontend assets"
|
||||
cd $APP_DIR
|
||||
npm install
|
||||
npm run build
|
||||
fi
|
||||
|
||||
# Create systemd service
|
||||
print_status "Creating systemd service"
|
||||
cat > /etc/systemd/system/talk2me.service <<EOF
|
||||
[Unit]
|
||||
Description=Talk2Me Translation Service
|
||||
After=network.target
|
||||
|
||||
[Service]
|
||||
Type=notify
|
||||
User=$APP_USER
|
||||
Group=$APP_USER
|
||||
WorkingDirectory=$APP_DIR
|
||||
Environment="PATH=$VENV_DIR/bin"
|
||||
Environment="FLASK_ENV=production"
|
||||
Environment="UPLOAD_FOLDER=/tmp/talk2me_uploads"
|
||||
Environment="LOGS_DIR=$LOG_DIR"
|
||||
ExecStart=$VENV_DIR/bin/gunicorn --config gunicorn_config.py wsgi:application
|
||||
ExecReload=/bin/kill -s HUP \$MAINPID
|
||||
KillMode=mixed
|
||||
TimeoutStopSec=5
|
||||
Restart=always
|
||||
RestartSec=10
|
||||
|
||||
# Security settings
|
||||
NoNewPrivileges=true
|
||||
PrivateTmp=true
|
||||
ProtectSystem=strict
|
||||
ProtectHome=true
|
||||
ReadWritePaths=$LOG_DIR /tmp
|
||||
|
||||
[Install]
|
||||
WantedBy=multi-user.target
|
||||
EOF
|
||||
|
||||
# Create nginx configuration
|
||||
print_status "Creating nginx configuration"
|
||||
cat > /etc/nginx/sites-available/talk2me <<EOF
|
||||
server {
|
||||
listen 80;
|
||||
server_name _; # Replace with your domain
|
||||
|
||||
# Security headers
|
||||
add_header X-Content-Type-Options nosniff;
|
||||
add_header X-Frame-Options DENY;
|
||||
add_header X-XSS-Protection "1; mode=block";
|
||||
add_header Referrer-Policy "strict-origin-when-cross-origin";
|
||||
|
||||
# File upload size limit
|
||||
client_max_body_size 50M;
|
||||
client_body_buffer_size 1M;
|
||||
|
||||
# Timeouts for long audio processing
|
||||
proxy_connect_timeout 120s;
|
||||
proxy_send_timeout 120s;
|
||||
proxy_read_timeout 120s;
|
||||
|
||||
location / {
|
||||
proxy_pass http://127.0.0.1:5005;
|
||||
proxy_http_version 1.1;
|
||||
proxy_set_header Upgrade \$http_upgrade;
|
||||
proxy_set_header Connection 'upgrade';
|
||||
proxy_set_header Host \$host;
|
||||
proxy_set_header X-Real-IP \$remote_addr;
|
||||
proxy_set_header X-Forwarded-For \$proxy_add_x_forwarded_for;
|
||||
proxy_set_header X-Forwarded-Proto \$scheme;
|
||||
proxy_cache_bypass \$http_upgrade;
|
||||
|
||||
# Don't buffer responses
|
||||
proxy_buffering off;
|
||||
|
||||
# WebSocket support
|
||||
proxy_set_header Connection "upgrade";
|
||||
}
|
||||
|
||||
location /static {
|
||||
alias $APP_DIR/static;
|
||||
expires 1y;
|
||||
add_header Cache-Control "public, immutable";
|
||||
}
|
||||
|
||||
# Health check endpoint
|
||||
location /health {
|
||||
proxy_pass http://127.0.0.1:5005/health;
|
||||
access_log off;
|
||||
}
|
||||
}
|
||||
EOF
|
||||
|
||||
# Enable nginx site
|
||||
if [ -f /etc/nginx/sites-enabled/default ]; then
|
||||
rm /etc/nginx/sites-enabled/default
|
||||
fi
|
||||
ln -sf /etc/nginx/sites-available/talk2me /etc/nginx/sites-enabled/
|
||||
|
||||
# Set permissions
|
||||
chown -R $APP_USER:$APP_USER $APP_DIR
|
||||
|
||||
# Reload systemd
|
||||
print_status "Reloading systemd"
|
||||
systemctl daemon-reload
|
||||
|
||||
# Start services
|
||||
print_status "Starting services"
|
||||
systemctl enable talk2me
|
||||
systemctl restart talk2me
|
||||
systemctl restart nginx
|
||||
|
||||
# Wait for service to start
|
||||
sleep 5
|
||||
|
||||
# Check service status
|
||||
if systemctl is-active --quiet talk2me; then
|
||||
print_status "Talk2Me service is running"
|
||||
else
|
||||
print_error "Talk2Me service failed to start"
|
||||
journalctl -u talk2me -n 50
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Test health endpoint
|
||||
if curl -s http://localhost:5005/health | grep -q "healthy"; then
|
||||
print_status "Health check passed"
|
||||
else
|
||||
print_error "Health check failed"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
print_status "Deployment complete!"
|
||||
print_status "Talk2Me is now running at http://$(hostname -I | awk '{print $1}')"
|
||||
print_status "Check logs at: $LOG_DIR"
|
||||
print_status "Service status: systemctl status talk2me"
|
121
diagnose-pwa.py
Executable file
121
diagnose-pwa.py
Executable file
@ -0,0 +1,121 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
PWA Diagnostic Script for Talk2Me
|
||||
Checks common PWA installation issues
|
||||
"""
|
||||
|
||||
import requests
|
||||
import json
|
||||
import sys
|
||||
from urllib.parse import urljoin
|
||||
|
||||
def check_pwa(base_url):
|
||||
"""Check PWA requirements for the given URL"""
|
||||
|
||||
if not base_url.startswith(('http://', 'https://')):
|
||||
base_url = 'https://' + base_url
|
||||
|
||||
if not base_url.endswith('/'):
|
||||
base_url += '/'
|
||||
|
||||
print(f"Checking PWA for: {base_url}\n")
|
||||
|
||||
# Check HTTPS
|
||||
if not base_url.startswith('https://'):
|
||||
print("❌ PWA requires HTTPS (except for localhost)")
|
||||
return
|
||||
else:
|
||||
print("✅ HTTPS is enabled")
|
||||
|
||||
# Check main page
|
||||
try:
|
||||
response = requests.get(base_url, timeout=10)
|
||||
if response.status_code == 200:
|
||||
print("✅ Main page loads successfully")
|
||||
else:
|
||||
print(f"❌ Main page returned status code: {response.status_code}")
|
||||
except Exception as e:
|
||||
print(f"❌ Error loading main page: {e}")
|
||||
return
|
||||
|
||||
# Check manifest
|
||||
manifest_url = urljoin(base_url, '/static/manifest.json')
|
||||
print(f"\nChecking manifest at: {manifest_url}")
|
||||
try:
|
||||
response = requests.get(manifest_url, timeout=10)
|
||||
if response.status_code == 200:
|
||||
print("✅ Manifest file found")
|
||||
|
||||
# Parse manifest
|
||||
try:
|
||||
manifest = response.json()
|
||||
print(f" - Name: {manifest.get('name', 'Not set')}")
|
||||
print(f" - Short name: {manifest.get('short_name', 'Not set')}")
|
||||
print(f" - Display: {manifest.get('display', 'Not set')}")
|
||||
print(f" - Start URL: {manifest.get('start_url', 'Not set')}")
|
||||
print(f" - Icons: {len(manifest.get('icons', []))} defined")
|
||||
|
||||
# Check icons
|
||||
for icon in manifest.get('icons', []):
|
||||
icon_url = urljoin(base_url, icon['src'])
|
||||
try:
|
||||
icon_response = requests.head(icon_url, timeout=5)
|
||||
if icon_response.status_code == 200:
|
||||
print(f" ✅ {icon['sizes']}: {icon['src']}")
|
||||
else:
|
||||
print(f" ❌ {icon['sizes']}: {icon['src']} (Status: {icon_response.status_code})")
|
||||
except:
|
||||
print(f" ❌ {icon['sizes']}: {icon['src']} (Failed to load)")
|
||||
|
||||
except json.JSONDecodeError:
|
||||
print("❌ Manifest is not valid JSON")
|
||||
else:
|
||||
print(f"❌ Manifest returned status code: {response.status_code}")
|
||||
except Exception as e:
|
||||
print(f"❌ Error loading manifest: {e}")
|
||||
|
||||
# Check service worker
|
||||
sw_url = urljoin(base_url, '/service-worker.js')
|
||||
print(f"\nChecking service worker at: {sw_url}")
|
||||
try:
|
||||
response = requests.get(sw_url, timeout=10)
|
||||
if response.status_code == 200:
|
||||
print("✅ Service worker file found")
|
||||
content_type = response.headers.get('Content-Type', '')
|
||||
if 'javascript' in content_type:
|
||||
print("✅ Service worker has correct content type")
|
||||
else:
|
||||
print(f"⚠️ Service worker content type: {content_type}")
|
||||
else:
|
||||
print(f"❌ Service worker returned status code: {response.status_code}")
|
||||
except Exception as e:
|
||||
print(f"❌ Error loading service worker: {e}")
|
||||
|
||||
# Check favicon
|
||||
favicon_url = urljoin(base_url, '/static/icons/favicon.ico')
|
||||
print(f"\nChecking favicon at: {favicon_url}")
|
||||
try:
|
||||
response = requests.head(favicon_url, timeout=5)
|
||||
if response.status_code == 200:
|
||||
print("✅ Favicon found")
|
||||
else:
|
||||
print(f"⚠️ Favicon returned status code: {response.status_code}")
|
||||
except Exception as e:
|
||||
print(f"⚠️ Error loading favicon: {e}")
|
||||
|
||||
print("\n" + "="*50)
|
||||
print("PWA Installation Tips:")
|
||||
print("1. Clear browser cache and app data")
|
||||
print("2. Visit the site in Chrome on Android")
|
||||
print("3. Wait a few seconds for the install prompt")
|
||||
print("4. Or tap menu (⋮) → 'Add to Home screen'")
|
||||
print("5. Check Chrome DevTools → Application → Manifest")
|
||||
print("="*50)
|
||||
|
||||
if __name__ == "__main__":
|
||||
if len(sys.argv) > 1:
|
||||
url = sys.argv[1]
|
||||
else:
|
||||
url = input("Enter the URL to check (e.g., talk2me.dr74.net): ")
|
||||
|
||||
check_pwa(url)
|
19
docker-compose.amd.yml
Normal file
19
docker-compose.amd.yml
Normal file
@ -0,0 +1,19 @@
|
||||
version: '3.8'
|
||||
|
||||
# Docker Compose override for AMD GPU support (ROCm)
|
||||
# Usage: docker-compose -f docker-compose.yml -f docker-compose.amd.yml up
|
||||
|
||||
services:
|
||||
talk2me:
|
||||
environment:
|
||||
- HSA_OVERRIDE_GFX_VERSION=10.3.0 # Adjust based on your GPU model
|
||||
- ROCR_VISIBLE_DEVICES=0 # Use first GPU
|
||||
volumes:
|
||||
- /dev/kfd:/dev/kfd # ROCm KFD interface
|
||||
- /dev/dri:/dev/dri # Direct Rendering Interface
|
||||
devices:
|
||||
- /dev/kfd
|
||||
- /dev/dri
|
||||
group_add:
|
||||
- video # Required for GPU access
|
||||
- render # Required for GPU access
|
11
docker-compose.apple.yml
Normal file
11
docker-compose.apple.yml
Normal file
@ -0,0 +1,11 @@
|
||||
version: '3.8'
|
||||
|
||||
# Docker Compose override for Apple Silicon
|
||||
# Usage: docker-compose -f docker-compose.yml -f docker-compose.apple.yml up
|
||||
|
||||
services:
|
||||
talk2me:
|
||||
platform: linux/arm64/v8 # For M1/M2/M3 Macs
|
||||
environment:
|
||||
- PYTORCH_ENABLE_MPS_FALLBACK=1 # Enable Metal Performance Shaders fallback
|
||||
- PYTORCH_MPS_HIGH_WATERMARK_RATIO=0.7 # Memory management for MPS
|
16
docker-compose.nvidia.yml
Normal file
16
docker-compose.nvidia.yml
Normal file
@ -0,0 +1,16 @@
|
||||
version: '3.8'
|
||||
|
||||
# Docker Compose override for NVIDIA GPU support
|
||||
# Usage: docker-compose -f docker-compose.yml -f docker-compose.nvidia.yml up
|
||||
|
||||
services:
|
||||
talk2me:
|
||||
environment:
|
||||
- CUDA_VISIBLE_DEVICES=0 # Use first GPU
|
||||
deploy:
|
||||
resources:
|
||||
reservations:
|
||||
devices:
|
||||
- driver: nvidia
|
||||
count: 1
|
||||
capabilities: [gpu]
|
92
docker-compose.yml
Normal file
92
docker-compose.yml
Normal file
@ -0,0 +1,92 @@
|
||||
version: '3.8'
|
||||
|
||||
services:
|
||||
talk2me:
|
||||
build: .
|
||||
container_name: talk2me
|
||||
restart: unless-stopped
|
||||
ports:
|
||||
- "5005:5005"
|
||||
environment:
|
||||
- FLASK_ENV=production
|
||||
- UPLOAD_FOLDER=/tmp/talk2me_uploads
|
||||
- LOGS_DIR=/app/logs
|
||||
- TTS_SERVER_URL=${TTS_SERVER_URL:-http://localhost:5050/v1/audio/speech}
|
||||
- TTS_API_KEY=${TTS_API_KEY}
|
||||
- ADMIN_TOKEN=${ADMIN_TOKEN:-change-me-in-production}
|
||||
- SECRET_KEY=${SECRET_KEY:-change-me-in-production}
|
||||
- GUNICORN_WORKERS=${GUNICORN_WORKERS:-4}
|
||||
- GUNICORN_THREADS=${GUNICORN_THREADS:-2}
|
||||
- MEMORY_THRESHOLD_MB=${MEMORY_THRESHOLD_MB:-4096}
|
||||
- GPU_MEMORY_THRESHOLD_MB=${GPU_MEMORY_THRESHOLD_MB:-2048}
|
||||
volumes:
|
||||
- ./logs:/app/logs
|
||||
- talk2me_uploads:/tmp/talk2me_uploads
|
||||
- talk2me_models:/root/.cache/whisper # Whisper models cache
|
||||
deploy:
|
||||
resources:
|
||||
limits:
|
||||
memory: 4G
|
||||
reservations:
|
||||
memory: 2G
|
||||
healthcheck:
|
||||
test: ["CMD", "curl", "-f", "http://localhost:5005/health"]
|
||||
interval: 30s
|
||||
timeout: 10s
|
||||
retries: 3
|
||||
start_period: 40s
|
||||
networks:
|
||||
- talk2me_network
|
||||
|
||||
# Nginx reverse proxy (optional, for production)
|
||||
nginx:
|
||||
image: nginx:alpine
|
||||
container_name: talk2me_nginx
|
||||
restart: unless-stopped
|
||||
ports:
|
||||
- "80:80"
|
||||
- "443:443"
|
||||
volumes:
|
||||
- ./nginx.conf:/etc/nginx/conf.d/default.conf:ro
|
||||
- ./static:/app/static:ro
|
||||
- nginx_ssl:/etc/nginx/ssl
|
||||
depends_on:
|
||||
- talk2me
|
||||
networks:
|
||||
- talk2me_network
|
||||
|
||||
# Redis for session storage (optional)
|
||||
redis:
|
||||
image: redis:7-alpine
|
||||
container_name: talk2me_redis
|
||||
restart: unless-stopped
|
||||
command: redis-server --maxmemory 256mb --maxmemory-policy allkeys-lru
|
||||
volumes:
|
||||
- redis_data:/data
|
||||
networks:
|
||||
- talk2me_network
|
||||
|
||||
# PostgreSQL for persistent storage (optional)
|
||||
postgres:
|
||||
image: postgres:15-alpine
|
||||
container_name: talk2me_postgres
|
||||
restart: unless-stopped
|
||||
environment:
|
||||
- POSTGRES_DB=talk2me
|
||||
- POSTGRES_USER=talk2me
|
||||
- POSTGRES_PASSWORD=${POSTGRES_PASSWORD:-change-me-in-production}
|
||||
volumes:
|
||||
- postgres_data:/var/lib/postgresql/data
|
||||
networks:
|
||||
- talk2me_network
|
||||
|
||||
volumes:
|
||||
talk2me_uploads:
|
||||
talk2me_models:
|
||||
redis_data:
|
||||
postgres_data:
|
||||
nginx_ssl:
|
||||
|
||||
networks:
|
||||
talk2me_network:
|
||||
driver: bridge
|
564
error_logger.py
Normal file
564
error_logger.py
Normal file
@ -0,0 +1,564 @@
|
||||
# Comprehensive error logging system for production debugging
|
||||
import logging
|
||||
import logging.handlers
|
||||
import os
|
||||
import sys
|
||||
import json
|
||||
import traceback
|
||||
import time
|
||||
from datetime import datetime
|
||||
from typing import Dict, Any, Optional, Union
|
||||
from functools import wraps
|
||||
import socket
|
||||
import threading
|
||||
from flask import request, g
|
||||
from contextlib import contextmanager
|
||||
import hashlib
|
||||
|
||||
# Create logs directory if it doesn't exist
|
||||
LOGS_DIR = os.environ.get('LOGS_DIR', 'logs')
|
||||
os.makedirs(LOGS_DIR, exist_ok=True)
|
||||
|
||||
class StructuredFormatter(logging.Formatter):
|
||||
"""
|
||||
Custom formatter that outputs structured JSON logs
|
||||
"""
|
||||
def __init__(self, app_name='talk2me', environment='development'):
|
||||
super().__init__()
|
||||
self.app_name = app_name
|
||||
self.environment = environment
|
||||
self.hostname = socket.gethostname()
|
||||
|
||||
def format(self, record):
|
||||
# Base log structure
|
||||
log_data = {
|
||||
'timestamp': datetime.utcnow().isoformat() + 'Z',
|
||||
'level': record.levelname,
|
||||
'logger': record.name,
|
||||
'message': record.getMessage(),
|
||||
'app': self.app_name,
|
||||
'environment': self.environment,
|
||||
'hostname': self.hostname,
|
||||
'thread': threading.current_thread().name,
|
||||
'process': os.getpid()
|
||||
}
|
||||
|
||||
# Add exception info if present
|
||||
if record.exc_info:
|
||||
log_data['exception'] = {
|
||||
'type': record.exc_info[0].__name__,
|
||||
'message': str(record.exc_info[1]),
|
||||
'traceback': traceback.format_exception(*record.exc_info)
|
||||
}
|
||||
|
||||
# Add extra fields
|
||||
if hasattr(record, 'extra_fields'):
|
||||
log_data.update(record.extra_fields)
|
||||
|
||||
# Add Flask request context if available
|
||||
if hasattr(record, 'request_id'):
|
||||
log_data['request_id'] = record.request_id
|
||||
|
||||
if hasattr(record, 'user_id'):
|
||||
log_data['user_id'] = record.user_id
|
||||
|
||||
if hasattr(record, 'session_id'):
|
||||
log_data['session_id'] = record.session_id
|
||||
|
||||
# Add performance metrics if available
|
||||
if hasattr(record, 'duration'):
|
||||
log_data['duration_ms'] = record.duration
|
||||
|
||||
if hasattr(record, 'memory_usage'):
|
||||
log_data['memory_usage_mb'] = record.memory_usage
|
||||
|
||||
return json.dumps(log_data, default=str)
|
||||
|
||||
class ErrorLogger:
|
||||
"""
|
||||
Comprehensive error logging system with multiple handlers and features
|
||||
"""
|
||||
def __init__(self, app=None, config=None):
|
||||
self.config = config or {}
|
||||
self.loggers = {}
|
||||
self.error_counts = {}
|
||||
self.error_signatures = {}
|
||||
|
||||
if app:
|
||||
self.init_app(app)
|
||||
|
||||
def init_app(self, app):
|
||||
"""Initialize error logging for Flask app"""
|
||||
self.app = app
|
||||
|
||||
# Get configuration
|
||||
self.log_level = self.config.get('log_level',
|
||||
app.config.get('LOG_LEVEL', 'INFO'))
|
||||
self.log_file = self.config.get('log_file',
|
||||
app.config.get('LOG_FILE',
|
||||
os.path.join(LOGS_DIR, 'talk2me.log')))
|
||||
self.error_log_file = self.config.get('error_log_file',
|
||||
os.path.join(LOGS_DIR, 'errors.log'))
|
||||
self.max_bytes = self.config.get('max_bytes', 50 * 1024 * 1024) # 50MB
|
||||
self.backup_count = self.config.get('backup_count', 10)
|
||||
self.environment = app.config.get('FLASK_ENV', 'development')
|
||||
|
||||
# Set up loggers
|
||||
self._setup_app_logger()
|
||||
self._setup_error_logger()
|
||||
self._setup_access_logger()
|
||||
self._setup_security_logger()
|
||||
self._setup_performance_logger()
|
||||
|
||||
# Add Flask error handlers
|
||||
self._setup_flask_handlers(app)
|
||||
|
||||
# Add request logging
|
||||
app.before_request(self._before_request)
|
||||
app.after_request(self._after_request)
|
||||
|
||||
# Store logger in app
|
||||
app.error_logger = self
|
||||
|
||||
logging.info("Error logging system initialized")
|
||||
|
||||
def _setup_app_logger(self):
|
||||
"""Set up main application logger"""
|
||||
app_logger = logging.getLogger('talk2me')
|
||||
app_logger.setLevel(getattr(logging, self.log_level.upper()))
|
||||
|
||||
# Remove existing handlers
|
||||
app_logger.handlers = []
|
||||
|
||||
# Console handler with color support
|
||||
console_handler = logging.StreamHandler(sys.stdout)
|
||||
if sys.stdout.isatty():
|
||||
# Use colored output for terminals
|
||||
from colorlog import ColoredFormatter
|
||||
console_formatter = ColoredFormatter(
|
||||
'%(log_color)s%(asctime)s - %(name)s - %(levelname)s - %(message)s',
|
||||
log_colors={
|
||||
'DEBUG': 'cyan',
|
||||
'INFO': 'green',
|
||||
'WARNING': 'yellow',
|
||||
'ERROR': 'red',
|
||||
'CRITICAL': 'red,bg_white',
|
||||
}
|
||||
)
|
||||
console_handler.setFormatter(console_formatter)
|
||||
else:
|
||||
console_handler.setFormatter(
|
||||
StructuredFormatter('talk2me', self.environment)
|
||||
)
|
||||
app_logger.addHandler(console_handler)
|
||||
|
||||
# Rotating file handler
|
||||
file_handler = logging.handlers.RotatingFileHandler(
|
||||
self.log_file,
|
||||
maxBytes=self.max_bytes,
|
||||
backupCount=self.backup_count
|
||||
)
|
||||
file_handler.setFormatter(
|
||||
StructuredFormatter('talk2me', self.environment)
|
||||
)
|
||||
app_logger.addHandler(file_handler)
|
||||
|
||||
self.loggers['app'] = app_logger
|
||||
|
||||
def _setup_error_logger(self):
|
||||
"""Set up dedicated error logger"""
|
||||
error_logger = logging.getLogger('talk2me.errors')
|
||||
error_logger.setLevel(logging.ERROR)
|
||||
|
||||
# Error file handler
|
||||
error_handler = logging.handlers.RotatingFileHandler(
|
||||
self.error_log_file,
|
||||
maxBytes=self.max_bytes,
|
||||
backupCount=self.backup_count
|
||||
)
|
||||
error_handler.setFormatter(
|
||||
StructuredFormatter('talk2me', self.environment)
|
||||
)
|
||||
error_logger.addHandler(error_handler)
|
||||
|
||||
# Also send errors to syslog if available
|
||||
try:
|
||||
syslog_handler = logging.handlers.SysLogHandler(
|
||||
address='/dev/log' if os.path.exists('/dev/log') else ('localhost', 514)
|
||||
)
|
||||
syslog_handler.setFormatter(
|
||||
logging.Formatter('talk2me[%(process)d]: %(levelname)s %(message)s')
|
||||
)
|
||||
error_logger.addHandler(syslog_handler)
|
||||
except Exception:
|
||||
pass # Syslog not available
|
||||
|
||||
self.loggers['error'] = error_logger
|
||||
|
||||
def _setup_access_logger(self):
|
||||
"""Set up access logger for HTTP requests"""
|
||||
access_logger = logging.getLogger('talk2me.access')
|
||||
access_logger.setLevel(logging.INFO)
|
||||
|
||||
# Access log file
|
||||
access_handler = logging.handlers.TimedRotatingFileHandler(
|
||||
os.path.join(LOGS_DIR, 'access.log'),
|
||||
when='midnight',
|
||||
interval=1,
|
||||
backupCount=30
|
||||
)
|
||||
access_handler.setFormatter(
|
||||
StructuredFormatter('talk2me', self.environment)
|
||||
)
|
||||
access_logger.addHandler(access_handler)
|
||||
|
||||
self.loggers['access'] = access_logger
|
||||
|
||||
def _setup_security_logger(self):
|
||||
"""Set up security event logger"""
|
||||
security_logger = logging.getLogger('talk2me.security')
|
||||
security_logger.setLevel(logging.WARNING)
|
||||
|
||||
# Security log file
|
||||
security_handler = logging.handlers.RotatingFileHandler(
|
||||
os.path.join(LOGS_DIR, 'security.log'),
|
||||
maxBytes=self.max_bytes,
|
||||
backupCount=self.backup_count
|
||||
)
|
||||
security_handler.setFormatter(
|
||||
StructuredFormatter('talk2me', self.environment)
|
||||
)
|
||||
security_logger.addHandler(security_handler)
|
||||
|
||||
self.loggers['security'] = security_logger
|
||||
|
||||
def _setup_performance_logger(self):
|
||||
"""Set up performance metrics logger"""
|
||||
perf_logger = logging.getLogger('talk2me.performance')
|
||||
perf_logger.setLevel(logging.INFO)
|
||||
|
||||
# Performance log file
|
||||
perf_handler = logging.handlers.TimedRotatingFileHandler(
|
||||
os.path.join(LOGS_DIR, 'performance.log'),
|
||||
when='H', # Hourly rotation
|
||||
interval=1,
|
||||
backupCount=168 # 7 days
|
||||
)
|
||||
perf_handler.setFormatter(
|
||||
StructuredFormatter('talk2me', self.environment)
|
||||
)
|
||||
perf_logger.addHandler(perf_handler)
|
||||
|
||||
self.loggers['performance'] = perf_logger
|
||||
|
||||
def _setup_flask_handlers(self, app):
|
||||
"""Set up Flask error handlers"""
|
||||
@app.errorhandler(Exception)
|
||||
def handle_exception(error):
|
||||
# Get request ID
|
||||
request_id = getattr(g, 'request_id', 'unknown')
|
||||
|
||||
# Create error signature for deduplication
|
||||
error_signature = self._get_error_signature(error)
|
||||
|
||||
# Log the error
|
||||
self.log_error(
|
||||
error,
|
||||
request_id=request_id,
|
||||
endpoint=request.endpoint,
|
||||
method=request.method,
|
||||
path=request.path,
|
||||
ip=request.remote_addr,
|
||||
user_agent=request.headers.get('User-Agent'),
|
||||
signature=error_signature
|
||||
)
|
||||
|
||||
# Track error frequency
|
||||
self._track_error(error_signature)
|
||||
|
||||
# Return appropriate response
|
||||
if hasattr(error, 'code'):
|
||||
return {'error': str(error)}, error.code
|
||||
else:
|
||||
return {'error': 'Internal server error'}, 500
|
||||
|
||||
def _before_request(self):
|
||||
"""Log request start"""
|
||||
# Generate request ID
|
||||
g.request_id = self._generate_request_id()
|
||||
g.request_start_time = time.time()
|
||||
|
||||
# Log access
|
||||
self.log_access(
|
||||
'request_start',
|
||||
request_id=g.request_id,
|
||||
method=request.method,
|
||||
path=request.path,
|
||||
ip=request.remote_addr,
|
||||
user_agent=request.headers.get('User-Agent')
|
||||
)
|
||||
|
||||
def _after_request(self, response):
|
||||
"""Log request completion"""
|
||||
# Calculate duration
|
||||
duration = None
|
||||
if hasattr(g, 'request_start_time'):
|
||||
duration = int((time.time() - g.request_start_time) * 1000)
|
||||
|
||||
# Log access
|
||||
self.log_access(
|
||||
'request_complete',
|
||||
request_id=getattr(g, 'request_id', 'unknown'),
|
||||
method=request.method,
|
||||
path=request.path,
|
||||
status=response.status_code,
|
||||
duration_ms=duration,
|
||||
content_length=response.content_length
|
||||
)
|
||||
|
||||
# Log performance metrics for slow requests
|
||||
if duration and duration > 1000: # Over 1 second
|
||||
self.log_performance(
|
||||
'slow_request',
|
||||
request_id=getattr(g, 'request_id', 'unknown'),
|
||||
endpoint=request.endpoint,
|
||||
duration_ms=duration
|
||||
)
|
||||
|
||||
return response
|
||||
|
||||
def log_error(self, error: Exception, **kwargs):
|
||||
"""Log an error with context"""
|
||||
error_logger = self.loggers.get('error', logging.getLogger())
|
||||
|
||||
# Create log record with extra fields
|
||||
extra = {
|
||||
'extra_fields': kwargs,
|
||||
'request_id': kwargs.get('request_id'),
|
||||
'user_id': kwargs.get('user_id'),
|
||||
'session_id': kwargs.get('session_id')
|
||||
}
|
||||
|
||||
# Log with full traceback
|
||||
error_logger.error(
|
||||
f"Error in {kwargs.get('endpoint', 'unknown')}: {str(error)}",
|
||||
exc_info=sys.exc_info(),
|
||||
extra=extra
|
||||
)
|
||||
|
||||
def log_access(self, message: str, **kwargs):
|
||||
"""Log access event"""
|
||||
access_logger = self.loggers.get('access', logging.getLogger())
|
||||
|
||||
extra = {
|
||||
'extra_fields': kwargs,
|
||||
'request_id': kwargs.get('request_id')
|
||||
}
|
||||
|
||||
access_logger.info(message, extra=extra)
|
||||
|
||||
def log_security(self, event: str, severity: str = 'warning', **kwargs):
|
||||
"""Log security event"""
|
||||
security_logger = self.loggers.get('security', logging.getLogger())
|
||||
|
||||
extra = {
|
||||
'extra_fields': {
|
||||
'event': event,
|
||||
'severity': severity,
|
||||
**kwargs
|
||||
},
|
||||
'request_id': kwargs.get('request_id')
|
||||
}
|
||||
|
||||
log_method = getattr(security_logger, severity.lower(), security_logger.warning)
|
||||
log_method(f"Security event: {event}", extra=extra)
|
||||
|
||||
def log_performance(self, metric: str, value: Union[int, float] = None, **kwargs):
|
||||
"""Log performance metric"""
|
||||
perf_logger = self.loggers.get('performance', logging.getLogger())
|
||||
|
||||
extra = {
|
||||
'extra_fields': {
|
||||
'metric': metric,
|
||||
'value': value,
|
||||
**kwargs
|
||||
},
|
||||
'request_id': kwargs.get('request_id')
|
||||
}
|
||||
|
||||
perf_logger.info(f"Performance metric: {metric}", extra=extra)
|
||||
|
||||
def _generate_request_id(self):
|
||||
"""Generate unique request ID"""
|
||||
return f"{int(time.time() * 1000)}-{os.urandom(8).hex()}"
|
||||
|
||||
def _get_error_signature(self, error: Exception):
|
||||
"""Generate signature for error deduplication"""
|
||||
# Create signature from error type and key parts of traceback
|
||||
tb_summary = traceback.format_exception_only(type(error), error)
|
||||
signature_data = f"{type(error).__name__}:{tb_summary[0] if tb_summary else ''}"
|
||||
return hashlib.md5(signature_data.encode()).hexdigest()
|
||||
|
||||
def _track_error(self, signature: str):
|
||||
"""Track error frequency"""
|
||||
now = time.time()
|
||||
|
||||
if signature not in self.error_counts:
|
||||
self.error_counts[signature] = []
|
||||
|
||||
# Add current timestamp
|
||||
self.error_counts[signature].append(now)
|
||||
|
||||
# Clean old entries (keep last hour)
|
||||
self.error_counts[signature] = [
|
||||
ts for ts in self.error_counts[signature]
|
||||
if now - ts < 3600
|
||||
]
|
||||
|
||||
# Alert if error rate is high
|
||||
error_count = len(self.error_counts[signature])
|
||||
if error_count > 10: # More than 10 in an hour
|
||||
self.log_security(
|
||||
'high_error_rate',
|
||||
severity='error',
|
||||
signature=signature,
|
||||
count=error_count,
|
||||
message="High error rate detected"
|
||||
)
|
||||
|
||||
def get_error_summary(self):
|
||||
"""Get summary of recent errors"""
|
||||
summary = {}
|
||||
now = time.time()
|
||||
|
||||
for signature, timestamps in self.error_counts.items():
|
||||
recent_count = len([ts for ts in timestamps if now - ts < 3600])
|
||||
if recent_count > 0:
|
||||
summary[signature] = {
|
||||
'count_last_hour': recent_count,
|
||||
'last_seen': max(timestamps)
|
||||
}
|
||||
|
||||
return summary
|
||||
|
||||
# Decorators for easy logging
|
||||
def log_errors(logger_name='talk2me'):
|
||||
"""Decorator to log function errors"""
|
||||
def decorator(func):
|
||||
@wraps(func)
|
||||
def wrapper(*args, **kwargs):
|
||||
try:
|
||||
return func(*args, **kwargs)
|
||||
except Exception as e:
|
||||
logger = logging.getLogger(logger_name)
|
||||
logger.error(
|
||||
f"Error in {func.__name__}: {str(e)}",
|
||||
exc_info=sys.exc_info(),
|
||||
extra={
|
||||
'extra_fields': {
|
||||
'function': func.__name__,
|
||||
'module': func.__module__
|
||||
}
|
||||
}
|
||||
)
|
||||
raise
|
||||
return wrapper
|
||||
return decorator
|
||||
|
||||
def log_performance(metric_name=None):
|
||||
"""Decorator to log function performance"""
|
||||
def decorator(func):
|
||||
@wraps(func)
|
||||
def wrapper(*args, **kwargs):
|
||||
start_time = time.time()
|
||||
|
||||
try:
|
||||
result = func(*args, **kwargs)
|
||||
duration = int((time.time() - start_time) * 1000)
|
||||
|
||||
# Log performance
|
||||
logger = logging.getLogger('talk2me.performance')
|
||||
logger.info(
|
||||
f"Performance: {metric_name or func.__name__}",
|
||||
extra={
|
||||
'extra_fields': {
|
||||
'metric': metric_name or func.__name__,
|
||||
'duration_ms': duration,
|
||||
'function': func.__name__,
|
||||
'module': func.__module__
|
||||
}
|
||||
}
|
||||
)
|
||||
|
||||
return result
|
||||
except Exception:
|
||||
duration = int((time.time() - start_time) * 1000)
|
||||
logger = logging.getLogger('talk2me.performance')
|
||||
logger.warning(
|
||||
f"Performance (failed): {metric_name or func.__name__}",
|
||||
extra={
|
||||
'extra_fields': {
|
||||
'metric': metric_name or func.__name__,
|
||||
'duration_ms': duration,
|
||||
'function': func.__name__,
|
||||
'module': func.__module__,
|
||||
'status': 'failed'
|
||||
}
|
||||
}
|
||||
)
|
||||
raise
|
||||
return wrapper
|
||||
return decorator
|
||||
|
||||
@contextmanager
|
||||
def log_context(**kwargs):
|
||||
"""Context manager to add context to logs"""
|
||||
# Store current context
|
||||
old_context = {}
|
||||
for key, value in kwargs.items():
|
||||
if hasattr(g, key):
|
||||
old_context[key] = getattr(g, key)
|
||||
setattr(g, key, value)
|
||||
|
||||
try:
|
||||
yield
|
||||
finally:
|
||||
# Restore old context
|
||||
for key in kwargs:
|
||||
if key in old_context:
|
||||
setattr(g, key, old_context[key])
|
||||
else:
|
||||
delattr(g, key)
|
||||
|
||||
# Utility functions
|
||||
def configure_logging(app, **kwargs):
|
||||
"""Configure logging for the application"""
|
||||
config = {
|
||||
'log_level': kwargs.get('log_level', app.config.get('LOG_LEVEL', 'INFO')),
|
||||
'log_file': kwargs.get('log_file', app.config.get('LOG_FILE')),
|
||||
'error_log_file': kwargs.get('error_log_file'),
|
||||
'max_bytes': kwargs.get('max_bytes', 50 * 1024 * 1024),
|
||||
'backup_count': kwargs.get('backup_count', 10)
|
||||
}
|
||||
|
||||
error_logger = ErrorLogger(app, config)
|
||||
return error_logger
|
||||
|
||||
def get_logger(name='talk2me'):
|
||||
"""Get a logger instance"""
|
||||
return logging.getLogger(name)
|
||||
|
||||
def log_exception(error, message=None, **kwargs):
|
||||
"""Log an exception with context"""
|
||||
logger = logging.getLogger('talk2me.errors')
|
||||
|
||||
extra = {
|
||||
'extra_fields': kwargs,
|
||||
'request_id': getattr(g, 'request_id', None)
|
||||
}
|
||||
|
||||
logger.error(
|
||||
message or f"Exception: {str(error)}",
|
||||
exc_info=(type(error), error, error.__traceback__),
|
||||
extra=extra
|
||||
)
|
86
gunicorn_config.py
Normal file
86
gunicorn_config.py
Normal file
@ -0,0 +1,86 @@
|
||||
"""
|
||||
Gunicorn configuration for production deployment
|
||||
"""
|
||||
import multiprocessing
|
||||
import os
|
||||
|
||||
# Server socket
|
||||
bind = os.environ.get('GUNICORN_BIND', '0.0.0.0:5005')
|
||||
backlog = 2048
|
||||
|
||||
# Worker processes
|
||||
# Use 2-4 workers per CPU core
|
||||
workers = int(os.environ.get('GUNICORN_WORKERS', multiprocessing.cpu_count() * 2 + 1))
|
||||
worker_class = 'sync' # Use 'gevent' for async if needed
|
||||
worker_connections = 1000
|
||||
timeout = 120 # Increased for audio processing
|
||||
keepalive = 5
|
||||
|
||||
# Restart workers after this many requests, to help prevent memory leaks
|
||||
max_requests = 1000
|
||||
max_requests_jitter = 50
|
||||
|
||||
# Preload the application
|
||||
preload_app = True
|
||||
|
||||
# Server mechanics
|
||||
daemon = False
|
||||
pidfile = os.environ.get('GUNICORN_PID', '/tmp/talk2me.pid')
|
||||
user = None
|
||||
group = None
|
||||
tmp_upload_dir = None
|
||||
|
||||
# Logging
|
||||
accesslog = os.environ.get('GUNICORN_ACCESS_LOG', '-')
|
||||
errorlog = os.environ.get('GUNICORN_ERROR_LOG', '-')
|
||||
loglevel = os.environ.get('GUNICORN_LOG_LEVEL', 'info')
|
||||
access_log_format = '%(h)s %(l)s %(u)s %(t)s "%(r)s" %(s)s %(b)s "%(f)s" "%(a)s" %(D)s'
|
||||
|
||||
# Process naming
|
||||
proc_name = 'talk2me'
|
||||
|
||||
# Server hooks
|
||||
def when_ready(server):
|
||||
"""Called just after the server is started."""
|
||||
server.log.info("Server is ready. Spawning workers")
|
||||
|
||||
def worker_int(worker):
|
||||
"""Called just after a worker exited on SIGINT or SIGQUIT."""
|
||||
worker.log.info("Worker received INT or QUIT signal")
|
||||
|
||||
def pre_fork(server, worker):
|
||||
"""Called just before a worker is forked."""
|
||||
server.log.info(f"Forking worker {worker}")
|
||||
|
||||
def post_fork(server, worker):
|
||||
"""Called just after a worker has been forked."""
|
||||
server.log.info(f"Worker spawned (pid: {worker.pid})")
|
||||
|
||||
def worker_exit(server, worker):
|
||||
"""Called just after a worker has been killed."""
|
||||
server.log.info(f"Worker exit (pid: {worker.pid})")
|
||||
|
||||
def pre_request(worker, req):
|
||||
"""Called just before a worker processes the request."""
|
||||
worker.log.debug(f"{req.method} {req.path}")
|
||||
|
||||
def post_request(worker, req, environ, resp):
|
||||
"""Called after a worker processes the request."""
|
||||
worker.log.debug(f"{req.method} {req.path} - {resp.status}")
|
||||
|
||||
# SSL/TLS (uncomment if using HTTPS directly)
|
||||
# keyfile = '/path/to/keyfile'
|
||||
# certfile = '/path/to/certfile'
|
||||
# ssl_version = 'TLSv1_2'
|
||||
# cert_reqs = 'required'
|
||||
# ca_certs = '/path/to/ca_certs'
|
||||
|
||||
# Thread option (if using threaded workers)
|
||||
threads = int(os.environ.get('GUNICORN_THREADS', 1))
|
||||
|
||||
# Silent health checks in logs
|
||||
def pre_request(worker, req):
|
||||
if req.path in ['/health', '/health/live']:
|
||||
# Don't log health checks
|
||||
return
|
||||
worker.log.debug(f"{req.method} {req.path}")
|
91
health-monitor.py
Executable file
91
health-monitor.py
Executable file
@ -0,0 +1,91 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
Health monitoring script for Talk2Me application
|
||||
Usage: python health-monitor.py [--detailed] [--interval SECONDS]
|
||||
"""
|
||||
|
||||
import requests
|
||||
import time
|
||||
import argparse
|
||||
import json
|
||||
from datetime import datetime
|
||||
|
||||
def check_health(url, detailed=False):
|
||||
"""Check health of the Talk2Me service"""
|
||||
endpoint = f"{url}/health/detailed" if detailed else f"{url}/health"
|
||||
|
||||
try:
|
||||
response = requests.get(endpoint, timeout=5)
|
||||
data = response.json()
|
||||
|
||||
if detailed:
|
||||
print(f"\n=== Health Check at {datetime.now().strftime('%Y-%m-%d %H:%M:%S')} ===")
|
||||
print(f"Overall Status: {data['status'].upper()}")
|
||||
print("\nComponent Status:")
|
||||
for component, status in data['components'].items():
|
||||
status_icon = "✅" if status.get('status') == 'healthy' else "❌"
|
||||
print(f" {status_icon} {component}: {status.get('status', 'unknown')}")
|
||||
if 'error' in status:
|
||||
print(f" Error: {status['error']}")
|
||||
if 'device' in status:
|
||||
print(f" Device: {status['device']}")
|
||||
if 'model_size' in status:
|
||||
print(f" Model: {status['model_size']}")
|
||||
|
||||
if 'metrics' in data:
|
||||
print("\nMetrics:")
|
||||
uptime = data['metrics'].get('uptime', 0)
|
||||
hours = int(uptime // 3600)
|
||||
minutes = int((uptime % 3600) // 60)
|
||||
print(f" Uptime: {hours}h {minutes}m")
|
||||
print(f" Request Count: {data['metrics'].get('request_count', 0)}")
|
||||
else:
|
||||
status_icon = "✅" if response.status_code == 200 else "❌"
|
||||
print(f"{status_icon} {datetime.now().strftime('%H:%M:%S')} - Status: {data.get('status', 'unknown')}")
|
||||
|
||||
return response.status_code == 200
|
||||
|
||||
except requests.exceptions.ConnectionError:
|
||||
print(f"❌ {datetime.now().strftime('%H:%M:%S')} - Connection failed")
|
||||
return False
|
||||
except requests.exceptions.Timeout:
|
||||
print(f"❌ {datetime.now().strftime('%H:%M:%S')} - Request timeout")
|
||||
return False
|
||||
except Exception as e:
|
||||
print(f"❌ {datetime.now().strftime('%H:%M:%S')} - Error: {str(e)}")
|
||||
return False
|
||||
|
||||
def main():
|
||||
parser = argparse.ArgumentParser(description='Monitor Talk2Me service health')
|
||||
parser.add_argument('--url', default='http://localhost:5005', help='Service URL')
|
||||
parser.add_argument('--detailed', action='store_true', help='Show detailed health info')
|
||||
parser.add_argument('--interval', type=int, default=30, help='Check interval in seconds')
|
||||
parser.add_argument('--once', action='store_true', help='Run once and exit')
|
||||
|
||||
args = parser.parse_args()
|
||||
|
||||
print(f"Monitoring {args.url}")
|
||||
print("Press Ctrl+C to stop\n")
|
||||
|
||||
consecutive_failures = 0
|
||||
|
||||
try:
|
||||
while True:
|
||||
success = check_health(args.url, args.detailed)
|
||||
|
||||
if not success:
|
||||
consecutive_failures += 1
|
||||
if consecutive_failures >= 3:
|
||||
print(f"\n⚠️ ALERT: Service has been down for {consecutive_failures} consecutive checks!")
|
||||
else:
|
||||
consecutive_failures = 0
|
||||
|
||||
if args.once:
|
||||
break
|
||||
|
||||
time.sleep(args.interval)
|
||||
except KeyboardInterrupt:
|
||||
print("\n\nMonitoring stopped.")
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
75
init_all_databases.py
Executable file
75
init_all_databases.py
Executable file
@ -0,0 +1,75 @@
|
||||
#!/usr/bin/env python3
|
||||
"""Initialize all database tables for Talk2Me"""
|
||||
|
||||
import os
|
||||
import sys
|
||||
import subprocess
|
||||
import logging
|
||||
from dotenv import load_dotenv
|
||||
|
||||
# Load environment variables
|
||||
load_dotenv()
|
||||
|
||||
logging.basicConfig(level=logging.INFO)
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
def run_script(script_name):
|
||||
"""Run a Python script and return success status"""
|
||||
try:
|
||||
logger.info(f"Running {script_name}...")
|
||||
result = subprocess.run([sys.executable, script_name], capture_output=True, text=True)
|
||||
|
||||
if result.returncode == 0:
|
||||
logger.info(f"✓ {script_name} completed successfully")
|
||||
return True
|
||||
else:
|
||||
logger.error(f"✗ {script_name} failed with return code {result.returncode}")
|
||||
if result.stderr:
|
||||
logger.error(f"Error output: {result.stderr}")
|
||||
return False
|
||||
except Exception as e:
|
||||
logger.error(f"✗ Failed to run {script_name}: {e}")
|
||||
return False
|
||||
|
||||
def main():
|
||||
"""Initialize all databases"""
|
||||
logger.info("=== Talk2Me Database Initialization ===")
|
||||
|
||||
# Check if DATABASE_URL is set
|
||||
if not os.environ.get('DATABASE_URL'):
|
||||
logger.error("DATABASE_URL environment variable not set!")
|
||||
logger.info("Please set DATABASE_URL in your .env file")
|
||||
logger.info("Example: DATABASE_URL=postgresql://postgres:password@localhost:5432/talk2me")
|
||||
return False
|
||||
|
||||
logger.info(f"Using database: {os.environ.get('DATABASE_URL')}")
|
||||
|
||||
scripts = [
|
||||
"database_init.py", # Initialize SQLAlchemy models
|
||||
"init_auth_db.py", # Initialize authentication tables
|
||||
"init_analytics_db.py" # Initialize analytics tables
|
||||
]
|
||||
|
||||
success = True
|
||||
for script in scripts:
|
||||
if os.path.exists(script):
|
||||
if not run_script(script):
|
||||
success = False
|
||||
else:
|
||||
logger.warning(f"Script {script} not found, skipping...")
|
||||
|
||||
if success:
|
||||
logger.info("\n✅ All database initialization completed successfully!")
|
||||
logger.info("\nYou can now:")
|
||||
logger.info("1. Create an admin user by calling POST /api/init-admin-user")
|
||||
logger.info("2. Or use the admin token to log in and create users")
|
||||
logger.info("3. Check /api/test-auth to verify authentication is working")
|
||||
else:
|
||||
logger.error("\n❌ Some database initialization steps failed!")
|
||||
logger.info("Please check the errors above and try again")
|
||||
|
||||
return success
|
||||
|
||||
if __name__ == "__main__":
|
||||
success = main()
|
||||
sys.exit(0 if success else 1)
|
72
init_analytics_db.py
Executable file
72
init_analytics_db.py
Executable file
@ -0,0 +1,72 @@
|
||||
#!/usr/bin/env python3
|
||||
"""Initialize analytics database tables"""
|
||||
|
||||
import os
|
||||
import sys
|
||||
import psycopg2
|
||||
from psycopg2.extensions import ISOLATION_LEVEL_AUTOCOMMIT
|
||||
import logging
|
||||
from dotenv import load_dotenv
|
||||
|
||||
# Load environment variables
|
||||
load_dotenv()
|
||||
|
||||
logging.basicConfig(level=logging.INFO)
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
def init_analytics_db():
|
||||
"""Initialize analytics database tables"""
|
||||
|
||||
# Get database URL from environment
|
||||
database_url = os.environ.get('DATABASE_URL', 'postgresql://localhost/talk2me')
|
||||
|
||||
try:
|
||||
# Connect to PostgreSQL
|
||||
conn = psycopg2.connect(database_url)
|
||||
conn.set_isolation_level(ISOLATION_LEVEL_AUTOCOMMIT)
|
||||
cursor = conn.cursor()
|
||||
|
||||
logger.info("Connected to PostgreSQL database")
|
||||
|
||||
# Read SQL file
|
||||
sql_file = os.path.join(os.path.dirname(__file__), 'migrations', 'create_analytics_tables.sql')
|
||||
|
||||
if not os.path.exists(sql_file):
|
||||
logger.error(f"SQL file not found: {sql_file}")
|
||||
return False
|
||||
|
||||
with open(sql_file, 'r') as f:
|
||||
sql_content = f.read()
|
||||
|
||||
# Execute SQL commands
|
||||
logger.info("Creating analytics tables...")
|
||||
cursor.execute(sql_content)
|
||||
|
||||
logger.info("Analytics tables created successfully!")
|
||||
|
||||
# Verify tables were created
|
||||
cursor.execute("""
|
||||
SELECT table_name
|
||||
FROM information_schema.tables
|
||||
WHERE table_schema = 'public'
|
||||
AND table_name IN (
|
||||
'error_logs', 'request_logs', 'translation_logs',
|
||||
'transcription_logs', 'tts_logs', 'daily_stats'
|
||||
)
|
||||
""")
|
||||
|
||||
created_tables = [row[0] for row in cursor.fetchall()]
|
||||
logger.info(f"Created tables: {', '.join(created_tables)}")
|
||||
|
||||
cursor.close()
|
||||
conn.close()
|
||||
|
||||
return True
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to initialize analytics database: {e}")
|
||||
return False
|
||||
|
||||
if __name__ == "__main__":
|
||||
success = init_analytics_db()
|
||||
sys.exit(0 if success else 1)
|
149
init_auth_db.py
Normal file
149
init_auth_db.py
Normal file
@ -0,0 +1,149 @@
|
||||
#!/usr/bin/env python3
|
||||
"""Initialize authentication database and create default admin user"""
|
||||
|
||||
import os
|
||||
import sys
|
||||
import getpass
|
||||
from flask import Flask
|
||||
from flask_sqlalchemy import SQLAlchemy
|
||||
from dotenv import load_dotenv
|
||||
|
||||
# Load environment variables
|
||||
load_dotenv()
|
||||
|
||||
# Add parent directory to path
|
||||
sys.path.insert(0, os.path.dirname(os.path.abspath(__file__)))
|
||||
|
||||
from config import init_app as init_config
|
||||
from database import db, init_db
|
||||
from auth_models import User, bcrypt
|
||||
from auth import create_user
|
||||
|
||||
def create_admin_user():
|
||||
"""Create the default admin user"""
|
||||
# Skip if running non-interactively
|
||||
if not sys.stdin.isatty():
|
||||
print("Running non-interactively, skipping interactive admin creation.")
|
||||
return True
|
||||
|
||||
print("\n=== Talk2Me Admin User Setup ===\n")
|
||||
|
||||
# Get admin credentials
|
||||
while True:
|
||||
email = input("Admin email: ").strip()
|
||||
if '@' in email and '.' in email:
|
||||
break
|
||||
print("Please enter a valid email address.")
|
||||
|
||||
while True:
|
||||
username = input("Admin username: ").strip()
|
||||
if len(username) >= 3:
|
||||
break
|
||||
print("Username must be at least 3 characters.")
|
||||
|
||||
while True:
|
||||
password = getpass.getpass("Admin password (min 8 chars): ")
|
||||
if len(password) >= 8:
|
||||
password_confirm = getpass.getpass("Confirm password: ")
|
||||
if password == password_confirm:
|
||||
break
|
||||
print("Passwords don't match. Try again.")
|
||||
else:
|
||||
print("Password must be at least 8 characters.")
|
||||
|
||||
full_name = input("Full name (optional): ").strip() or None
|
||||
|
||||
# Create admin user
|
||||
print("\nCreating admin user...")
|
||||
user, error = create_user(
|
||||
email=email,
|
||||
username=username,
|
||||
password=password,
|
||||
full_name=full_name,
|
||||
role='admin',
|
||||
is_verified=True
|
||||
)
|
||||
|
||||
if error:
|
||||
print(f"Error creating admin: {error}")
|
||||
return False
|
||||
|
||||
# Set higher rate limits for admin
|
||||
user.rate_limit_per_minute = 300
|
||||
user.rate_limit_per_hour = 5000
|
||||
user.rate_limit_per_day = 50000
|
||||
|
||||
# Add all permissions
|
||||
user.permissions = ['all']
|
||||
|
||||
db.session.commit()
|
||||
|
||||
print(f"\n✅ Admin user created successfully!")
|
||||
print(f" Email: {user.email}")
|
||||
print(f" Username: {user.username}")
|
||||
print(f" API Key: {user.api_key}")
|
||||
print(f"\n📝 Save your API key securely. You can use it to authenticate API requests.")
|
||||
print(f"\n🔐 Login at: http://localhost:5005/login")
|
||||
print(f"📊 Admin dashboard: http://localhost:5005/admin/users")
|
||||
|
||||
return True
|
||||
|
||||
|
||||
def init_database():
|
||||
"""Initialize the database with all tables"""
|
||||
# Create Flask app
|
||||
app = Flask(__name__)
|
||||
|
||||
# Initialize configuration
|
||||
init_config(app)
|
||||
|
||||
# Initialize bcrypt
|
||||
bcrypt.init_app(app)
|
||||
|
||||
# Initialize database
|
||||
init_db(app)
|
||||
|
||||
with app.app_context():
|
||||
print("Creating database tables...")
|
||||
|
||||
# Import all models to ensure they're registered
|
||||
from auth_models import User, LoginHistory, UserSession, RevokedToken
|
||||
from database import Translation, Transcription, UserPreferences, UsageAnalytics, ApiKey
|
||||
|
||||
# Create all tables
|
||||
db.create_all()
|
||||
print("✅ Database tables created successfully!")
|
||||
|
||||
# Check if admin user already exists
|
||||
admin_exists = User.query.filter_by(role='admin').first()
|
||||
|
||||
if admin_exists:
|
||||
print(f"\n⚠️ Admin user already exists: {admin_exists.username}")
|
||||
# Skip creating new admin if running non-interactively
|
||||
if not sys.stdin.isatty():
|
||||
print("Running non-interactively, skipping admin user creation.")
|
||||
return
|
||||
create_new = input("Create another admin user? (y/n): ").lower().strip()
|
||||
if create_new != 'y':
|
||||
print("\nExiting without creating new admin.")
|
||||
return
|
||||
|
||||
# Create admin user
|
||||
if not create_admin_user():
|
||||
print("\n❌ Failed to create admin user.")
|
||||
sys.exit(1)
|
||||
|
||||
print("\n✨ Authentication system initialized successfully!")
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
try:
|
||||
init_database()
|
||||
except KeyboardInterrupt:
|
||||
print("\n\nSetup cancelled by user.")
|
||||
sys.exit(0)
|
||||
except Exception as e:
|
||||
print(f"\n❌ Error during setup: {str(e)}")
|
||||
import traceback
|
||||
traceback.print_exc()
|
||||
sys.exit(1)
|
117
maintenance.sh
Executable file
117
maintenance.sh
Executable file
@ -0,0 +1,117 @@
|
||||
#!/bin/bash
|
||||
|
||||
# Maintenance script for Talk2Me application
|
||||
# This script helps manage temporary files and disk space
|
||||
|
||||
UPLOAD_FOLDER="${UPLOAD_FOLDER:-/tmp/talk2me_uploads}"
|
||||
MAX_AGE_MINUTES=5
|
||||
|
||||
echo "Talk2Me Maintenance Script"
|
||||
echo "========================="
|
||||
|
||||
# Function to check disk usage
|
||||
check_disk_usage() {
|
||||
echo -e "\nDisk Usage:"
|
||||
df -h "$UPLOAD_FOLDER" 2>/dev/null || df -h /tmp
|
||||
}
|
||||
|
||||
# Function to show temp file stats
|
||||
show_temp_stats() {
|
||||
echo -e "\nTemporary File Statistics:"
|
||||
if [ -d "$UPLOAD_FOLDER" ]; then
|
||||
file_count=$(find "$UPLOAD_FOLDER" -type f 2>/dev/null | wc -l)
|
||||
total_size=$(du -sh "$UPLOAD_FOLDER" 2>/dev/null | cut -f1)
|
||||
echo " Upload folder: $UPLOAD_FOLDER"
|
||||
echo " File count: $file_count"
|
||||
echo " Total size: ${total_size:-0}"
|
||||
|
||||
if [ $file_count -gt 0 ]; then
|
||||
echo -e "\n Oldest files:"
|
||||
find "$UPLOAD_FOLDER" -type f -printf '%T+ %p\n' 2>/dev/null | sort | head -5
|
||||
fi
|
||||
else
|
||||
echo " Upload folder does not exist: $UPLOAD_FOLDER"
|
||||
fi
|
||||
}
|
||||
|
||||
# Function to clean old temp files
|
||||
clean_temp_files() {
|
||||
echo -e "\nCleaning temporary files older than $MAX_AGE_MINUTES minutes..."
|
||||
if [ -d "$UPLOAD_FOLDER" ]; then
|
||||
# Count files before cleanup
|
||||
before_count=$(find "$UPLOAD_FOLDER" -type f 2>/dev/null | wc -l)
|
||||
|
||||
# Remove old files
|
||||
find "$UPLOAD_FOLDER" -type f -mmin +$MAX_AGE_MINUTES -delete 2>/dev/null
|
||||
|
||||
# Count files after cleanup
|
||||
after_count=$(find "$UPLOAD_FOLDER" -type f 2>/dev/null | wc -l)
|
||||
removed=$((before_count - after_count))
|
||||
|
||||
echo " Removed $removed files"
|
||||
else
|
||||
echo " Upload folder does not exist: $UPLOAD_FOLDER"
|
||||
fi
|
||||
}
|
||||
|
||||
# Function to setup upload folder
|
||||
setup_upload_folder() {
|
||||
echo -e "\nSetting up upload folder..."
|
||||
if [ ! -d "$UPLOAD_FOLDER" ]; then
|
||||
mkdir -p "$UPLOAD_FOLDER"
|
||||
chmod 755 "$UPLOAD_FOLDER"
|
||||
echo " Created: $UPLOAD_FOLDER"
|
||||
else
|
||||
echo " Exists: $UPLOAD_FOLDER"
|
||||
fi
|
||||
}
|
||||
|
||||
# Function to monitor in real-time
|
||||
monitor_realtime() {
|
||||
echo -e "\nMonitoring temporary files (Press Ctrl+C to stop)..."
|
||||
while true; do
|
||||
clear
|
||||
echo "Talk2Me File Monitor - $(date)"
|
||||
echo "================================"
|
||||
show_temp_stats
|
||||
check_disk_usage
|
||||
sleep 5
|
||||
done
|
||||
}
|
||||
|
||||
# Main menu
|
||||
case "${1:-help}" in
|
||||
status)
|
||||
show_temp_stats
|
||||
check_disk_usage
|
||||
;;
|
||||
clean)
|
||||
clean_temp_files
|
||||
show_temp_stats
|
||||
;;
|
||||
setup)
|
||||
setup_upload_folder
|
||||
;;
|
||||
monitor)
|
||||
monitor_realtime
|
||||
;;
|
||||
all)
|
||||
setup_upload_folder
|
||||
clean_temp_files
|
||||
show_temp_stats
|
||||
check_disk_usage
|
||||
;;
|
||||
*)
|
||||
echo "Usage: $0 {status|clean|setup|monitor|all}"
|
||||
echo ""
|
||||
echo "Commands:"
|
||||
echo " status - Show current temp file statistics"
|
||||
echo " clean - Clean old temporary files"
|
||||
echo " setup - Create upload folder if needed"
|
||||
echo " monitor - Real-time monitoring"
|
||||
echo " all - Run setup, clean, and show status"
|
||||
echo ""
|
||||
echo "Environment Variables:"
|
||||
echo " UPLOAD_FOLDER - Set custom upload folder (default: /tmp/talk2me_uploads)"
|
||||
;;
|
||||
esac
|
271
manage_secrets.py
Executable file
271
manage_secrets.py
Executable file
@ -0,0 +1,271 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
Secret management CLI tool for Talk2Me
|
||||
|
||||
Usage:
|
||||
python manage_secrets.py list
|
||||
python manage_secrets.py get <key>
|
||||
python manage_secrets.py set <key> <value>
|
||||
python manage_secrets.py rotate <key>
|
||||
python manage_secrets.py delete <key>
|
||||
python manage_secrets.py check-rotation
|
||||
python manage_secrets.py verify
|
||||
python manage_secrets.py migrate
|
||||
"""
|
||||
|
||||
import sys
|
||||
import os
|
||||
import click
|
||||
import getpass
|
||||
from datetime import datetime
|
||||
from secrets_manager import get_secrets_manager, SecretsManager
|
||||
import json
|
||||
|
||||
# Initialize secrets manager
|
||||
manager = get_secrets_manager()
|
||||
|
||||
@click.group()
|
||||
def cli():
|
||||
"""Talk2Me Secrets Management Tool"""
|
||||
pass
|
||||
|
||||
@cli.command()
|
||||
def list():
|
||||
"""List all secrets (without values)"""
|
||||
secrets = manager.list_secrets()
|
||||
|
||||
if not secrets:
|
||||
click.echo("No secrets found.")
|
||||
return
|
||||
|
||||
click.echo(f"\nFound {len(secrets)} secrets:\n")
|
||||
|
||||
# Format as table
|
||||
click.echo(f"{'Key':<30} {'Created':<20} {'Last Rotated':<20} {'Has Value'}")
|
||||
click.echo("-" * 90)
|
||||
|
||||
for secret in secrets:
|
||||
created = secret['created'][:10] if secret['created'] else 'Unknown'
|
||||
rotated = secret['rotated'][:10] if secret['rotated'] else 'Never'
|
||||
has_value = '✓' if secret['has_value'] else '✗'
|
||||
|
||||
click.echo(f"{secret['key']:<30} {created:<20} {rotated:<20} {has_value}")
|
||||
|
||||
@cli.command()
|
||||
@click.argument('key')
|
||||
def get(key):
|
||||
"""Get a secret value (requires confirmation)"""
|
||||
if not click.confirm(f"Are you sure you want to display the value of '{key}'?"):
|
||||
return
|
||||
|
||||
value = manager.get(key)
|
||||
|
||||
if value is None:
|
||||
click.echo(f"Secret '{key}' not found.")
|
||||
else:
|
||||
click.echo(f"\nSecret '{key}':")
|
||||
click.echo(f"Value: {value}")
|
||||
|
||||
# Show metadata
|
||||
secrets = manager.list_secrets()
|
||||
for secret in secrets:
|
||||
if secret['key'] == key:
|
||||
if secret.get('metadata'):
|
||||
click.echo(f"Metadata: {json.dumps(secret['metadata'], indent=2)}")
|
||||
break
|
||||
|
||||
@cli.command()
|
||||
@click.argument('key')
|
||||
@click.option('--value', help='Secret value (will prompt if not provided)')
|
||||
@click.option('--metadata', help='JSON metadata')
|
||||
def set(key, value, metadata):
|
||||
"""Set a secret value"""
|
||||
if not value:
|
||||
value = getpass.getpass(f"Enter value for '{key}': ")
|
||||
confirm = getpass.getpass(f"Confirm value for '{key}': ")
|
||||
|
||||
if value != confirm:
|
||||
click.echo("Values do not match. Aborted.")
|
||||
return
|
||||
|
||||
# Parse metadata if provided
|
||||
metadata_dict = None
|
||||
if metadata:
|
||||
try:
|
||||
metadata_dict = json.loads(metadata)
|
||||
except json.JSONDecodeError:
|
||||
click.echo("Invalid JSON metadata")
|
||||
return
|
||||
|
||||
# Validate the secret if validator exists
|
||||
if not manager.validate(key, value):
|
||||
click.echo(f"Validation failed for '{key}'")
|
||||
return
|
||||
|
||||
manager.set(key, value, metadata_dict, user='cli')
|
||||
click.echo(f"Secret '{key}' set successfully.")
|
||||
|
||||
@cli.command()
|
||||
@click.argument('key')
|
||||
@click.option('--new-value', help='New secret value (will auto-generate if not provided)')
|
||||
def rotate(key):
|
||||
"""Rotate a secret"""
|
||||
try:
|
||||
if not click.confirm(f"Are you sure you want to rotate '{key}'?"):
|
||||
return
|
||||
|
||||
old_value, new_value = manager.rotate(key, new_value, user='cli')
|
||||
|
||||
click.echo(f"\nSecret '{key}' rotated successfully.")
|
||||
click.echo(f"New value: {new_value}")
|
||||
|
||||
if click.confirm("Do you want to see the old value?"):
|
||||
click.echo(f"Old value: {old_value}")
|
||||
|
||||
except KeyError:
|
||||
click.echo(f"Secret '{key}' not found.")
|
||||
except ValueError as e:
|
||||
click.echo(f"Error: {e}")
|
||||
|
||||
@cli.command()
|
||||
@click.argument('key')
|
||||
def delete(key):
|
||||
"""Delete a secret"""
|
||||
if not click.confirm(f"Are you sure you want to delete '{key}'? This cannot be undone."):
|
||||
return
|
||||
|
||||
if manager.delete(key, user='cli'):
|
||||
click.echo(f"Secret '{key}' deleted successfully.")
|
||||
else:
|
||||
click.echo(f"Secret '{key}' not found.")
|
||||
|
||||
@cli.command()
|
||||
def check_rotation():
|
||||
"""Check which secrets need rotation"""
|
||||
needs_rotation = manager.check_rotation_needed()
|
||||
|
||||
if not needs_rotation:
|
||||
click.echo("No secrets need rotation.")
|
||||
return
|
||||
|
||||
click.echo(f"\n{len(needs_rotation)} secrets need rotation:")
|
||||
for key in needs_rotation:
|
||||
click.echo(f" - {key}")
|
||||
|
||||
if click.confirm("\nDo you want to rotate all of them now?"):
|
||||
for key in needs_rotation:
|
||||
try:
|
||||
old_value, new_value = manager.rotate(key, user='cli')
|
||||
click.echo(f"✓ Rotated {key}")
|
||||
except Exception as e:
|
||||
click.echo(f"✗ Failed to rotate {key}: {e}")
|
||||
|
||||
@cli.command()
|
||||
def verify():
|
||||
"""Verify integrity of all secrets"""
|
||||
click.echo("Verifying secrets integrity...")
|
||||
|
||||
if manager.verify_integrity():
|
||||
click.echo("✓ All secrets passed integrity check")
|
||||
else:
|
||||
click.echo("✗ Integrity check failed!")
|
||||
click.echo("Some secrets may be corrupted. Check logs for details.")
|
||||
|
||||
@cli.command()
|
||||
def migrate():
|
||||
"""Migrate secrets from environment variables"""
|
||||
click.echo("Migrating secrets from environment variables...")
|
||||
|
||||
# List of known secrets to migrate
|
||||
secrets_to_migrate = [
|
||||
('TTS_API_KEY', 'TTS API Key'),
|
||||
('SECRET_KEY', 'Flask Secret Key'),
|
||||
('ADMIN_TOKEN', 'Admin Token'),
|
||||
('DATABASE_URL', 'Database URL'),
|
||||
('REDIS_URL', 'Redis URL'),
|
||||
]
|
||||
|
||||
migrated = 0
|
||||
|
||||
for env_key, description in secrets_to_migrate:
|
||||
value = os.environ.get(env_key)
|
||||
if value and value != manager.get(env_key):
|
||||
if click.confirm(f"Migrate {description} from environment?"):
|
||||
manager.set(env_key, value, {'migrated_from': 'environment'}, user='migration')
|
||||
click.echo(f"✓ Migrated {env_key}")
|
||||
migrated += 1
|
||||
|
||||
click.echo(f"\nMigrated {migrated} secrets.")
|
||||
|
||||
@cli.command()
|
||||
@click.argument('key')
|
||||
@click.argument('days', type=int)
|
||||
def schedule_rotation(key, days):
|
||||
"""Schedule automatic rotation for a secret"""
|
||||
manager.schedule_rotation(key, days)
|
||||
click.echo(f"Scheduled rotation for '{key}' every {days} days.")
|
||||
|
||||
@cli.command()
|
||||
@click.argument('key', required=False)
|
||||
@click.option('--limit', default=20, help='Number of entries to show')
|
||||
def audit(key, limit):
|
||||
"""Show audit log"""
|
||||
logs = manager.get_audit_log(key, limit)
|
||||
|
||||
if not logs:
|
||||
click.echo("No audit log entries found.")
|
||||
return
|
||||
|
||||
click.echo(f"\nShowing last {len(logs)} audit log entries:\n")
|
||||
|
||||
for entry in logs:
|
||||
timestamp = entry['timestamp'][:19] # Trim microseconds
|
||||
action = entry['action'].ljust(15)
|
||||
key_str = entry['key'].ljust(20)
|
||||
user = entry['user']
|
||||
|
||||
click.echo(f"{timestamp} | {action} | {key_str} | {user}")
|
||||
|
||||
if entry.get('details'):
|
||||
click.echo(f"{'':>20} Details: {json.dumps(entry['details'])}")
|
||||
|
||||
@cli.command()
|
||||
def init():
|
||||
"""Initialize secrets configuration"""
|
||||
click.echo("Initializing Talk2Me secrets configuration...")
|
||||
|
||||
# Check if already initialized
|
||||
if os.path.exists('.secrets.json'):
|
||||
if not click.confirm(".secrets.json already exists. Overwrite?"):
|
||||
return
|
||||
|
||||
# Generate initial secrets
|
||||
import secrets as py_secrets
|
||||
|
||||
initial_secrets = {
|
||||
'FLASK_SECRET_KEY': py_secrets.token_hex(32),
|
||||
'ADMIN_TOKEN': py_secrets.token_urlsafe(32),
|
||||
}
|
||||
|
||||
click.echo("\nGenerating initial secrets...")
|
||||
|
||||
for key, value in initial_secrets.items():
|
||||
manager.set(key, value, {'generated': True}, user='init')
|
||||
click.echo(f"✓ Generated {key}")
|
||||
|
||||
# Prompt for required secrets
|
||||
click.echo("\nPlease provide the following secrets:")
|
||||
|
||||
tts_api_key = getpass.getpass("TTS API Key (press Enter to skip): ")
|
||||
if tts_api_key:
|
||||
manager.set('TTS_API_KEY', tts_api_key, user='init')
|
||||
click.echo("✓ Set TTS_API_KEY")
|
||||
|
||||
click.echo("\nSecrets initialized successfully!")
|
||||
click.echo("\nIMPORTANT:")
|
||||
click.echo("1. Keep .secrets.json secure and never commit it to version control")
|
||||
click.echo("2. Back up your master key from .master_key")
|
||||
click.echo("3. Set appropriate file permissions (owner read/write only)")
|
||||
|
||||
if __name__ == '__main__':
|
||||
cli()
|
238
memory_manager.py
Normal file
238
memory_manager.py
Normal file
@ -0,0 +1,238 @@
|
||||
"""Memory management for Talk2Me application"""
|
||||
|
||||
import gc
|
||||
import logging
|
||||
import psutil
|
||||
import torch
|
||||
import os
|
||||
import time
|
||||
from contextlib import contextmanager
|
||||
from functools import wraps
|
||||
from dataclasses import dataclass
|
||||
from typing import Optional, Dict, Any
|
||||
import threading
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
@dataclass
|
||||
class MemoryStats:
|
||||
"""Memory statistics"""
|
||||
process_memory_mb: float
|
||||
available_memory_mb: float
|
||||
memory_percent: float
|
||||
gpu_memory_mb: float = 0.0
|
||||
gpu_memory_percent: float = 0.0
|
||||
|
||||
class MemoryManager:
|
||||
"""Manage memory usage for the application"""
|
||||
|
||||
def __init__(self, app=None, config: Optional[Dict[str, Any]] = None):
|
||||
self.app = app
|
||||
self.config = config or {}
|
||||
self.memory_threshold_mb = self.config.get('memory_threshold_mb', 4096)
|
||||
self.gpu_memory_threshold_mb = self.config.get('gpu_memory_threshold_mb', 2048)
|
||||
self.cleanup_interval = self.config.get('cleanup_interval', 30)
|
||||
self.whisper_model = None
|
||||
self._cleanup_thread = None
|
||||
self._stop_cleanup = threading.Event()
|
||||
|
||||
if app:
|
||||
self.init_app(app)
|
||||
|
||||
def init_app(self, app):
|
||||
"""Initialize with Flask app"""
|
||||
self.app = app
|
||||
app.memory_manager = self
|
||||
|
||||
# Start cleanup thread
|
||||
self._start_cleanup_thread()
|
||||
|
||||
logger.info(f"Memory manager initialized with thresholds: "
|
||||
f"Process={self.memory_threshold_mb}MB, "
|
||||
f"GPU={self.gpu_memory_threshold_mb}MB")
|
||||
|
||||
def set_whisper_model(self, model):
|
||||
"""Set reference to Whisper model for memory management"""
|
||||
self.whisper_model = model
|
||||
|
||||
def get_memory_stats(self) -> MemoryStats:
|
||||
"""Get current memory statistics"""
|
||||
process = psutil.Process()
|
||||
memory_info = process.memory_info()
|
||||
|
||||
stats = MemoryStats(
|
||||
process_memory_mb=memory_info.rss / 1024 / 1024,
|
||||
available_memory_mb=psutil.virtual_memory().available / 1024 / 1024,
|
||||
memory_percent=process.memory_percent()
|
||||
)
|
||||
|
||||
# Check GPU memory if available
|
||||
if torch.cuda.is_available():
|
||||
try:
|
||||
stats.gpu_memory_mb = torch.cuda.memory_allocated() / 1024 / 1024
|
||||
stats.gpu_memory_percent = (torch.cuda.memory_allocated() /
|
||||
torch.cuda.get_device_properties(0).total_memory * 100)
|
||||
except Exception as e:
|
||||
logger.error(f"Error getting GPU memory stats: {e}")
|
||||
|
||||
return stats
|
||||
|
||||
def check_memory_pressure(self) -> bool:
|
||||
"""Check if system is under memory pressure"""
|
||||
stats = self.get_memory_stats()
|
||||
|
||||
# Check process memory
|
||||
if stats.process_memory_mb > self.memory_threshold_mb:
|
||||
logger.warning(f"High process memory usage: {stats.process_memory_mb:.1f}MB")
|
||||
return True
|
||||
|
||||
# Check system memory
|
||||
if stats.memory_percent > 80:
|
||||
logger.warning(f"High system memory usage: {stats.memory_percent:.1f}%")
|
||||
return True
|
||||
|
||||
# Check GPU memory
|
||||
if stats.gpu_memory_mb > self.gpu_memory_threshold_mb:
|
||||
logger.warning(f"High GPU memory usage: {stats.gpu_memory_mb:.1f}MB")
|
||||
return True
|
||||
|
||||
return False
|
||||
|
||||
def cleanup_memory(self, aggressive: bool = False):
|
||||
"""Clean up memory"""
|
||||
logger.info("Starting memory cleanup...")
|
||||
|
||||
# Run garbage collection
|
||||
collected = gc.collect()
|
||||
logger.info(f"Garbage collector: collected {collected} objects")
|
||||
|
||||
# Clear GPU cache if available
|
||||
if torch.cuda.is_available():
|
||||
torch.cuda.empty_cache()
|
||||
torch.cuda.synchronize()
|
||||
logger.info("Cleared GPU cache")
|
||||
|
||||
if aggressive:
|
||||
# Force garbage collection of all generations
|
||||
for i in range(3):
|
||||
gc.collect(i)
|
||||
|
||||
# Clear Whisper model cache if needed
|
||||
if self.whisper_model and hasattr(self.whisper_model, 'clear_cache'):
|
||||
self.whisper_model.clear_cache()
|
||||
logger.info("Cleared Whisper model cache")
|
||||
|
||||
def _cleanup_worker(self):
|
||||
"""Background cleanup worker"""
|
||||
while not self._stop_cleanup.wait(self.cleanup_interval):
|
||||
try:
|
||||
if self.check_memory_pressure():
|
||||
self.cleanup_memory(aggressive=True)
|
||||
else:
|
||||
# Light cleanup
|
||||
gc.collect(0)
|
||||
|
||||
if torch.cuda.is_available():
|
||||
torch.cuda.empty_cache()
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error in memory cleanup worker: {e}")
|
||||
|
||||
def _start_cleanup_thread(self):
|
||||
"""Start background cleanup thread"""
|
||||
if self._cleanup_thread and self._cleanup_thread.is_alive():
|
||||
return
|
||||
|
||||
self._stop_cleanup.clear()
|
||||
self._cleanup_thread = threading.Thread(target=self._cleanup_worker, daemon=True)
|
||||
self._cleanup_thread.start()
|
||||
logger.info("Started memory cleanup thread")
|
||||
|
||||
def stop(self):
|
||||
"""Stop memory manager"""
|
||||
self._stop_cleanup.set()
|
||||
if self._cleanup_thread:
|
||||
self._cleanup_thread.join(timeout=5)
|
||||
|
||||
def get_metrics(self) -> Dict[str, Any]:
|
||||
"""Get memory metrics for monitoring"""
|
||||
stats = self.get_memory_stats()
|
||||
|
||||
return {
|
||||
'process_memory_mb': round(stats.process_memory_mb, 2),
|
||||
'available_memory_mb': round(stats.available_memory_mb, 2),
|
||||
'memory_percent': round(stats.memory_percent, 2),
|
||||
'gpu_memory_mb': round(stats.gpu_memory_mb, 2),
|
||||
'gpu_memory_percent': round(stats.gpu_memory_percent, 2),
|
||||
'thresholds': {
|
||||
'process_mb': self.memory_threshold_mb,
|
||||
'gpu_mb': self.gpu_memory_threshold_mb
|
||||
},
|
||||
'under_pressure': self.check_memory_pressure()
|
||||
}
|
||||
|
||||
class AudioProcessingContext:
|
||||
"""Context manager for audio processing with memory management"""
|
||||
|
||||
def __init__(self, memory_manager: MemoryManager, name: str = "audio_processing"):
|
||||
self.memory_manager = memory_manager
|
||||
self.name = name
|
||||
self.temp_files = []
|
||||
self.start_time = None
|
||||
|
||||
def __enter__(self):
|
||||
self.start_time = time.time()
|
||||
|
||||
# Check memory before processing
|
||||
if self.memory_manager and self.memory_manager.check_memory_pressure():
|
||||
logger.warning(f"Memory pressure detected before {self.name}")
|
||||
self.memory_manager.cleanup_memory()
|
||||
|
||||
return self
|
||||
|
||||
def __exit__(self, exc_type, exc_val, exc_tb):
|
||||
# Clean up temporary files
|
||||
for temp_file in self.temp_files:
|
||||
try:
|
||||
if os.path.exists(temp_file):
|
||||
os.remove(temp_file)
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to remove temp file {temp_file}: {e}")
|
||||
|
||||
# Clean up memory after processing
|
||||
if self.memory_manager:
|
||||
gc.collect()
|
||||
if torch.cuda.is_available():
|
||||
torch.cuda.empty_cache()
|
||||
|
||||
duration = time.time() - self.start_time
|
||||
logger.info(f"{self.name} completed in {duration:.2f}s")
|
||||
|
||||
def add_temp_file(self, filepath: str):
|
||||
"""Add a temporary file to be cleaned up"""
|
||||
self.temp_files.append(filepath)
|
||||
|
||||
def with_memory_management(func):
|
||||
"""Decorator to add memory management to functions"""
|
||||
@wraps(func)
|
||||
def wrapper(*args, **kwargs):
|
||||
# Get memory manager from app context
|
||||
from flask import current_app
|
||||
memory_manager = getattr(current_app, 'memory_manager', None)
|
||||
|
||||
if memory_manager:
|
||||
# Check memory before
|
||||
if memory_manager.check_memory_pressure():
|
||||
logger.warning(f"Memory pressure before {func.__name__}")
|
||||
memory_manager.cleanup_memory()
|
||||
|
||||
try:
|
||||
result = func(*args, **kwargs)
|
||||
return result
|
||||
finally:
|
||||
# Light cleanup after
|
||||
gc.collect(0)
|
||||
if torch.cuda.is_available():
|
||||
torch.cuda.empty_cache()
|
||||
|
||||
return wrapper
|
135
migrations.py
Normal file
135
migrations.py
Normal file
@ -0,0 +1,135 @@
|
||||
# Database migration scripts
|
||||
import os
|
||||
import sys
|
||||
import logging
|
||||
from flask import Flask
|
||||
from flask_migrate import Migrate, init, migrate, upgrade
|
||||
from database import db, init_db
|
||||
from config import Config
|
||||
|
||||
logging.basicConfig(level=logging.INFO)
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
def create_app():
|
||||
"""Create Flask app for migrations"""
|
||||
app = Flask(__name__)
|
||||
|
||||
# Load configuration
|
||||
config = Config()
|
||||
app.config.from_object(config)
|
||||
|
||||
# Initialize database
|
||||
init_db(app)
|
||||
|
||||
return app
|
||||
|
||||
def init_migrations():
|
||||
"""Initialize migration repository"""
|
||||
app = create_app()
|
||||
|
||||
with app.app_context():
|
||||
# Initialize Flask-Migrate
|
||||
migrate_instance = Migrate(app, db)
|
||||
|
||||
# Initialize migration repository
|
||||
try:
|
||||
init()
|
||||
logger.info("Migration repository initialized")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to initialize migrations: {e}")
|
||||
return False
|
||||
|
||||
return True
|
||||
|
||||
def create_migration(message="Auto migration"):
|
||||
"""Create a new migration"""
|
||||
app = create_app()
|
||||
|
||||
with app.app_context():
|
||||
# Initialize Flask-Migrate
|
||||
migrate_instance = Migrate(app, db)
|
||||
|
||||
try:
|
||||
migrate(message=message)
|
||||
logger.info(f"Migration created: {message}")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to create migration: {e}")
|
||||
return False
|
||||
|
||||
return True
|
||||
|
||||
def run_migrations():
|
||||
"""Run pending migrations"""
|
||||
app = create_app()
|
||||
|
||||
with app.app_context():
|
||||
# Initialize Flask-Migrate
|
||||
migrate_instance = Migrate(app, db)
|
||||
|
||||
try:
|
||||
upgrade()
|
||||
logger.info("Migrations completed successfully")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to run migrations: {e}")
|
||||
return False
|
||||
|
||||
return True
|
||||
|
||||
def create_initial_data():
|
||||
"""Create initial data if needed"""
|
||||
app = create_app()
|
||||
|
||||
with app.app_context():
|
||||
try:
|
||||
# Add any initial data here
|
||||
# For example, creating default API keys, admin users, etc.
|
||||
|
||||
db.session.commit()
|
||||
logger.info("Initial data created")
|
||||
except Exception as e:
|
||||
db.session.rollback()
|
||||
logger.error(f"Failed to create initial data: {e}")
|
||||
return False
|
||||
|
||||
return True
|
||||
|
||||
if __name__ == "__main__":
|
||||
if len(sys.argv) < 2:
|
||||
print("Usage: python migrations.py [init|create|run|seed]")
|
||||
sys.exit(1)
|
||||
|
||||
command = sys.argv[1]
|
||||
|
||||
if command == "init":
|
||||
if init_migrations():
|
||||
print("Migration repository initialized successfully")
|
||||
else:
|
||||
print("Failed to initialize migrations")
|
||||
sys.exit(1)
|
||||
|
||||
elif command == "create":
|
||||
message = sys.argv[2] if len(sys.argv) > 2 else "Auto migration"
|
||||
if create_migration(message):
|
||||
print(f"Migration created: {message}")
|
||||
else:
|
||||
print("Failed to create migration")
|
||||
sys.exit(1)
|
||||
|
||||
elif command == "run":
|
||||
if run_migrations():
|
||||
print("Migrations completed successfully")
|
||||
else:
|
||||
print("Failed to run migrations")
|
||||
sys.exit(1)
|
||||
|
||||
elif command == "seed":
|
||||
if create_initial_data():
|
||||
print("Initial data created successfully")
|
||||
else:
|
||||
print("Failed to create initial data")
|
||||
sys.exit(1)
|
||||
|
||||
else:
|
||||
print(f"Unknown command: {command}")
|
||||
print("Available commands: init, create, run, seed")
|
||||
sys.exit(1)
|
221
migrations/add_user_authentication.py
Normal file
221
migrations/add_user_authentication.py
Normal file
@ -0,0 +1,221 @@
|
||||
"""Add user authentication tables and update existing models
|
||||
|
||||
This migration:
|
||||
1. Creates user authentication tables (users, login_history, user_sessions, revoked_tokens)
|
||||
2. Updates translation and transcription tables to link to users
|
||||
3. Adds proper foreign key constraints and indexes
|
||||
"""
|
||||
|
||||
from alembic import op
|
||||
import sqlalchemy as sa
|
||||
from sqlalchemy.dialects import postgresql
|
||||
import uuid
|
||||
|
||||
# revision identifiers
|
||||
revision = 'add_user_authentication'
|
||||
down_revision = None
|
||||
branch_labels = None
|
||||
depends_on = None
|
||||
|
||||
|
||||
def upgrade():
|
||||
# Create users table
|
||||
op.create_table('users',
|
||||
sa.Column('id', postgresql.UUID(as_uuid=True), nullable=False, default=uuid.uuid4),
|
||||
sa.Column('email', sa.String(255), nullable=False),
|
||||
sa.Column('username', sa.String(100), nullable=False),
|
||||
sa.Column('password_hash', sa.String(255), nullable=False),
|
||||
sa.Column('full_name', sa.String(255), nullable=True),
|
||||
sa.Column('avatar_url', sa.String(500), nullable=True),
|
||||
sa.Column('api_key', sa.String(64), nullable=False),
|
||||
sa.Column('api_key_created_at', sa.DateTime(), nullable=False),
|
||||
sa.Column('is_active', sa.Boolean(), nullable=False, default=True),
|
||||
sa.Column('is_verified', sa.Boolean(), nullable=False, default=False),
|
||||
sa.Column('is_suspended', sa.Boolean(), nullable=False, default=False),
|
||||
sa.Column('suspension_reason', sa.Text(), nullable=True),
|
||||
sa.Column('suspended_at', sa.DateTime(), nullable=True),
|
||||
sa.Column('suspended_until', sa.DateTime(), nullable=True),
|
||||
sa.Column('role', sa.String(20), nullable=False, default='user'),
|
||||
sa.Column('permissions', postgresql.JSONB(astext_type=sa.Text()), nullable=False, default=[]),
|
||||
sa.Column('rate_limit_per_minute', sa.Integer(), nullable=False, default=30),
|
||||
sa.Column('rate_limit_per_hour', sa.Integer(), nullable=False, default=500),
|
||||
sa.Column('rate_limit_per_day', sa.Integer(), nullable=False, default=5000),
|
||||
sa.Column('total_requests', sa.Integer(), nullable=False, default=0),
|
||||
sa.Column('total_translations', sa.Integer(), nullable=False, default=0),
|
||||
sa.Column('total_transcriptions', sa.Integer(), nullable=False, default=0),
|
||||
sa.Column('total_tts_requests', sa.Integer(), nullable=False, default=0),
|
||||
sa.Column('created_at', sa.DateTime(), nullable=False),
|
||||
sa.Column('updated_at', sa.DateTime(), nullable=False),
|
||||
sa.Column('last_login_at', sa.DateTime(), nullable=True),
|
||||
sa.Column('last_active_at', sa.DateTime(), nullable=True),
|
||||
sa.Column('password_changed_at', sa.DateTime(), nullable=False),
|
||||
sa.Column('failed_login_attempts', sa.Integer(), nullable=False, default=0),
|
||||
sa.Column('locked_until', sa.DateTime(), nullable=True),
|
||||
sa.Column('settings', postgresql.JSONB(astext_type=sa.Text()), nullable=False, default={}),
|
||||
sa.PrimaryKeyConstraint('id'),
|
||||
sa.UniqueConstraint('email'),
|
||||
sa.UniqueConstraint('username'),
|
||||
sa.UniqueConstraint('api_key')
|
||||
)
|
||||
|
||||
# Create indexes on users table
|
||||
op.create_index('idx_users_email', 'users', ['email'])
|
||||
op.create_index('idx_users_username', 'users', ['username'])
|
||||
op.create_index('idx_users_api_key', 'users', ['api_key'])
|
||||
op.create_index('idx_users_email_active', 'users', ['email', 'is_active'])
|
||||
op.create_index('idx_users_role_active', 'users', ['role', 'is_active'])
|
||||
op.create_index('idx_users_created_at', 'users', ['created_at'])
|
||||
|
||||
# Create login_history table
|
||||
op.create_table('login_history',
|
||||
sa.Column('id', postgresql.UUID(as_uuid=True), nullable=False, default=uuid.uuid4),
|
||||
sa.Column('user_id', postgresql.UUID(as_uuid=True), nullable=False),
|
||||
sa.Column('login_at', sa.DateTime(), nullable=False),
|
||||
sa.Column('logout_at', sa.DateTime(), nullable=True),
|
||||
sa.Column('login_method', sa.String(20), nullable=False),
|
||||
sa.Column('success', sa.Boolean(), nullable=False),
|
||||
sa.Column('failure_reason', sa.String(255), nullable=True),
|
||||
sa.Column('session_id', sa.String(255), nullable=True),
|
||||
sa.Column('jwt_jti', sa.String(255), nullable=True),
|
||||
sa.Column('ip_address', sa.String(45), nullable=False),
|
||||
sa.Column('user_agent', sa.String(500), nullable=True),
|
||||
sa.Column('device_info', postgresql.JSONB(astext_type=sa.Text()), nullable=True),
|
||||
sa.Column('country', sa.String(2), nullable=True),
|
||||
sa.Column('city', sa.String(100), nullable=True),
|
||||
sa.Column('is_suspicious', sa.Boolean(), nullable=False, default=False),
|
||||
sa.Column('security_notes', sa.Text(), nullable=True),
|
||||
sa.ForeignKeyConstraint(['user_id'], ['users.id'], ondelete='CASCADE'),
|
||||
sa.PrimaryKeyConstraint('id')
|
||||
)
|
||||
|
||||
# Create indexes on login_history
|
||||
op.create_index('idx_login_history_user_id', 'login_history', ['user_id'])
|
||||
op.create_index('idx_login_history_user_time', 'login_history', ['user_id', 'login_at'])
|
||||
op.create_index('idx_login_history_session', 'login_history', ['session_id'])
|
||||
op.create_index('idx_login_history_jwt_jti', 'login_history', ['jwt_jti'])
|
||||
op.create_index('idx_login_history_ip', 'login_history', ['ip_address'])
|
||||
|
||||
# Create user_sessions table
|
||||
op.create_table('user_sessions',
|
||||
sa.Column('id', postgresql.UUID(as_uuid=True), nullable=False, default=uuid.uuid4),
|
||||
sa.Column('session_id', sa.String(255), nullable=False),
|
||||
sa.Column('user_id', postgresql.UUID(as_uuid=True), nullable=False),
|
||||
sa.Column('access_token_jti', sa.String(255), nullable=True),
|
||||
sa.Column('refresh_token_jti', sa.String(255), nullable=True),
|
||||
sa.Column('created_at', sa.DateTime(), nullable=False),
|
||||
sa.Column('last_active_at', sa.DateTime(), nullable=False),
|
||||
sa.Column('expires_at', sa.DateTime(), nullable=False),
|
||||
sa.Column('ip_address', sa.String(45), nullable=False),
|
||||
sa.Column('user_agent', sa.String(500), nullable=True),
|
||||
sa.Column('data', postgresql.JSONB(astext_type=sa.Text()), nullable=False, default={}),
|
||||
sa.ForeignKeyConstraint(['user_id'], ['users.id'], ondelete='CASCADE'),
|
||||
sa.PrimaryKeyConstraint('id'),
|
||||
sa.UniqueConstraint('session_id')
|
||||
)
|
||||
|
||||
# Create indexes on user_sessions
|
||||
op.create_index('idx_user_sessions_session_id', 'user_sessions', ['session_id'])
|
||||
op.create_index('idx_user_sessions_user_id', 'user_sessions', ['user_id'])
|
||||
op.create_index('idx_user_sessions_user_active', 'user_sessions', ['user_id', 'expires_at'])
|
||||
op.create_index('idx_user_sessions_token', 'user_sessions', ['access_token_jti'])
|
||||
|
||||
# Create revoked_tokens table
|
||||
op.create_table('revoked_tokens',
|
||||
sa.Column('id', postgresql.UUID(as_uuid=True), nullable=False, default=uuid.uuid4),
|
||||
sa.Column('jti', sa.String(255), nullable=False),
|
||||
sa.Column('token_type', sa.String(20), nullable=False),
|
||||
sa.Column('user_id', postgresql.UUID(as_uuid=True), nullable=True),
|
||||
sa.Column('revoked_at', sa.DateTime(), nullable=False),
|
||||
sa.Column('expires_at', sa.DateTime(), nullable=False),
|
||||
sa.Column('reason', sa.String(255), nullable=True),
|
||||
sa.PrimaryKeyConstraint('id'),
|
||||
sa.UniqueConstraint('jti')
|
||||
)
|
||||
|
||||
# Create indexes on revoked_tokens
|
||||
op.create_index('idx_revoked_tokens_jti', 'revoked_tokens', ['jti'])
|
||||
op.create_index('idx_revoked_tokens_user_id', 'revoked_tokens', ['user_id'])
|
||||
op.create_index('idx_revoked_tokens_expires', 'revoked_tokens', ['expires_at'])
|
||||
|
||||
# Update translations table to add user_id with proper foreign key
|
||||
# First, check if user_id column exists
|
||||
try:
|
||||
op.add_column('translations', sa.Column('user_id', postgresql.UUID(as_uuid=True), nullable=True))
|
||||
op.create_foreign_key('fk_translations_user_id', 'translations', 'users', ['user_id'], ['id'], ondelete='SET NULL')
|
||||
op.create_index('idx_translations_user_id', 'translations', ['user_id'])
|
||||
except:
|
||||
pass # Column might already exist
|
||||
|
||||
# Update transcriptions table to add user_id with proper foreign key
|
||||
try:
|
||||
op.add_column('transcriptions', sa.Column('user_id', postgresql.UUID(as_uuid=True), nullable=True))
|
||||
op.create_foreign_key('fk_transcriptions_user_id', 'transcriptions', 'users', ['user_id'], ['id'], ondelete='SET NULL')
|
||||
op.create_index('idx_transcriptions_user_id', 'transcriptions', ['user_id'])
|
||||
except:
|
||||
pass # Column might already exist
|
||||
|
||||
# Update user_preferences table to add proper foreign key if not exists
|
||||
try:
|
||||
op.create_foreign_key('fk_user_preferences_user_id', 'user_preferences', 'users', ['user_id'], ['id'], ondelete='CASCADE')
|
||||
except:
|
||||
pass # Foreign key might already exist
|
||||
|
||||
# Update api_keys table to add proper foreign key if not exists
|
||||
try:
|
||||
op.add_column('api_keys', sa.Column('user_id_new', postgresql.UUID(as_uuid=True), nullable=True))
|
||||
op.create_foreign_key('fk_api_keys_user_id', 'api_keys', 'users', ['user_id_new'], ['id'], ondelete='CASCADE')
|
||||
except:
|
||||
pass # Column/FK might already exist
|
||||
|
||||
# Create function for updating updated_at timestamp
|
||||
op.execute("""
|
||||
CREATE OR REPLACE FUNCTION update_updated_at_column()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
NEW.updated_at = NOW();
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ language 'plpgsql';
|
||||
""")
|
||||
|
||||
# Drop existing trigger if it exists and recreate it
|
||||
op.execute("""
|
||||
DROP TRIGGER IF EXISTS update_users_updated_at ON users;
|
||||
""")
|
||||
|
||||
# Create trigger for users table
|
||||
op.execute("""
|
||||
CREATE TRIGGER update_users_updated_at
|
||||
BEFORE UPDATE ON users
|
||||
FOR EACH ROW
|
||||
EXECUTE FUNCTION update_updated_at_column();
|
||||
""")
|
||||
|
||||
|
||||
def downgrade():
|
||||
# Drop triggers
|
||||
op.execute("DROP TRIGGER IF EXISTS update_users_updated_at ON users")
|
||||
op.execute("DROP FUNCTION IF EXISTS update_updated_at_column()")
|
||||
|
||||
# Drop foreign keys
|
||||
try:
|
||||
op.drop_constraint('fk_translations_user_id', 'translations', type_='foreignkey')
|
||||
op.drop_constraint('fk_transcriptions_user_id', 'transcriptions', type_='foreignkey')
|
||||
op.drop_constraint('fk_user_preferences_user_id', 'user_preferences', type_='foreignkey')
|
||||
op.drop_constraint('fk_api_keys_user_id', 'api_keys', type_='foreignkey')
|
||||
except:
|
||||
pass
|
||||
|
||||
# Drop columns
|
||||
try:
|
||||
op.drop_column('translations', 'user_id')
|
||||
op.drop_column('transcriptions', 'user_id')
|
||||
op.drop_column('api_keys', 'user_id_new')
|
||||
except:
|
||||
pass
|
||||
|
||||
# Drop tables
|
||||
op.drop_table('revoked_tokens')
|
||||
op.drop_table('user_sessions')
|
||||
op.drop_table('login_history')
|
||||
op.drop_table('users')
|
135
migrations/create_analytics_tables.sql
Normal file
135
migrations/create_analytics_tables.sql
Normal file
@ -0,0 +1,135 @@
|
||||
-- Create analytics tables for Talk2Me admin dashboard
|
||||
|
||||
-- Error logs table
|
||||
CREATE TABLE IF NOT EXISTS error_logs (
|
||||
id SERIAL PRIMARY KEY,
|
||||
error_type VARCHAR(100) NOT NULL,
|
||||
error_message TEXT,
|
||||
endpoint VARCHAR(255),
|
||||
method VARCHAR(10),
|
||||
status_code INTEGER,
|
||||
ip_address INET,
|
||||
user_agent TEXT,
|
||||
request_id VARCHAR(100),
|
||||
stack_trace TEXT,
|
||||
created_at TIMESTAMP WITH TIME ZONE DEFAULT CURRENT_TIMESTAMP
|
||||
);
|
||||
|
||||
-- Create indexes for error_logs
|
||||
CREATE INDEX IF NOT EXISTS idx_error_logs_created_at ON error_logs(created_at DESC);
|
||||
CREATE INDEX IF NOT EXISTS idx_error_logs_error_type ON error_logs(error_type);
|
||||
CREATE INDEX IF NOT EXISTS idx_error_logs_endpoint ON error_logs(endpoint);
|
||||
|
||||
-- Request logs table for detailed analytics
|
||||
CREATE TABLE IF NOT EXISTS request_logs (
|
||||
id SERIAL PRIMARY KEY,
|
||||
endpoint VARCHAR(255) NOT NULL,
|
||||
method VARCHAR(10) NOT NULL,
|
||||
status_code INTEGER,
|
||||
response_time_ms INTEGER,
|
||||
ip_address INET,
|
||||
user_agent TEXT,
|
||||
request_size_bytes INTEGER,
|
||||
response_size_bytes INTEGER,
|
||||
session_id VARCHAR(100),
|
||||
created_at TIMESTAMP WITH TIME ZONE DEFAULT CURRENT_TIMESTAMP
|
||||
);
|
||||
|
||||
-- Create indexes for request_logs
|
||||
CREATE INDEX IF NOT EXISTS idx_request_logs_created_at ON request_logs(created_at DESC);
|
||||
CREATE INDEX IF NOT EXISTS idx_request_logs_endpoint ON request_logs(endpoint);
|
||||
CREATE INDEX IF NOT EXISTS idx_request_logs_session_id ON request_logs(session_id);
|
||||
CREATE INDEX IF NOT EXISTS idx_request_logs_response_time ON request_logs(response_time_ms);
|
||||
|
||||
-- Translation logs table
|
||||
CREATE TABLE IF NOT EXISTS translation_logs (
|
||||
id SERIAL PRIMARY KEY,
|
||||
source_language VARCHAR(10),
|
||||
target_language VARCHAR(10),
|
||||
text_length INTEGER,
|
||||
response_time_ms INTEGER,
|
||||
success BOOLEAN DEFAULT TRUE,
|
||||
error_message TEXT,
|
||||
session_id VARCHAR(100),
|
||||
created_at TIMESTAMP WITH TIME ZONE DEFAULT CURRENT_TIMESTAMP
|
||||
);
|
||||
|
||||
-- Create indexes for translation_logs
|
||||
CREATE INDEX IF NOT EXISTS idx_translation_logs_created_at ON translation_logs(created_at DESC);
|
||||
CREATE INDEX IF NOT EXISTS idx_translation_logs_languages ON translation_logs(source_language, target_language);
|
||||
|
||||
-- Transcription logs table
|
||||
CREATE TABLE IF NOT EXISTS transcription_logs (
|
||||
id SERIAL PRIMARY KEY,
|
||||
detected_language VARCHAR(10),
|
||||
audio_duration_seconds FLOAT,
|
||||
file_size_bytes INTEGER,
|
||||
response_time_ms INTEGER,
|
||||
success BOOLEAN DEFAULT TRUE,
|
||||
error_message TEXT,
|
||||
session_id VARCHAR(100),
|
||||
created_at TIMESTAMP WITH TIME ZONE DEFAULT CURRENT_TIMESTAMP
|
||||
);
|
||||
|
||||
-- Create indexes for transcription_logs
|
||||
CREATE INDEX IF NOT EXISTS idx_transcription_logs_created_at ON transcription_logs(created_at DESC);
|
||||
CREATE INDEX IF NOT EXISTS idx_transcription_logs_language ON transcription_logs(detected_language);
|
||||
|
||||
-- TTS logs table
|
||||
CREATE TABLE IF NOT EXISTS tts_logs (
|
||||
id SERIAL PRIMARY KEY,
|
||||
language VARCHAR(10),
|
||||
text_length INTEGER,
|
||||
voice VARCHAR(50),
|
||||
response_time_ms INTEGER,
|
||||
success BOOLEAN DEFAULT TRUE,
|
||||
error_message TEXT,
|
||||
session_id VARCHAR(100),
|
||||
created_at TIMESTAMP WITH TIME ZONE DEFAULT CURRENT_TIMESTAMP
|
||||
);
|
||||
|
||||
-- Create indexes for tts_logs
|
||||
CREATE INDEX IF NOT EXISTS idx_tts_logs_created_at ON tts_logs(created_at DESC);
|
||||
CREATE INDEX IF NOT EXISTS idx_tts_logs_language ON tts_logs(language);
|
||||
|
||||
-- Daily aggregated stats table for faster queries
|
||||
CREATE TABLE IF NOT EXISTS daily_stats (
|
||||
date DATE PRIMARY KEY,
|
||||
total_requests INTEGER DEFAULT 0,
|
||||
total_translations INTEGER DEFAULT 0,
|
||||
total_transcriptions INTEGER DEFAULT 0,
|
||||
total_tts INTEGER DEFAULT 0,
|
||||
total_errors INTEGER DEFAULT 0,
|
||||
unique_sessions INTEGER DEFAULT 0,
|
||||
avg_response_time_ms FLOAT,
|
||||
created_at TIMESTAMP WITH TIME ZONE DEFAULT CURRENT_TIMESTAMP,
|
||||
updated_at TIMESTAMP WITH TIME ZONE DEFAULT CURRENT_TIMESTAMP
|
||||
);
|
||||
|
||||
-- Create function to update updated_at timestamp
|
||||
CREATE OR REPLACE FUNCTION update_updated_at_column()
|
||||
RETURNS TRIGGER AS $$
|
||||
BEGIN
|
||||
NEW.updated_at = CURRENT_TIMESTAMP;
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ language 'plpgsql';
|
||||
|
||||
-- Create trigger for daily_stats
|
||||
DROP TRIGGER IF EXISTS update_daily_stats_updated_at ON daily_stats;
|
||||
CREATE TRIGGER update_daily_stats_updated_at
|
||||
BEFORE UPDATE ON daily_stats
|
||||
FOR EACH ROW
|
||||
EXECUTE FUNCTION update_updated_at_column();
|
||||
|
||||
-- Create view for language pair statistics
|
||||
CREATE OR REPLACE VIEW language_pair_stats AS
|
||||
SELECT
|
||||
source_language || ' -> ' || target_language as language_pair,
|
||||
COUNT(*) as usage_count,
|
||||
AVG(response_time_ms) as avg_response_time,
|
||||
MAX(created_at) as last_used
|
||||
FROM translation_logs
|
||||
WHERE success = TRUE
|
||||
GROUP BY source_language, target_language
|
||||
ORDER BY usage_count DESC;
|
108
nginx.conf
Normal file
108
nginx.conf
Normal file
@ -0,0 +1,108 @@
|
||||
upstream talk2me {
|
||||
server talk2me:5005 fail_timeout=0;
|
||||
}
|
||||
|
||||
server {
|
||||
listen 80;
|
||||
server_name _;
|
||||
|
||||
# Redirect to HTTPS in production
|
||||
# return 301 https://$server_name$request_uri;
|
||||
|
||||
# Security headers
|
||||
add_header X-Content-Type-Options nosniff always;
|
||||
add_header X-Frame-Options DENY always;
|
||||
add_header X-XSS-Protection "1; mode=block" always;
|
||||
add_header Referrer-Policy "strict-origin-when-cross-origin" always;
|
||||
add_header Content-Security-Policy "default-src 'self'; script-src 'self' 'unsafe-inline'; style-src 'self' 'unsafe-inline'; img-src 'self' data:; font-src 'self'; connect-src 'self'; media-src 'self';" always;
|
||||
|
||||
# File upload limits
|
||||
client_max_body_size 50M;
|
||||
client_body_buffer_size 1M;
|
||||
client_body_timeout 120s;
|
||||
|
||||
# Timeouts
|
||||
proxy_connect_timeout 120s;
|
||||
proxy_send_timeout 120s;
|
||||
proxy_read_timeout 120s;
|
||||
send_timeout 120s;
|
||||
|
||||
# Gzip compression
|
||||
gzip on;
|
||||
gzip_vary on;
|
||||
gzip_min_length 1024;
|
||||
gzip_types text/plain text/css text/xml text/javascript application/x-javascript application/xml+rss application/json application/javascript;
|
||||
|
||||
# Static files
|
||||
location /static {
|
||||
alias /app/static;
|
||||
expires 1y;
|
||||
add_header Cache-Control "public, immutable";
|
||||
|
||||
# Gzip static files
|
||||
gzip_static on;
|
||||
}
|
||||
|
||||
# Service worker
|
||||
location /service-worker.js {
|
||||
proxy_pass http://talk2me;
|
||||
proxy_set_header Host $host;
|
||||
proxy_set_header X-Real-IP $remote_addr;
|
||||
add_header Cache-Control "no-cache, no-store, must-revalidate";
|
||||
}
|
||||
|
||||
# WebSocket support for future features
|
||||
location /ws {
|
||||
proxy_pass http://talk2me;
|
||||
proxy_http_version 1.1;
|
||||
proxy_set_header Upgrade $http_upgrade;
|
||||
proxy_set_header Connection "upgrade";
|
||||
proxy_set_header Host $host;
|
||||
proxy_set_header X-Real-IP $remote_addr;
|
||||
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
||||
proxy_set_header X-Forwarded-Proto $scheme;
|
||||
|
||||
# WebSocket timeouts
|
||||
proxy_read_timeout 86400s;
|
||||
proxy_send_timeout 86400s;
|
||||
}
|
||||
|
||||
# Health check (don't log)
|
||||
location /health {
|
||||
proxy_pass http://talk2me/health;
|
||||
access_log off;
|
||||
proxy_set_header Host $host;
|
||||
proxy_set_header X-Real-IP $remote_addr;
|
||||
}
|
||||
|
||||
# Main application
|
||||
location / {
|
||||
proxy_pass http://talk2me;
|
||||
proxy_redirect off;
|
||||
proxy_buffering off;
|
||||
|
||||
proxy_set_header Host $host;
|
||||
proxy_set_header X-Real-IP $remote_addr;
|
||||
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
||||
proxy_set_header X-Forwarded-Proto $scheme;
|
||||
proxy_set_header X-Forwarded-Host $server_name;
|
||||
|
||||
# Don't buffer responses
|
||||
proxy_buffering off;
|
||||
proxy_request_buffering off;
|
||||
}
|
||||
}
|
||||
|
||||
# HTTPS configuration (uncomment for production)
|
||||
# server {
|
||||
# listen 443 ssl http2;
|
||||
# server_name your-domain.com;
|
||||
#
|
||||
# ssl_certificate /etc/nginx/ssl/cert.pem;
|
||||
# ssl_certificate_key /etc/nginx/ssl/key.pem;
|
||||
# ssl_protocols TLSv1.2 TLSv1.3;
|
||||
# ssl_ciphers ECDHE-ECDSA-AES128-GCM-SHA256:ECDHE-RSA-AES128-GCM-SHA256:ECDHE-ECDSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-GCM-SHA384;
|
||||
# ssl_prefer_server_ciphers off;
|
||||
#
|
||||
# # Include all location blocks from above
|
||||
# }
|
1698
package-lock.json
generated
Normal file
1698
package-lock.json
generated
Normal file
File diff suppressed because it is too large
Load Diff
29
package.json
Normal file
29
package.json
Normal file
@ -0,0 +1,29 @@
|
||||
{
|
||||
"name": "talk2me",
|
||||
"version": "1.0.0",
|
||||
"description": "Real-time voice translation web application",
|
||||
"main": "index.js",
|
||||
"scripts": {
|
||||
"build": "webpack",
|
||||
"build-tsc": "tsc",
|
||||
"watch": "webpack --watch",
|
||||
"dev": "webpack --watch",
|
||||
"clean": "rm -rf static/js/dist",
|
||||
"type-check": "tsc --noEmit"
|
||||
},
|
||||
"keywords": [
|
||||
"translation",
|
||||
"voice",
|
||||
"pwa",
|
||||
"typescript"
|
||||
],
|
||||
"author": "",
|
||||
"license": "ISC",
|
||||
"devDependencies": {
|
||||
"@types/node": "^20.10.0",
|
||||
"ts-loader": "^9.5.2",
|
||||
"typescript": "^5.3.0",
|
||||
"webpack": "^5.99.9",
|
||||
"webpack-cli": "^6.0.1"
|
||||
}
|
||||
}
|
437
rate_limiter.py
Normal file
437
rate_limiter.py
Normal file
@ -0,0 +1,437 @@
|
||||
# Rate limiting implementation for Flask
|
||||
import time
|
||||
import logging
|
||||
from functools import wraps
|
||||
from collections import defaultdict, deque
|
||||
from threading import Lock
|
||||
from flask import request, jsonify, g
|
||||
from datetime import datetime, timedelta
|
||||
import hashlib
|
||||
import json
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
class RateLimiter:
|
||||
"""
|
||||
Token bucket rate limiter with sliding window and multiple strategies
|
||||
"""
|
||||
def __init__(self):
|
||||
self.buckets = defaultdict(lambda: {
|
||||
'tokens': 5, # Start with some tokens to avoid immediate burst errors
|
||||
'last_update': time.time(),
|
||||
'requests': deque(maxlen=1000) # Track last 1000 requests
|
||||
})
|
||||
self.lock = Lock()
|
||||
|
||||
# Default limits (can be overridden per endpoint)
|
||||
self.default_limits = {
|
||||
'requests_per_minute': 30,
|
||||
'requests_per_hour': 500,
|
||||
'burst_size': 10,
|
||||
'token_refresh_rate': 0.5 # tokens per second
|
||||
}
|
||||
|
||||
# Endpoint-specific limits
|
||||
self.endpoint_limits = {
|
||||
'/transcribe': {
|
||||
'requests_per_minute': 10,
|
||||
'requests_per_hour': 100,
|
||||
'burst_size': 3,
|
||||
'token_refresh_rate': 0.167, # 1 token per 6 seconds
|
||||
'max_request_size': 10 * 1024 * 1024 # 10MB
|
||||
},
|
||||
'/translate': {
|
||||
'requests_per_minute': 20,
|
||||
'requests_per_hour': 300,
|
||||
'burst_size': 5,
|
||||
'token_refresh_rate': 0.333, # 1 token per 3 seconds
|
||||
'max_request_size': 100 * 1024 # 100KB
|
||||
},
|
||||
'/translate/stream': {
|
||||
'requests_per_minute': 10,
|
||||
'requests_per_hour': 150,
|
||||
'burst_size': 3,
|
||||
'token_refresh_rate': 0.167,
|
||||
'max_request_size': 100 * 1024 # 100KB
|
||||
},
|
||||
'/speak': {
|
||||
'requests_per_minute': 15,
|
||||
'requests_per_hour': 200,
|
||||
'burst_size': 3,
|
||||
'token_refresh_rate': 0.25, # 1 token per 4 seconds
|
||||
'max_request_size': 50 * 1024 # 50KB
|
||||
}
|
||||
}
|
||||
|
||||
# IP-based blocking
|
||||
self.blocked_ips = set()
|
||||
self.temp_blocked_ips = {} # IP -> unblock_time
|
||||
|
||||
# Global limits
|
||||
self.global_limits = {
|
||||
'total_requests_per_minute': 1000,
|
||||
'total_requests_per_hour': 10000,
|
||||
'concurrent_requests': 50
|
||||
}
|
||||
self.global_requests = deque(maxlen=10000)
|
||||
self.concurrent_requests = 0
|
||||
|
||||
def get_client_id(self, request):
|
||||
"""Get unique client identifier"""
|
||||
# Use IP address + user agent for better identification
|
||||
ip = request.remote_addr or 'unknown'
|
||||
user_agent = request.headers.get('User-Agent', '')
|
||||
|
||||
# Handle proxied requests
|
||||
forwarded_for = request.headers.get('X-Forwarded-For')
|
||||
if forwarded_for:
|
||||
ip = forwarded_for.split(',')[0].strip()
|
||||
|
||||
# Create unique identifier
|
||||
identifier = f"{ip}:{user_agent}"
|
||||
return hashlib.md5(identifier.encode()).hexdigest()
|
||||
|
||||
def get_limits(self, endpoint):
|
||||
"""Get rate limits for endpoint"""
|
||||
return self.endpoint_limits.get(endpoint, self.default_limits)
|
||||
|
||||
def is_ip_blocked(self, ip):
|
||||
"""Check if IP is blocked"""
|
||||
# Check permanent blocks
|
||||
if ip in self.blocked_ips:
|
||||
return True
|
||||
|
||||
# Check temporary blocks
|
||||
if ip in self.temp_blocked_ips:
|
||||
if time.time() < self.temp_blocked_ips[ip]:
|
||||
return True
|
||||
else:
|
||||
# Unblock if time expired
|
||||
del self.temp_blocked_ips[ip]
|
||||
|
||||
return False
|
||||
|
||||
def block_ip_temporarily(self, ip, duration=3600):
|
||||
"""Block IP temporarily (default 1 hour)"""
|
||||
self.temp_blocked_ips[ip] = time.time() + duration
|
||||
logger.warning(f"IP {ip} temporarily blocked for {duration} seconds")
|
||||
|
||||
def check_global_limits(self):
|
||||
"""Check global rate limits"""
|
||||
now = time.time()
|
||||
|
||||
# Clean old requests
|
||||
minute_ago = now - 60
|
||||
hour_ago = now - 3600
|
||||
|
||||
self.global_requests = deque(
|
||||
(t for t in self.global_requests if t > hour_ago),
|
||||
maxlen=10000
|
||||
)
|
||||
|
||||
# Count requests
|
||||
requests_last_minute = sum(1 for t in self.global_requests if t > minute_ago)
|
||||
requests_last_hour = len(self.global_requests)
|
||||
|
||||
# Check limits
|
||||
if requests_last_minute >= self.global_limits['total_requests_per_minute']:
|
||||
return False, "Global rate limit exceeded (per minute)"
|
||||
|
||||
if requests_last_hour >= self.global_limits['total_requests_per_hour']:
|
||||
return False, "Global rate limit exceeded (per hour)"
|
||||
|
||||
if self.concurrent_requests >= self.global_limits['concurrent_requests']:
|
||||
return False, "Too many concurrent requests"
|
||||
|
||||
return True, None
|
||||
|
||||
def is_exempt_path(self, path):
|
||||
"""Check if path is exempt from rate limiting"""
|
||||
# Handle both path strings and endpoint names
|
||||
if path.startswith('admin.'):
|
||||
return True
|
||||
|
||||
exempt_paths = ['/admin', '/health', '/static']
|
||||
return any(path.startswith(p) for p in exempt_paths)
|
||||
|
||||
def check_rate_limit(self, client_id, endpoint, request_size=0):
|
||||
"""Check if request should be allowed"""
|
||||
# Log what endpoint we're checking
|
||||
logger.debug(f"Checking rate limit for endpoint: {endpoint}")
|
||||
|
||||
# Skip rate limiting for exempt paths before any processing
|
||||
if self.is_exempt_path(endpoint):
|
||||
logger.debug(f"Endpoint {endpoint} is exempt from rate limiting")
|
||||
return True, None, None
|
||||
|
||||
with self.lock:
|
||||
# Check global limits first
|
||||
global_ok, global_msg = self.check_global_limits()
|
||||
if not global_ok:
|
||||
return False, global_msg, None
|
||||
|
||||
# Get limits for endpoint
|
||||
limits = self.get_limits(endpoint)
|
||||
|
||||
# Check request size if applicable
|
||||
if request_size > 0 and 'max_request_size' in limits:
|
||||
if request_size > limits['max_request_size']:
|
||||
return False, "Request too large", None
|
||||
|
||||
# Get or create bucket
|
||||
bucket = self.buckets[client_id]
|
||||
now = time.time()
|
||||
|
||||
# Update tokens based on time passed
|
||||
time_passed = now - bucket['last_update']
|
||||
new_tokens = time_passed * limits['token_refresh_rate']
|
||||
bucket['tokens'] = min(
|
||||
limits['burst_size'],
|
||||
bucket['tokens'] + new_tokens
|
||||
)
|
||||
bucket['last_update'] = now
|
||||
|
||||
# Clean old requests from sliding window
|
||||
minute_ago = now - 60
|
||||
hour_ago = now - 3600
|
||||
bucket['requests'] = deque(
|
||||
(t for t in bucket['requests'] if t > hour_ago),
|
||||
maxlen=1000
|
||||
)
|
||||
|
||||
# Count requests in windows
|
||||
requests_last_minute = sum(1 for t in bucket['requests'] if t > minute_ago)
|
||||
requests_last_hour = len(bucket['requests'])
|
||||
|
||||
# Check sliding window limits
|
||||
if requests_last_minute >= limits['requests_per_minute']:
|
||||
return False, "Rate limit exceeded (per minute)", {
|
||||
'retry_after': 60,
|
||||
'limit': limits['requests_per_minute'],
|
||||
'remaining': 0,
|
||||
'reset': int(minute_ago + 60)
|
||||
}
|
||||
|
||||
if requests_last_hour >= limits['requests_per_hour']:
|
||||
return False, "Rate limit exceeded (per hour)", {
|
||||
'retry_after': 3600,
|
||||
'limit': limits['requests_per_hour'],
|
||||
'remaining': 0,
|
||||
'reset': int(hour_ago + 3600)
|
||||
}
|
||||
|
||||
# Check token bucket
|
||||
if bucket['tokens'] < 1:
|
||||
retry_after = int(1 / limits['token_refresh_rate'])
|
||||
return False, "Rate limit exceeded (burst)", {
|
||||
'retry_after': retry_after,
|
||||
'limit': limits['burst_size'],
|
||||
'remaining': 0,
|
||||
'reset': int(now + retry_after)
|
||||
}
|
||||
|
||||
# Request allowed - consume token and record
|
||||
bucket['tokens'] -= 1
|
||||
bucket['requests'].append(now)
|
||||
self.global_requests.append(now)
|
||||
|
||||
# Calculate remaining
|
||||
remaining_minute = limits['requests_per_minute'] - requests_last_minute - 1
|
||||
remaining_hour = limits['requests_per_hour'] - requests_last_hour - 1
|
||||
|
||||
return True, None, {
|
||||
'limit': limits['requests_per_minute'],
|
||||
'remaining': remaining_minute,
|
||||
'reset': int(minute_ago + 60)
|
||||
}
|
||||
|
||||
def increment_concurrent(self):
|
||||
"""Increment concurrent request counter"""
|
||||
with self.lock:
|
||||
self.concurrent_requests += 1
|
||||
|
||||
def decrement_concurrent(self):
|
||||
"""Decrement concurrent request counter"""
|
||||
with self.lock:
|
||||
self.concurrent_requests = max(0, self.concurrent_requests - 1)
|
||||
|
||||
def get_client_stats(self, client_id):
|
||||
"""Get statistics for a client"""
|
||||
with self.lock:
|
||||
if client_id not in self.buckets:
|
||||
return None
|
||||
|
||||
bucket = self.buckets[client_id]
|
||||
now = time.time()
|
||||
minute_ago = now - 60
|
||||
hour_ago = now - 3600
|
||||
|
||||
requests_last_minute = sum(1 for t in bucket['requests'] if t > minute_ago)
|
||||
requests_last_hour = len([t for t in bucket['requests'] if t > hour_ago])
|
||||
|
||||
return {
|
||||
'requests_last_minute': requests_last_minute,
|
||||
'requests_last_hour': requests_last_hour,
|
||||
'tokens_available': bucket['tokens'],
|
||||
'last_request': bucket['last_update']
|
||||
}
|
||||
|
||||
def cleanup_old_buckets(self, max_age=86400):
|
||||
"""Clean up old unused buckets (default 24 hours)"""
|
||||
with self.lock:
|
||||
now = time.time()
|
||||
to_remove = []
|
||||
|
||||
for client_id, bucket in self.buckets.items():
|
||||
if now - bucket['last_update'] > max_age:
|
||||
to_remove.append(client_id)
|
||||
|
||||
for client_id in to_remove:
|
||||
del self.buckets[client_id]
|
||||
|
||||
if to_remove:
|
||||
logger.info(f"Cleaned up {len(to_remove)} old rate limit buckets")
|
||||
|
||||
# Global rate limiter instance
|
||||
rate_limiter = RateLimiter()
|
||||
|
||||
def rate_limit(endpoint=None,
|
||||
requests_per_minute=None,
|
||||
requests_per_hour=None,
|
||||
burst_size=None,
|
||||
check_size=False):
|
||||
"""
|
||||
Rate limiting decorator for Flask routes
|
||||
|
||||
Usage:
|
||||
@app.route('/api/endpoint')
|
||||
@rate_limit(requests_per_minute=10, check_size=True)
|
||||
def endpoint():
|
||||
return jsonify({'status': 'ok'})
|
||||
"""
|
||||
def decorator(f):
|
||||
@wraps(f)
|
||||
def decorated_function(*args, **kwargs):
|
||||
# Skip rate limiting for admin routes - check both path and endpoint
|
||||
if request.path.startswith('/admin'):
|
||||
return f(*args, **kwargs)
|
||||
|
||||
# Also check endpoint name
|
||||
if request.endpoint and request.endpoint.startswith('admin.'):
|
||||
return f(*args, **kwargs)
|
||||
|
||||
# Get client ID
|
||||
client_id = rate_limiter.get_client_id(request)
|
||||
ip = request.remote_addr
|
||||
|
||||
# Check if IP is blocked
|
||||
if rate_limiter.is_ip_blocked(ip):
|
||||
return jsonify({
|
||||
'error': 'IP temporarily blocked due to excessive requests'
|
||||
}), 429
|
||||
|
||||
# Get endpoint
|
||||
endpoint_path = endpoint or request.endpoint
|
||||
|
||||
# Override default limits if specified
|
||||
if any([requests_per_minute, requests_per_hour, burst_size]):
|
||||
limits = rate_limiter.get_limits(endpoint_path).copy()
|
||||
if requests_per_minute:
|
||||
limits['requests_per_minute'] = requests_per_minute
|
||||
if requests_per_hour:
|
||||
limits['requests_per_hour'] = requests_per_hour
|
||||
if burst_size:
|
||||
limits['burst_size'] = burst_size
|
||||
rate_limiter.endpoint_limits[endpoint_path] = limits
|
||||
|
||||
# Check request size if needed
|
||||
request_size = 0
|
||||
if check_size:
|
||||
request_size = request.content_length or 0
|
||||
|
||||
# Check rate limit
|
||||
allowed, message, headers = rate_limiter.check_rate_limit(
|
||||
client_id, endpoint_path, request_size
|
||||
)
|
||||
|
||||
if not allowed:
|
||||
# Log excessive requests
|
||||
logger.warning(f"Rate limit exceeded for {client_id} on {endpoint_path}: {message}")
|
||||
|
||||
# Check if we should temporarily block this IP
|
||||
stats = rate_limiter.get_client_stats(client_id)
|
||||
if stats and stats['requests_last_minute'] > 100:
|
||||
rate_limiter.block_ip_temporarily(ip, 3600) # 1 hour block
|
||||
|
||||
response = jsonify({
|
||||
'error': message,
|
||||
'retry_after': headers.get('retry_after') if headers else 60
|
||||
})
|
||||
response.status_code = 429
|
||||
|
||||
# Add rate limit headers
|
||||
if headers:
|
||||
response.headers['X-RateLimit-Limit'] = str(headers['limit'])
|
||||
response.headers['X-RateLimit-Remaining'] = str(headers['remaining'])
|
||||
response.headers['X-RateLimit-Reset'] = str(headers['reset'])
|
||||
response.headers['Retry-After'] = str(headers['retry_after'])
|
||||
|
||||
return response
|
||||
|
||||
# Track concurrent requests
|
||||
rate_limiter.increment_concurrent()
|
||||
|
||||
try:
|
||||
# Add rate limit info to response
|
||||
g.rate_limit_headers = headers
|
||||
response = f(*args, **kwargs)
|
||||
|
||||
# Add headers to successful response
|
||||
if headers and hasattr(response, 'headers'):
|
||||
response.headers['X-RateLimit-Limit'] = str(headers['limit'])
|
||||
response.headers['X-RateLimit-Remaining'] = str(headers['remaining'])
|
||||
response.headers['X-RateLimit-Reset'] = str(headers['reset'])
|
||||
|
||||
return response
|
||||
finally:
|
||||
rate_limiter.decrement_concurrent()
|
||||
|
||||
return decorated_function
|
||||
return decorator
|
||||
|
||||
def cleanup_rate_limiter():
|
||||
"""Cleanup function to be called periodically"""
|
||||
rate_limiter.cleanup_old_buckets()
|
||||
|
||||
# IP whitelist/blacklist management
|
||||
class IPFilter:
|
||||
def __init__(self):
|
||||
self.whitelist = set()
|
||||
self.blacklist = set()
|
||||
|
||||
def add_to_whitelist(self, ip):
|
||||
self.whitelist.add(ip)
|
||||
self.blacklist.discard(ip)
|
||||
|
||||
def add_to_blacklist(self, ip):
|
||||
self.blacklist.add(ip)
|
||||
self.whitelist.discard(ip)
|
||||
|
||||
def is_allowed(self, ip):
|
||||
if ip in self.blacklist:
|
||||
return False
|
||||
if self.whitelist and ip not in self.whitelist:
|
||||
return False
|
||||
return True
|
||||
|
||||
ip_filter = IPFilter()
|
||||
|
||||
def ip_filter_check():
|
||||
"""Middleware to check IP filtering"""
|
||||
# Skip IP filtering for admin routes
|
||||
if request.path.startswith('/admin'):
|
||||
return None
|
||||
|
||||
ip = request.remote_addr
|
||||
if not ip_filter.is_allowed(ip):
|
||||
return jsonify({'error': 'Access denied'}), 403
|
446
redis_manager.py
Normal file
446
redis_manager.py
Normal file
@ -0,0 +1,446 @@
|
||||
# Redis connection and caching management
|
||||
import redis
|
||||
import json
|
||||
import pickle
|
||||
import logging
|
||||
from typing import Optional, Any, Dict, List, Union
|
||||
from datetime import timedelta
|
||||
from functools import wraps
|
||||
import hashlib
|
||||
import time
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
class RedisManager:
|
||||
"""Manage Redis connections and operations"""
|
||||
|
||||
def __init__(self, app=None, config=None):
|
||||
self.redis_client = None
|
||||
self.config = config or {}
|
||||
self.key_prefix = self.config.get('key_prefix', 'talk2me:')
|
||||
|
||||
if app:
|
||||
self.init_app(app)
|
||||
|
||||
def init_app(self, app):
|
||||
"""Initialize Redis with Flask app"""
|
||||
# Get Redis configuration
|
||||
redis_url = app.config.get('REDIS_URL', 'redis://localhost:6379/0')
|
||||
|
||||
# Parse connection options
|
||||
decode_responses = app.config.get('REDIS_DECODE_RESPONSES', False)
|
||||
max_connections = app.config.get('REDIS_MAX_CONNECTIONS', 50)
|
||||
socket_timeout = app.config.get('REDIS_SOCKET_TIMEOUT', 5)
|
||||
|
||||
# Create connection pool
|
||||
pool = redis.ConnectionPool.from_url(
|
||||
redis_url,
|
||||
max_connections=max_connections,
|
||||
socket_timeout=socket_timeout,
|
||||
decode_responses=decode_responses
|
||||
)
|
||||
|
||||
self.redis_client = redis.Redis(connection_pool=pool)
|
||||
|
||||
# Test connection
|
||||
try:
|
||||
self.redis_client.ping()
|
||||
logger.info(f"Redis connected successfully to {redis_url}")
|
||||
except redis.ConnectionError as e:
|
||||
logger.error(f"Failed to connect to Redis: {e}")
|
||||
raise
|
||||
|
||||
# Store reference in app
|
||||
app.redis = self
|
||||
|
||||
def _make_key(self, key: str) -> str:
|
||||
"""Create a prefixed key"""
|
||||
return f"{self.key_prefix}{key}"
|
||||
|
||||
# Basic operations
|
||||
def get(self, key: str, default=None) -> Any:
|
||||
"""Get value from Redis"""
|
||||
try:
|
||||
value = self.redis_client.get(self._make_key(key))
|
||||
if value is None:
|
||||
return default
|
||||
|
||||
# Try to deserialize JSON first
|
||||
try:
|
||||
return json.loads(value)
|
||||
except (json.JSONDecodeError, TypeError):
|
||||
# Try pickle for complex objects
|
||||
try:
|
||||
return pickle.loads(value)
|
||||
except:
|
||||
# Return as string
|
||||
return value.decode('utf-8') if isinstance(value, bytes) else value
|
||||
except Exception as e:
|
||||
logger.error(f"Redis get error for key {key}: {e}")
|
||||
return default
|
||||
|
||||
def set(self, key: str, value: Any, expire: Optional[int] = None) -> bool:
|
||||
"""Set value in Redis with optional expiration"""
|
||||
try:
|
||||
# Serialize value
|
||||
if isinstance(value, (str, int, float)):
|
||||
serialized = str(value)
|
||||
elif isinstance(value, (dict, list)):
|
||||
serialized = json.dumps(value)
|
||||
else:
|
||||
serialized = pickle.dumps(value)
|
||||
|
||||
return self.redis_client.set(
|
||||
self._make_key(key),
|
||||
serialized,
|
||||
ex=expire
|
||||
)
|
||||
except Exception as e:
|
||||
logger.error(f"Redis set error for key {key}: {e}")
|
||||
return False
|
||||
|
||||
def delete(self, *keys) -> int:
|
||||
"""Delete one or more keys"""
|
||||
try:
|
||||
prefixed_keys = [self._make_key(k) for k in keys]
|
||||
return self.redis_client.delete(*prefixed_keys)
|
||||
except Exception as e:
|
||||
logger.error(f"Redis delete error: {e}")
|
||||
return 0
|
||||
|
||||
def exists(self, key: str) -> bool:
|
||||
"""Check if key exists"""
|
||||
try:
|
||||
return bool(self.redis_client.exists(self._make_key(key)))
|
||||
except Exception as e:
|
||||
logger.error(f"Redis exists error for key {key}: {e}")
|
||||
return False
|
||||
|
||||
def expire(self, key: str, seconds: int) -> bool:
|
||||
"""Set expiration on a key"""
|
||||
try:
|
||||
return bool(self.redis_client.expire(self._make_key(key), seconds))
|
||||
except Exception as e:
|
||||
logger.error(f"Redis expire error for key {key}: {e}")
|
||||
return False
|
||||
|
||||
# Hash operations for session/rate limiting
|
||||
def hget(self, name: str, key: str, default=None) -> Any:
|
||||
"""Get value from hash"""
|
||||
try:
|
||||
value = self.redis_client.hget(self._make_key(name), key)
|
||||
if value is None:
|
||||
return default
|
||||
|
||||
try:
|
||||
return json.loads(value)
|
||||
except (json.JSONDecodeError, TypeError):
|
||||
return value.decode('utf-8') if isinstance(value, bytes) else value
|
||||
except Exception as e:
|
||||
logger.error(f"Redis hget error for {name}:{key}: {e}")
|
||||
return default
|
||||
|
||||
def hset(self, name: str, key: str, value: Any) -> bool:
|
||||
"""Set value in hash"""
|
||||
try:
|
||||
if isinstance(value, (dict, list)):
|
||||
value = json.dumps(value)
|
||||
return bool(self.redis_client.hset(self._make_key(name), key, value))
|
||||
except Exception as e:
|
||||
logger.error(f"Redis hset error for {name}:{key}: {e}")
|
||||
return False
|
||||
|
||||
def hgetall(self, name: str) -> Dict[str, Any]:
|
||||
"""Get all values from hash"""
|
||||
try:
|
||||
data = self.redis_client.hgetall(self._make_key(name))
|
||||
result = {}
|
||||
for k, v in data.items():
|
||||
key = k.decode('utf-8') if isinstance(k, bytes) else k
|
||||
try:
|
||||
result[key] = json.loads(v)
|
||||
except:
|
||||
result[key] = v.decode('utf-8') if isinstance(v, bytes) else v
|
||||
return result
|
||||
except Exception as e:
|
||||
logger.error(f"Redis hgetall error for {name}: {e}")
|
||||
return {}
|
||||
|
||||
def hdel(self, name: str, *keys) -> int:
|
||||
"""Delete fields from hash"""
|
||||
try:
|
||||
return self.redis_client.hdel(self._make_key(name), *keys)
|
||||
except Exception as e:
|
||||
logger.error(f"Redis hdel error for {name}: {e}")
|
||||
return 0
|
||||
|
||||
# List operations for queues
|
||||
def lpush(self, key: str, *values) -> int:
|
||||
"""Push values to the left of list"""
|
||||
try:
|
||||
serialized = []
|
||||
for v in values:
|
||||
if isinstance(v, (dict, list)):
|
||||
serialized.append(json.dumps(v))
|
||||
else:
|
||||
serialized.append(v)
|
||||
return self.redis_client.lpush(self._make_key(key), *serialized)
|
||||
except Exception as e:
|
||||
logger.error(f"Redis lpush error for {key}: {e}")
|
||||
return 0
|
||||
|
||||
def rpop(self, key: str, default=None) -> Any:
|
||||
"""Pop value from the right of list"""
|
||||
try:
|
||||
value = self.redis_client.rpop(self._make_key(key))
|
||||
if value is None:
|
||||
return default
|
||||
|
||||
try:
|
||||
return json.loads(value)
|
||||
except:
|
||||
return value.decode('utf-8') if isinstance(value, bytes) else value
|
||||
except Exception as e:
|
||||
logger.error(f"Redis rpop error for {key}: {e}")
|
||||
return default
|
||||
|
||||
def llen(self, key: str) -> int:
|
||||
"""Get length of list"""
|
||||
try:
|
||||
return self.redis_client.llen(self._make_key(key))
|
||||
except Exception as e:
|
||||
logger.error(f"Redis llen error for {key}: {e}")
|
||||
return 0
|
||||
|
||||
# Set operations for unique tracking
|
||||
def sadd(self, key: str, *values) -> int:
|
||||
"""Add members to set"""
|
||||
try:
|
||||
return self.redis_client.sadd(self._make_key(key), *values)
|
||||
except Exception as e:
|
||||
logger.error(f"Redis sadd error for {key}: {e}")
|
||||
return 0
|
||||
|
||||
def srem(self, key: str, *values) -> int:
|
||||
"""Remove members from set"""
|
||||
try:
|
||||
return self.redis_client.srem(self._make_key(key), *values)
|
||||
except Exception as e:
|
||||
logger.error(f"Redis srem error for {key}: {e}")
|
||||
return 0
|
||||
|
||||
def sismember(self, key: str, value: Any) -> bool:
|
||||
"""Check if value is member of set"""
|
||||
try:
|
||||
return bool(self.redis_client.sismember(self._make_key(key), value))
|
||||
except Exception as e:
|
||||
logger.error(f"Redis sismember error for {key}: {e}")
|
||||
return False
|
||||
|
||||
def scard(self, key: str) -> int:
|
||||
"""Get number of members in set"""
|
||||
try:
|
||||
return self.redis_client.scard(self._make_key(key))
|
||||
except Exception as e:
|
||||
logger.error(f"Redis scard error for {key}: {e}")
|
||||
return 0
|
||||
|
||||
def smembers(self, key: str) -> set:
|
||||
"""Get all members of set"""
|
||||
try:
|
||||
members = self.redis_client.smembers(self._make_key(key))
|
||||
return {m.decode('utf-8') if isinstance(m, bytes) else m for m in members}
|
||||
except Exception as e:
|
||||
logger.error(f"Redis smembers error for {key}: {e}")
|
||||
return set()
|
||||
|
||||
# Atomic counters
|
||||
def incr(self, key: str, amount: int = 1) -> int:
|
||||
"""Increment counter"""
|
||||
try:
|
||||
return self.redis_client.incr(self._make_key(key), amount)
|
||||
except Exception as e:
|
||||
logger.error(f"Redis incr error for {key}: {e}")
|
||||
return 0
|
||||
|
||||
def decr(self, key: str, amount: int = 1) -> int:
|
||||
"""Decrement counter"""
|
||||
try:
|
||||
return self.redis_client.decr(self._make_key(key), amount)
|
||||
except Exception as e:
|
||||
logger.error(f"Redis decr error for {key}: {e}")
|
||||
return 0
|
||||
|
||||
# Transaction support
|
||||
def pipeline(self):
|
||||
"""Create a pipeline for atomic operations"""
|
||||
return self.redis_client.pipeline()
|
||||
|
||||
# Pub/Sub support
|
||||
def publish(self, channel: str, message: Any) -> int:
|
||||
"""Publish message to channel"""
|
||||
try:
|
||||
if isinstance(message, (dict, list)):
|
||||
message = json.dumps(message)
|
||||
return self.redis_client.publish(self._make_key(channel), message)
|
||||
except Exception as e:
|
||||
logger.error(f"Redis publish error for {channel}: {e}")
|
||||
return 0
|
||||
|
||||
def subscribe(self, *channels):
|
||||
"""Subscribe to channels"""
|
||||
pubsub = self.redis_client.pubsub()
|
||||
prefixed_channels = [self._make_key(c) for c in channels]
|
||||
pubsub.subscribe(*prefixed_channels)
|
||||
return pubsub
|
||||
|
||||
# Cache helpers
|
||||
def cache_translation(self, source_text: str, source_lang: str,
|
||||
target_lang: str, translation: str,
|
||||
expire_hours: int = 24) -> bool:
|
||||
"""Cache a translation"""
|
||||
key = self._translation_key(source_text, source_lang, target_lang)
|
||||
data = {
|
||||
'translation': translation,
|
||||
'timestamp': time.time(),
|
||||
'hits': 0
|
||||
}
|
||||
return self.set(key, data, expire=expire_hours * 3600)
|
||||
|
||||
def get_cached_translation(self, source_text: str, source_lang: str,
|
||||
target_lang: str) -> Optional[str]:
|
||||
"""Get cached translation and increment hit counter"""
|
||||
key = self._translation_key(source_text, source_lang, target_lang)
|
||||
data = self.get(key)
|
||||
|
||||
if data and isinstance(data, dict):
|
||||
# Increment hit counter
|
||||
data['hits'] = data.get('hits', 0) + 1
|
||||
self.set(key, data)
|
||||
return data.get('translation')
|
||||
|
||||
return None
|
||||
|
||||
def _translation_key(self, text: str, source_lang: str, target_lang: str) -> str:
|
||||
"""Generate cache key for translation"""
|
||||
# Create a hash of the text to handle long texts
|
||||
text_hash = hashlib.md5(text.encode()).hexdigest()
|
||||
return f"translation:{source_lang}:{target_lang}:{text_hash}"
|
||||
|
||||
# Session management
|
||||
def save_session(self, session_id: str, data: Dict[str, Any],
|
||||
expire_seconds: int = 3600) -> bool:
|
||||
"""Save session data"""
|
||||
key = f"session:{session_id}"
|
||||
return self.set(key, data, expire=expire_seconds)
|
||||
|
||||
def get_session(self, session_id: str) -> Optional[Dict[str, Any]]:
|
||||
"""Get session data"""
|
||||
key = f"session:{session_id}"
|
||||
return self.get(key)
|
||||
|
||||
def delete_session(self, session_id: str) -> bool:
|
||||
"""Delete session data"""
|
||||
key = f"session:{session_id}"
|
||||
return bool(self.delete(key))
|
||||
|
||||
def extend_session(self, session_id: str, expire_seconds: int = 3600) -> bool:
|
||||
"""Extend session expiration"""
|
||||
key = f"session:{session_id}"
|
||||
return self.expire(key, expire_seconds)
|
||||
|
||||
# Rate limiting
|
||||
def check_rate_limit(self, identifier: str, limit: int,
|
||||
window_seconds: int) -> tuple[bool, int]:
|
||||
"""Check rate limit using sliding window"""
|
||||
key = f"rate_limit:{identifier}:{window_seconds}"
|
||||
now = time.time()
|
||||
window_start = now - window_seconds
|
||||
|
||||
pipe = self.pipeline()
|
||||
|
||||
# Remove old entries
|
||||
pipe.zremrangebyscore(self._make_key(key), 0, window_start)
|
||||
|
||||
# Count current entries
|
||||
pipe.zcard(self._make_key(key))
|
||||
|
||||
# Add current request
|
||||
pipe.zadd(self._make_key(key), {str(now): now})
|
||||
|
||||
# Set expiration
|
||||
pipe.expire(self._make_key(key), window_seconds + 1)
|
||||
|
||||
results = pipe.execute()
|
||||
current_count = results[1]
|
||||
|
||||
if current_count >= limit:
|
||||
return False, limit - current_count
|
||||
|
||||
return True, limit - current_count - 1
|
||||
|
||||
# Cleanup
|
||||
def cleanup_expired_keys(self, pattern: str = "*") -> int:
|
||||
"""Clean up expired keys matching pattern"""
|
||||
try:
|
||||
cursor = 0
|
||||
deleted = 0
|
||||
|
||||
while True:
|
||||
cursor, keys = self.redis_client.scan(
|
||||
cursor,
|
||||
match=self._make_key(pattern),
|
||||
count=100
|
||||
)
|
||||
|
||||
for key in keys:
|
||||
ttl = self.redis_client.ttl(key)
|
||||
if ttl == -2: # Key doesn't exist
|
||||
continue
|
||||
elif ttl == -1: # Key exists but no TTL
|
||||
# Set a default TTL of 24 hours for keys without expiration
|
||||
self.redis_client.expire(key, 86400)
|
||||
|
||||
if cursor == 0:
|
||||
break
|
||||
|
||||
return deleted
|
||||
except Exception as e:
|
||||
logger.error(f"Redis cleanup error: {e}")
|
||||
return 0
|
||||
|
||||
|
||||
# Cache decorator
|
||||
def redis_cache(expire_seconds: int = 300, key_prefix: str = ""):
|
||||
"""Decorator to cache function results in Redis"""
|
||||
def decorator(func):
|
||||
@wraps(func)
|
||||
def wrapper(*args, **kwargs):
|
||||
# Get Redis instance from app context
|
||||
from flask import current_app
|
||||
redis_manager = getattr(current_app, 'redis', None)
|
||||
|
||||
if not redis_manager:
|
||||
# No Redis, execute function normally
|
||||
return func(*args, **kwargs)
|
||||
|
||||
# Generate cache key
|
||||
cache_key = f"{key_prefix}:{func.__name__}:"
|
||||
cache_key += hashlib.md5(
|
||||
f"{args}:{kwargs}".encode()
|
||||
).hexdigest()
|
||||
|
||||
# Try to get from cache
|
||||
cached = redis_manager.get(cache_key)
|
||||
if cached is not None:
|
||||
logger.debug(f"Cache hit for {func.__name__}")
|
||||
return cached
|
||||
|
||||
# Execute function and cache result
|
||||
result = func(*args, **kwargs)
|
||||
redis_manager.set(cache_key, result, expire=expire_seconds)
|
||||
|
||||
return result
|
||||
|
||||
return wrapper
|
||||
return decorator
|
365
redis_rate_limiter.py
Normal file
365
redis_rate_limiter.py
Normal file
@ -0,0 +1,365 @@
|
||||
# Redis-based rate limiting implementation
|
||||
import time
|
||||
import logging
|
||||
from functools import wraps
|
||||
from flask import request, jsonify, g
|
||||
import hashlib
|
||||
from typing import Optional, Dict, Tuple
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
class RedisRateLimiter:
|
||||
"""Token bucket rate limiter using Redis for distributed rate limiting"""
|
||||
|
||||
def __init__(self, redis_manager):
|
||||
self.redis = redis_manager
|
||||
|
||||
# Default limits (can be overridden per endpoint)
|
||||
self.default_limits = {
|
||||
'requests_per_minute': 30,
|
||||
'requests_per_hour': 500,
|
||||
'burst_size': 10,
|
||||
'token_refresh_rate': 0.5 # tokens per second
|
||||
}
|
||||
|
||||
# Endpoint-specific limits
|
||||
self.endpoint_limits = {
|
||||
'/transcribe': {
|
||||
'requests_per_minute': 10,
|
||||
'requests_per_hour': 100,
|
||||
'burst_size': 3,
|
||||
'token_refresh_rate': 0.167,
|
||||
'max_request_size': 10 * 1024 * 1024 # 10MB
|
||||
},
|
||||
'/translate': {
|
||||
'requests_per_minute': 20,
|
||||
'requests_per_hour': 300,
|
||||
'burst_size': 5,
|
||||
'token_refresh_rate': 0.333,
|
||||
'max_request_size': 100 * 1024 # 100KB
|
||||
},
|
||||
'/translate/stream': {
|
||||
'requests_per_minute': 10,
|
||||
'requests_per_hour': 150,
|
||||
'burst_size': 3,
|
||||
'token_refresh_rate': 0.167,
|
||||
'max_request_size': 100 * 1024 # 100KB
|
||||
},
|
||||
'/speak': {
|
||||
'requests_per_minute': 15,
|
||||
'requests_per_hour': 200,
|
||||
'burst_size': 3,
|
||||
'token_refresh_rate': 0.25,
|
||||
'max_request_size': 50 * 1024 # 50KB
|
||||
}
|
||||
}
|
||||
|
||||
# Global limits
|
||||
self.global_limits = {
|
||||
'total_requests_per_minute': 1000,
|
||||
'total_requests_per_hour': 10000,
|
||||
'concurrent_requests': 50
|
||||
}
|
||||
|
||||
def get_client_id(self, req) -> str:
|
||||
"""Get unique client identifier"""
|
||||
ip = req.remote_addr or 'unknown'
|
||||
user_agent = req.headers.get('User-Agent', '')
|
||||
|
||||
# Handle proxied requests
|
||||
forwarded_for = req.headers.get('X-Forwarded-For')
|
||||
if forwarded_for:
|
||||
ip = forwarded_for.split(',')[0].strip()
|
||||
|
||||
# Create unique identifier
|
||||
identifier = f"{ip}:{user_agent}"
|
||||
return hashlib.md5(identifier.encode()).hexdigest()
|
||||
|
||||
def get_limits(self, endpoint: str) -> Dict:
|
||||
"""Get rate limits for endpoint"""
|
||||
return self.endpoint_limits.get(endpoint, self.default_limits)
|
||||
|
||||
def is_ip_blocked(self, ip: str) -> bool:
|
||||
"""Check if IP is blocked"""
|
||||
# Check permanent blocks
|
||||
if self.redis.sismember('blocked_ips:permanent', ip):
|
||||
return True
|
||||
|
||||
# Check temporary blocks
|
||||
block_key = f'blocked_ip:{ip}'
|
||||
if self.redis.exists(block_key):
|
||||
return True
|
||||
|
||||
return False
|
||||
|
||||
def block_ip_temporarily(self, ip: str, duration: int = 3600):
|
||||
"""Block IP temporarily"""
|
||||
block_key = f'blocked_ip:{ip}'
|
||||
self.redis.set(block_key, 1, expire=duration)
|
||||
logger.warning(f"IP {ip} temporarily blocked for {duration} seconds")
|
||||
|
||||
def check_global_limits(self) -> Tuple[bool, Optional[str]]:
|
||||
"""Check global rate limits"""
|
||||
now = time.time()
|
||||
|
||||
# Check requests per minute
|
||||
minute_key = 'global:requests:minute'
|
||||
allowed, remaining = self.redis.check_rate_limit(
|
||||
minute_key,
|
||||
self.global_limits['total_requests_per_minute'],
|
||||
60
|
||||
)
|
||||
if not allowed:
|
||||
return False, "Global rate limit exceeded (per minute)"
|
||||
|
||||
# Check requests per hour
|
||||
hour_key = 'global:requests:hour'
|
||||
allowed, remaining = self.redis.check_rate_limit(
|
||||
hour_key,
|
||||
self.global_limits['total_requests_per_hour'],
|
||||
3600
|
||||
)
|
||||
if not allowed:
|
||||
return False, "Global rate limit exceeded (per hour)"
|
||||
|
||||
# Check concurrent requests
|
||||
concurrent_key = 'global:concurrent'
|
||||
current_concurrent = self.redis.get(concurrent_key, 0)
|
||||
if current_concurrent >= self.global_limits['concurrent_requests']:
|
||||
return False, "Too many concurrent requests"
|
||||
|
||||
return True, None
|
||||
|
||||
def check_rate_limit(self, client_id: str, endpoint: str,
|
||||
request_size: int = 0) -> Tuple[bool, Optional[str], Optional[Dict]]:
|
||||
"""Check if request should be allowed"""
|
||||
# Check global limits first
|
||||
global_ok, global_msg = self.check_global_limits()
|
||||
if not global_ok:
|
||||
return False, global_msg, None
|
||||
|
||||
# Get limits for endpoint
|
||||
limits = self.get_limits(endpoint)
|
||||
|
||||
# Check request size if applicable
|
||||
if request_size > 0 and 'max_request_size' in limits:
|
||||
if request_size > limits['max_request_size']:
|
||||
return False, "Request too large", None
|
||||
|
||||
# Token bucket implementation using Redis
|
||||
bucket_key = f'bucket:{client_id}:{endpoint}'
|
||||
now = time.time()
|
||||
|
||||
# Get current bucket state
|
||||
bucket_data = self.redis.hgetall(bucket_key)
|
||||
|
||||
# Initialize bucket if empty
|
||||
if not bucket_data:
|
||||
bucket_data = {
|
||||
'tokens': limits['burst_size'],
|
||||
'last_update': now
|
||||
}
|
||||
else:
|
||||
# Update tokens based on time passed
|
||||
last_update = float(bucket_data.get('last_update', now))
|
||||
time_passed = now - last_update
|
||||
new_tokens = time_passed * limits['token_refresh_rate']
|
||||
|
||||
current_tokens = float(bucket_data.get('tokens', 0))
|
||||
bucket_data['tokens'] = min(
|
||||
limits['burst_size'],
|
||||
current_tokens + new_tokens
|
||||
)
|
||||
bucket_data['last_update'] = now
|
||||
|
||||
# Check sliding window limits
|
||||
minute_allowed, minute_remaining = self.redis.check_rate_limit(
|
||||
f'window:{client_id}:{endpoint}:minute',
|
||||
limits['requests_per_minute'],
|
||||
60
|
||||
)
|
||||
|
||||
if not minute_allowed:
|
||||
return False, "Rate limit exceeded (per minute)", {
|
||||
'retry_after': 60,
|
||||
'limit': limits['requests_per_minute'],
|
||||
'remaining': 0,
|
||||
'reset': int(now + 60)
|
||||
}
|
||||
|
||||
hour_allowed, hour_remaining = self.redis.check_rate_limit(
|
||||
f'window:{client_id}:{endpoint}:hour',
|
||||
limits['requests_per_hour'],
|
||||
3600
|
||||
)
|
||||
|
||||
if not hour_allowed:
|
||||
return False, "Rate limit exceeded (per hour)", {
|
||||
'retry_after': 3600,
|
||||
'limit': limits['requests_per_hour'],
|
||||
'remaining': 0,
|
||||
'reset': int(now + 3600)
|
||||
}
|
||||
|
||||
# Check token bucket
|
||||
if float(bucket_data['tokens']) < 1:
|
||||
retry_after = int(1 / limits['token_refresh_rate'])
|
||||
return False, "Rate limit exceeded (burst)", {
|
||||
'retry_after': retry_after,
|
||||
'limit': limits['burst_size'],
|
||||
'remaining': 0,
|
||||
'reset': int(now + retry_after)
|
||||
}
|
||||
|
||||
# Request allowed - update bucket
|
||||
bucket_data['tokens'] = float(bucket_data['tokens']) - 1
|
||||
|
||||
# Save bucket state
|
||||
self.redis.hset(bucket_key, 'tokens', bucket_data['tokens'])
|
||||
self.redis.hset(bucket_key, 'last_update', bucket_data['last_update'])
|
||||
self.redis.expire(bucket_key, 86400) # Expire after 24 hours
|
||||
|
||||
return True, None, {
|
||||
'limit': limits['requests_per_minute'],
|
||||
'remaining': minute_remaining,
|
||||
'reset': int(now + 60)
|
||||
}
|
||||
|
||||
def increment_concurrent(self):
|
||||
"""Increment concurrent request counter"""
|
||||
self.redis.incr('global:concurrent')
|
||||
|
||||
def decrement_concurrent(self):
|
||||
"""Decrement concurrent request counter"""
|
||||
self.redis.decr('global:concurrent')
|
||||
|
||||
def get_client_stats(self, client_id: str) -> Optional[Dict]:
|
||||
"""Get statistics for a client"""
|
||||
stats = {
|
||||
'requests_last_minute': 0,
|
||||
'requests_last_hour': 0,
|
||||
'buckets': {}
|
||||
}
|
||||
|
||||
# Get request counts from all endpoints
|
||||
for endpoint in self.endpoint_limits.keys():
|
||||
minute_key = f'window:{client_id}:{endpoint}:minute'
|
||||
hour_key = f'window:{client_id}:{endpoint}:hour'
|
||||
|
||||
# This is approximate since we're using sliding windows
|
||||
minute_count = self.redis.scard(minute_key)
|
||||
hour_count = self.redis.scard(hour_key)
|
||||
|
||||
stats['requests_last_minute'] += minute_count
|
||||
stats['requests_last_hour'] += hour_count
|
||||
|
||||
# Get bucket info
|
||||
bucket_key = f'bucket:{client_id}:{endpoint}'
|
||||
bucket_data = self.redis.hgetall(bucket_key)
|
||||
if bucket_data:
|
||||
stats['buckets'][endpoint] = {
|
||||
'tokens': float(bucket_data.get('tokens', 0)),
|
||||
'last_update': float(bucket_data.get('last_update', 0))
|
||||
}
|
||||
|
||||
return stats
|
||||
|
||||
|
||||
def rate_limit(endpoint=None,
|
||||
requests_per_minute=None,
|
||||
requests_per_hour=None,
|
||||
burst_size=None,
|
||||
check_size=False):
|
||||
"""
|
||||
Rate limiting decorator for Flask routes using Redis
|
||||
"""
|
||||
def decorator(f):
|
||||
@wraps(f)
|
||||
def decorated_function(*args, **kwargs):
|
||||
# Get Redis rate limiter from app
|
||||
from flask import current_app
|
||||
|
||||
if not hasattr(current_app, 'redis_rate_limiter'):
|
||||
# No Redis rate limiter, execute function normally
|
||||
return f(*args, **kwargs)
|
||||
|
||||
rate_limiter = current_app.redis_rate_limiter
|
||||
|
||||
# Get client ID
|
||||
client_id = rate_limiter.get_client_id(request)
|
||||
ip = request.remote_addr
|
||||
|
||||
# Check if IP is blocked
|
||||
if rate_limiter.is_ip_blocked(ip):
|
||||
return jsonify({
|
||||
'error': 'IP temporarily blocked due to excessive requests'
|
||||
}), 429
|
||||
|
||||
# Get endpoint
|
||||
endpoint_path = endpoint or request.endpoint
|
||||
|
||||
# Override default limits if specified
|
||||
if any([requests_per_minute, requests_per_hour, burst_size]):
|
||||
limits = rate_limiter.get_limits(endpoint_path).copy()
|
||||
if requests_per_minute:
|
||||
limits['requests_per_minute'] = requests_per_minute
|
||||
if requests_per_hour:
|
||||
limits['requests_per_hour'] = requests_per_hour
|
||||
if burst_size:
|
||||
limits['burst_size'] = burst_size
|
||||
rate_limiter.endpoint_limits[endpoint_path] = limits
|
||||
|
||||
# Check request size if needed
|
||||
request_size = 0
|
||||
if check_size:
|
||||
request_size = request.content_length or 0
|
||||
|
||||
# Check rate limit
|
||||
allowed, message, headers = rate_limiter.check_rate_limit(
|
||||
client_id, endpoint_path, request_size
|
||||
)
|
||||
|
||||
if not allowed:
|
||||
# Log excessive requests
|
||||
logger.warning(f"Rate limit exceeded for {client_id} on {endpoint_path}: {message}")
|
||||
|
||||
# Check if we should temporarily block this IP
|
||||
stats = rate_limiter.get_client_stats(client_id)
|
||||
if stats and stats['requests_last_minute'] > 100:
|
||||
rate_limiter.block_ip_temporarily(ip, 3600) # 1 hour block
|
||||
|
||||
response = jsonify({
|
||||
'error': message,
|
||||
'retry_after': headers.get('retry_after') if headers else 60
|
||||
})
|
||||
response.status_code = 429
|
||||
|
||||
# Add rate limit headers
|
||||
if headers:
|
||||
response.headers['X-RateLimit-Limit'] = str(headers['limit'])
|
||||
response.headers['X-RateLimit-Remaining'] = str(headers['remaining'])
|
||||
response.headers['X-RateLimit-Reset'] = str(headers['reset'])
|
||||
response.headers['Retry-After'] = str(headers['retry_after'])
|
||||
|
||||
return response
|
||||
|
||||
# Track concurrent requests
|
||||
rate_limiter.increment_concurrent()
|
||||
|
||||
try:
|
||||
# Add rate limit info to response
|
||||
g.rate_limit_headers = headers
|
||||
response = f(*args, **kwargs)
|
||||
|
||||
# Add headers to successful response
|
||||
if headers and hasattr(response, 'headers'):
|
||||
response.headers['X-RateLimit-Limit'] = str(headers['limit'])
|
||||
response.headers['X-RateLimit-Remaining'] = str(headers['remaining'])
|
||||
response.headers['X-RateLimit-Reset'] = str(headers['reset'])
|
||||
|
||||
return response
|
||||
finally:
|
||||
rate_limiter.decrement_concurrent()
|
||||
|
||||
return decorated_function
|
||||
return decorator
|
389
redis_session_manager.py
Normal file
389
redis_session_manager.py
Normal file
@ -0,0 +1,389 @@
|
||||
# Redis-based session management system
|
||||
import time
|
||||
import uuid
|
||||
import logging
|
||||
from datetime import datetime
|
||||
from typing import Dict, Any, Optional, List
|
||||
from dataclasses import dataclass, asdict
|
||||
from flask import session, request, g
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
@dataclass
|
||||
class SessionInfo:
|
||||
"""Session information stored in Redis"""
|
||||
session_id: str
|
||||
user_id: Optional[str] = None
|
||||
ip_address: Optional[str] = None
|
||||
user_agent: Optional[str] = None
|
||||
created_at: float = None
|
||||
last_activity: float = None
|
||||
request_count: int = 0
|
||||
resource_count: int = 0
|
||||
total_bytes_used: int = 0
|
||||
metadata: Dict[str, Any] = None
|
||||
|
||||
def __post_init__(self):
|
||||
if self.created_at is None:
|
||||
self.created_at = time.time()
|
||||
if self.last_activity is None:
|
||||
self.last_activity = time.time()
|
||||
if self.metadata is None:
|
||||
self.metadata = {}
|
||||
|
||||
|
||||
class RedisSessionManager:
|
||||
"""
|
||||
Session management using Redis for distributed sessions
|
||||
"""
|
||||
def __init__(self, redis_manager, config: Dict[str, Any] = None):
|
||||
self.redis = redis_manager
|
||||
self.config = config or {}
|
||||
|
||||
# Configuration
|
||||
self.max_session_duration = self.config.get('max_session_duration', 3600) # 1 hour
|
||||
self.max_idle_time = self.config.get('max_idle_time', 900) # 15 minutes
|
||||
self.max_resources_per_session = self.config.get('max_resources_per_session', 100)
|
||||
self.max_bytes_per_session = self.config.get('max_bytes_per_session', 100 * 1024 * 1024) # 100MB
|
||||
|
||||
logger.info("Redis session manager initialized")
|
||||
|
||||
def create_session(self, session_id: str = None, user_id: str = None,
|
||||
ip_address: str = None, user_agent: str = None) -> SessionInfo:
|
||||
"""Create a new session"""
|
||||
if not session_id:
|
||||
session_id = str(uuid.uuid4())
|
||||
|
||||
# Check if session already exists
|
||||
existing = self.get_session(session_id)
|
||||
if existing:
|
||||
logger.warning(f"Session {session_id} already exists")
|
||||
return existing
|
||||
|
||||
session_info = SessionInfo(
|
||||
session_id=session_id,
|
||||
user_id=user_id,
|
||||
ip_address=ip_address,
|
||||
user_agent=user_agent
|
||||
)
|
||||
|
||||
# Save to Redis
|
||||
self._save_session(session_info)
|
||||
|
||||
# Add to active sessions set
|
||||
self.redis.sadd('active_sessions', session_id)
|
||||
|
||||
# Update stats
|
||||
self.redis.incr('stats:sessions:created')
|
||||
|
||||
logger.info(f"Created session {session_id}")
|
||||
return session_info
|
||||
|
||||
def get_session(self, session_id: str) -> Optional[SessionInfo]:
|
||||
"""Get a session by ID"""
|
||||
data = self.redis.get(f'session:{session_id}')
|
||||
if not data:
|
||||
return None
|
||||
|
||||
# Update last activity
|
||||
session_info = SessionInfo(**data)
|
||||
session_info.last_activity = time.time()
|
||||
self._save_session(session_info)
|
||||
|
||||
return session_info
|
||||
|
||||
def update_session_activity(self, session_id: str):
|
||||
"""Update session last activity time"""
|
||||
session_info = self.get_session(session_id)
|
||||
if session_info:
|
||||
session_info.last_activity = time.time()
|
||||
session_info.request_count += 1
|
||||
self._save_session(session_info)
|
||||
|
||||
def add_resource(self, session_id: str, resource_type: str,
|
||||
resource_id: str = None, path: str = None,
|
||||
size_bytes: int = 0, metadata: Dict[str, Any] = None) -> bool:
|
||||
"""Add a resource to a session"""
|
||||
session_info = self.get_session(session_id)
|
||||
if not session_info:
|
||||
logger.warning(f"Session {session_id} not found")
|
||||
return False
|
||||
|
||||
# Check limits
|
||||
if session_info.resource_count >= self.max_resources_per_session:
|
||||
logger.warning(f"Session {session_id} reached resource limit")
|
||||
# Clean up oldest resources
|
||||
self._cleanup_oldest_resources(session_id, 1)
|
||||
|
||||
if session_info.total_bytes_used + size_bytes > self.max_bytes_per_session:
|
||||
logger.warning(f"Session {session_id} reached size limit")
|
||||
bytes_to_free = (session_info.total_bytes_used + size_bytes) - self.max_bytes_per_session
|
||||
self._cleanup_resources_by_size(session_id, bytes_to_free)
|
||||
|
||||
# Generate resource ID if not provided
|
||||
if not resource_id:
|
||||
resource_id = str(uuid.uuid4())
|
||||
|
||||
# Store resource info
|
||||
resource_key = f'session:{session_id}:resource:{resource_id}'
|
||||
resource_data = {
|
||||
'resource_id': resource_id,
|
||||
'resource_type': resource_type,
|
||||
'path': path,
|
||||
'size_bytes': size_bytes,
|
||||
'created_at': time.time(),
|
||||
'metadata': metadata or {}
|
||||
}
|
||||
|
||||
self.redis.set(resource_key, resource_data, expire=self.max_session_duration)
|
||||
|
||||
# Add to session's resource set
|
||||
self.redis.sadd(f'session:{session_id}:resources', resource_id)
|
||||
|
||||
# Update session info
|
||||
session_info.resource_count += 1
|
||||
session_info.total_bytes_used += size_bytes
|
||||
self._save_session(session_info)
|
||||
|
||||
# Update global stats
|
||||
self.redis.incr('stats:resources:active')
|
||||
self.redis.incr('stats:bytes:active', size_bytes)
|
||||
|
||||
logger.debug(f"Added {resource_type} resource {resource_id} to session {session_id}")
|
||||
return True
|
||||
|
||||
def remove_resource(self, session_id: str, resource_id: str) -> bool:
|
||||
"""Remove a resource from a session"""
|
||||
# Get resource info
|
||||
resource_key = f'session:{session_id}:resource:{resource_id}'
|
||||
resource_data = self.redis.get(resource_key)
|
||||
if not resource_data:
|
||||
return False
|
||||
|
||||
# Clean up the actual resource (file, etc.)
|
||||
self._cleanup_resource(resource_data)
|
||||
|
||||
# Remove from Redis
|
||||
self.redis.delete(resource_key)
|
||||
self.redis.srem(f'session:{session_id}:resources', resource_id)
|
||||
|
||||
# Update session info
|
||||
session_info = self.get_session(session_id)
|
||||
if session_info:
|
||||
session_info.resource_count -= 1
|
||||
session_info.total_bytes_used -= resource_data.get('size_bytes', 0)
|
||||
self._save_session(session_info)
|
||||
|
||||
# Update stats
|
||||
self.redis.decr('stats:resources:active')
|
||||
self.redis.decr('stats:bytes:active', resource_data.get('size_bytes', 0))
|
||||
self.redis.incr('stats:resources:cleaned')
|
||||
self.redis.incr('stats:bytes:cleaned', resource_data.get('size_bytes', 0))
|
||||
|
||||
logger.debug(f"Removed resource {resource_id} from session {session_id}")
|
||||
return True
|
||||
|
||||
def cleanup_session(self, session_id: str) -> bool:
|
||||
"""Clean up a session and all its resources"""
|
||||
session_info = self.get_session(session_id)
|
||||
if not session_info:
|
||||
return False
|
||||
|
||||
# Get all resources
|
||||
resource_ids = self.redis.smembers(f'session:{session_id}:resources')
|
||||
|
||||
# Clean up each resource
|
||||
for resource_id in resource_ids:
|
||||
self.remove_resource(session_id, resource_id)
|
||||
|
||||
# Remove session data
|
||||
self.redis.delete(f'session:{session_id}')
|
||||
self.redis.delete(f'session:{session_id}:resources')
|
||||
self.redis.srem('active_sessions', session_id)
|
||||
|
||||
# Update stats
|
||||
self.redis.incr('stats:sessions:cleaned')
|
||||
|
||||
logger.info(f"Cleaned up session {session_id}")
|
||||
return True
|
||||
|
||||
def cleanup_expired_sessions(self):
|
||||
"""Clean up sessions that have exceeded max duration"""
|
||||
now = time.time()
|
||||
active_sessions = self.redis.smembers('active_sessions')
|
||||
|
||||
for session_id in active_sessions:
|
||||
session_info = self.get_session(session_id)
|
||||
if session_info and (now - session_info.created_at > self.max_session_duration):
|
||||
logger.info(f"Cleaning up expired session {session_id}")
|
||||
self.cleanup_session(session_id)
|
||||
|
||||
def cleanup_idle_sessions(self):
|
||||
"""Clean up sessions that have been idle too long"""
|
||||
now = time.time()
|
||||
active_sessions = self.redis.smembers('active_sessions')
|
||||
|
||||
for session_id in active_sessions:
|
||||
session_info = self.get_session(session_id)
|
||||
if session_info and (now - session_info.last_activity > self.max_idle_time):
|
||||
logger.info(f"Cleaning up idle session {session_id}")
|
||||
self.cleanup_session(session_id)
|
||||
|
||||
def get_session_info(self, session_id: str) -> Optional[Dict[str, Any]]:
|
||||
"""Get detailed information about a session"""
|
||||
session_info = self.get_session(session_id)
|
||||
if not session_info:
|
||||
return None
|
||||
|
||||
# Get resources
|
||||
resource_ids = self.redis.smembers(f'session:{session_id}:resources')
|
||||
resources = []
|
||||
|
||||
for resource_id in resource_ids:
|
||||
resource_data = self.redis.get(f'session:{session_id}:resource:{resource_id}')
|
||||
if resource_data:
|
||||
resources.append({
|
||||
'resource_id': resource_data['resource_id'],
|
||||
'resource_type': resource_data['resource_type'],
|
||||
'size_bytes': resource_data['size_bytes'],
|
||||
'created_at': datetime.fromtimestamp(resource_data['created_at']).isoformat()
|
||||
})
|
||||
|
||||
return {
|
||||
'session_id': session_info.session_id,
|
||||
'user_id': session_info.user_id,
|
||||
'ip_address': session_info.ip_address,
|
||||
'created_at': datetime.fromtimestamp(session_info.created_at).isoformat(),
|
||||
'last_activity': datetime.fromtimestamp(session_info.last_activity).isoformat(),
|
||||
'duration_seconds': int(time.time() - session_info.created_at),
|
||||
'idle_seconds': int(time.time() - session_info.last_activity),
|
||||
'request_count': session_info.request_count,
|
||||
'resource_count': session_info.resource_count,
|
||||
'total_bytes_used': session_info.total_bytes_used,
|
||||
'resources': resources
|
||||
}
|
||||
|
||||
def get_all_sessions_info(self) -> List[Dict[str, Any]]:
|
||||
"""Get information about all active sessions"""
|
||||
active_sessions = self.redis.smembers('active_sessions')
|
||||
return [
|
||||
self.get_session_info(session_id)
|
||||
for session_id in active_sessions
|
||||
if self.get_session_info(session_id)
|
||||
]
|
||||
|
||||
def get_stats(self) -> Dict[str, Any]:
|
||||
"""Get session manager statistics"""
|
||||
active_sessions = self.redis.scard('active_sessions')
|
||||
|
||||
return {
|
||||
'active_sessions': active_sessions,
|
||||
'total_sessions_created': self.redis.get('stats:sessions:created', 0),
|
||||
'total_sessions_cleaned': self.redis.get('stats:sessions:cleaned', 0),
|
||||
'active_resources': self.redis.get('stats:resources:active', 0),
|
||||
'total_resources_cleaned': self.redis.get('stats:resources:cleaned', 0),
|
||||
'active_bytes': self.redis.get('stats:bytes:active', 0),
|
||||
'total_bytes_cleaned': self.redis.get('stats:bytes:cleaned', 0)
|
||||
}
|
||||
|
||||
def _save_session(self, session_info: SessionInfo):
|
||||
"""Save session info to Redis"""
|
||||
key = f'session:{session_info.session_id}'
|
||||
data = asdict(session_info)
|
||||
self.redis.set(key, data, expire=self.max_session_duration)
|
||||
|
||||
def _cleanup_resource(self, resource_data: Dict[str, Any]):
|
||||
"""Clean up a resource (e.g., delete file)"""
|
||||
import os
|
||||
|
||||
if resource_data.get('resource_type') in ['audio_file', 'temp_file']:
|
||||
path = resource_data.get('path')
|
||||
if path and os.path.exists(path):
|
||||
try:
|
||||
os.remove(path)
|
||||
logger.debug(f"Removed file {path}")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to remove file {path}: {e}")
|
||||
|
||||
def _cleanup_oldest_resources(self, session_id: str, count: int):
|
||||
"""Clean up oldest resources from a session"""
|
||||
resource_ids = list(self.redis.smembers(f'session:{session_id}:resources'))
|
||||
|
||||
# Get resource creation times
|
||||
resources_with_time = []
|
||||
for resource_id in resource_ids:
|
||||
resource_data = self.redis.get(f'session:{session_id}:resource:{resource_id}')
|
||||
if resource_data:
|
||||
resources_with_time.append((resource_id, resource_data.get('created_at', 0)))
|
||||
|
||||
# Sort by creation time and remove oldest
|
||||
resources_with_time.sort(key=lambda x: x[1])
|
||||
for resource_id, _ in resources_with_time[:count]:
|
||||
self.remove_resource(session_id, resource_id)
|
||||
|
||||
def _cleanup_resources_by_size(self, session_id: str, bytes_to_free: int):
|
||||
"""Clean up resources to free up space"""
|
||||
resource_ids = list(self.redis.smembers(f'session:{session_id}:resources'))
|
||||
|
||||
# Get resource sizes
|
||||
resources_with_size = []
|
||||
for resource_id in resource_ids:
|
||||
resource_data = self.redis.get(f'session:{session_id}:resource:{resource_id}')
|
||||
if resource_data:
|
||||
resources_with_size.append((resource_id, resource_data.get('size_bytes', 0)))
|
||||
|
||||
# Sort by size (largest first) and remove until we've freed enough
|
||||
resources_with_size.sort(key=lambda x: x[1], reverse=True)
|
||||
freed_bytes = 0
|
||||
|
||||
for resource_id, size in resources_with_size:
|
||||
if freed_bytes >= bytes_to_free:
|
||||
break
|
||||
freed_bytes += size
|
||||
self.remove_resource(session_id, resource_id)
|
||||
|
||||
|
||||
def init_app(app):
|
||||
"""Initialize Redis session management for Flask app"""
|
||||
# Get Redis manager
|
||||
redis_manager = getattr(app, 'redis', None)
|
||||
if not redis_manager:
|
||||
raise RuntimeError("Redis manager not initialized. Call init_redis() first.")
|
||||
|
||||
config = {
|
||||
'max_session_duration': app.config.get('MAX_SESSION_DURATION', 3600),
|
||||
'max_idle_time': app.config.get('MAX_SESSION_IDLE_TIME', 900),
|
||||
'max_resources_per_session': app.config.get('MAX_RESOURCES_PER_SESSION', 100),
|
||||
'max_bytes_per_session': app.config.get('MAX_BYTES_PER_SESSION', 100 * 1024 * 1024)
|
||||
}
|
||||
|
||||
manager = RedisSessionManager(redis_manager, config)
|
||||
app.redis_session_manager = manager
|
||||
|
||||
# Add before_request handler
|
||||
@app.before_request
|
||||
def before_request_session():
|
||||
# Get or create session
|
||||
session_id = session.get('session_id')
|
||||
if not session_id:
|
||||
session_id = str(uuid.uuid4())
|
||||
session['session_id'] = session_id
|
||||
session.permanent = True
|
||||
|
||||
# Get session from manager
|
||||
user_session = manager.get_session(session_id)
|
||||
if not user_session:
|
||||
user_session = manager.create_session(
|
||||
session_id=session_id,
|
||||
ip_address=request.remote_addr,
|
||||
user_agent=request.headers.get('User-Agent')
|
||||
)
|
||||
|
||||
# Update activity
|
||||
manager.update_session_activity(session_id)
|
||||
|
||||
# Store in g for request access
|
||||
g.user_session = user_session
|
||||
g.session_manager = manager
|
||||
|
||||
logger.info("Redis session management initialized")
|
302
request_size_limiter.py
Normal file
302
request_size_limiter.py
Normal file
@ -0,0 +1,302 @@
|
||||
# Request size limiting middleware for preventing memory exhaustion
|
||||
import logging
|
||||
from functools import wraps
|
||||
from flask import request, jsonify, current_app
|
||||
import os
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
# Default size limits (in bytes)
|
||||
DEFAULT_LIMITS = {
|
||||
'max_content_length': 50 * 1024 * 1024, # 50MB global max
|
||||
'max_audio_size': 25 * 1024 * 1024, # 25MB for audio files
|
||||
'max_json_size': 1 * 1024 * 1024, # 1MB for JSON payloads
|
||||
'max_image_size': 10 * 1024 * 1024, # 10MB for images
|
||||
'max_chunk_size': 1 * 1024 * 1024, # 1MB chunks for streaming
|
||||
}
|
||||
|
||||
# File extension to MIME type mapping
|
||||
AUDIO_EXTENSIONS = {'.wav', '.mp3', '.ogg', '.webm', '.m4a', '.flac', '.aac'}
|
||||
IMAGE_EXTENSIONS = {'.jpg', '.jpeg', '.png', '.gif', '.webp', '.bmp'}
|
||||
|
||||
class RequestSizeLimiter:
|
||||
"""
|
||||
Middleware to enforce request size limits and prevent memory exhaustion
|
||||
"""
|
||||
def __init__(self, app=None, config=None):
|
||||
self.config = config or {}
|
||||
self.limits = {**DEFAULT_LIMITS, **self.config}
|
||||
|
||||
if app:
|
||||
self.init_app(app)
|
||||
|
||||
def init_app(self, app):
|
||||
"""Initialize the Flask application with size limiting"""
|
||||
# Set Flask's MAX_CONTENT_LENGTH
|
||||
app.config['MAX_CONTENT_LENGTH'] = self.limits['max_content_length']
|
||||
|
||||
# Store limiter in app
|
||||
app.request_size_limiter = self
|
||||
|
||||
# Add before_request handler
|
||||
app.before_request(self.check_request_size)
|
||||
|
||||
# Add error handler for 413 Request Entity Too Large
|
||||
app.register_error_handler(413, self.handle_413)
|
||||
|
||||
logger.info(f"Request size limiter initialized with max content length: {self.limits['max_content_length'] / 1024 / 1024:.1f}MB")
|
||||
|
||||
def check_request_size(self):
|
||||
"""Check request size before processing"""
|
||||
# Skip size check for GET, HEAD, OPTIONS
|
||||
if request.method in ('GET', 'HEAD', 'OPTIONS'):
|
||||
return None
|
||||
|
||||
# Get content length
|
||||
content_length = request.content_length
|
||||
|
||||
if content_length is None:
|
||||
# No content-length header, check for chunked encoding
|
||||
if request.headers.get('Transfer-Encoding') == 'chunked':
|
||||
logger.warning(f"Chunked request from {request.remote_addr} to {request.endpoint}")
|
||||
# For chunked requests, we'll need to monitor the stream
|
||||
return None
|
||||
else:
|
||||
# No content, allow it
|
||||
return None
|
||||
|
||||
# Check against global limit
|
||||
if content_length > self.limits['max_content_length']:
|
||||
logger.warning(f"Request from {request.remote_addr} exceeds global limit: {content_length} bytes")
|
||||
return jsonify({
|
||||
'error': 'Request too large',
|
||||
'max_size': self.limits['max_content_length'],
|
||||
'your_size': content_length
|
||||
}), 413
|
||||
|
||||
# Check endpoint-specific limits
|
||||
endpoint = request.endpoint
|
||||
if endpoint:
|
||||
endpoint_limit = self.get_endpoint_limit(endpoint)
|
||||
if endpoint_limit and content_length > endpoint_limit:
|
||||
logger.warning(f"Request from {request.remote_addr} to {endpoint} exceeds endpoint limit: {content_length} bytes")
|
||||
return jsonify({
|
||||
'error': f'Request too large for {endpoint}',
|
||||
'max_size': endpoint_limit,
|
||||
'your_size': content_length
|
||||
}), 413
|
||||
|
||||
# Check file-specific limits
|
||||
if request.files:
|
||||
for file_key, file_obj in request.files.items():
|
||||
# Check file size
|
||||
file_obj.seek(0, os.SEEK_END)
|
||||
file_size = file_obj.tell()
|
||||
file_obj.seek(0) # Reset position
|
||||
|
||||
# Determine file type
|
||||
filename = file_obj.filename or ''
|
||||
file_ext = os.path.splitext(filename)[1].lower()
|
||||
|
||||
# Apply type-specific limits
|
||||
if file_ext in AUDIO_EXTENSIONS:
|
||||
max_size = self.limits.get('max_audio_size', self.limits['max_content_length'])
|
||||
if file_size > max_size:
|
||||
logger.warning(f"Audio file from {request.remote_addr} exceeds limit: {file_size} bytes")
|
||||
return jsonify({
|
||||
'error': 'Audio file too large',
|
||||
'max_size': max_size,
|
||||
'your_size': file_size,
|
||||
'max_size_mb': round(max_size / 1024 / 1024, 1)
|
||||
}), 413
|
||||
|
||||
elif file_ext in IMAGE_EXTENSIONS:
|
||||
max_size = self.limits.get('max_image_size', self.limits['max_content_length'])
|
||||
if file_size > max_size:
|
||||
logger.warning(f"Image file from {request.remote_addr} exceeds limit: {file_size} bytes")
|
||||
return jsonify({
|
||||
'error': 'Image file too large',
|
||||
'max_size': max_size,
|
||||
'your_size': file_size,
|
||||
'max_size_mb': round(max_size / 1024 / 1024, 1)
|
||||
}), 413
|
||||
|
||||
# Check JSON payload size
|
||||
if request.is_json:
|
||||
try:
|
||||
# Get raw data size
|
||||
data_size = len(request.get_data())
|
||||
max_json = self.limits.get('max_json_size', self.limits['max_content_length'])
|
||||
|
||||
if data_size > max_json:
|
||||
logger.warning(f"JSON payload from {request.remote_addr} exceeds limit: {data_size} bytes")
|
||||
return jsonify({
|
||||
'error': 'JSON payload too large',
|
||||
'max_size': max_json,
|
||||
'your_size': data_size,
|
||||
'max_size_kb': round(max_json / 1024, 1)
|
||||
}), 413
|
||||
except Exception as e:
|
||||
logger.error(f"Error checking JSON size: {e}")
|
||||
|
||||
return None
|
||||
|
||||
def get_endpoint_limit(self, endpoint):
|
||||
"""Get size limit for specific endpoint"""
|
||||
endpoint_limits = {
|
||||
'transcribe': self.limits.get('max_audio_size', 25 * 1024 * 1024),
|
||||
'speak': self.limits.get('max_json_size', 1 * 1024 * 1024),
|
||||
'translate': self.limits.get('max_json_size', 1 * 1024 * 1024),
|
||||
'translate_stream': self.limits.get('max_json_size', 1 * 1024 * 1024),
|
||||
}
|
||||
return endpoint_limits.get(endpoint)
|
||||
|
||||
def handle_413(self, error):
|
||||
"""Handle 413 Request Entity Too Large errors"""
|
||||
logger.warning(f"413 error from {request.remote_addr}: {error}")
|
||||
return jsonify({
|
||||
'error': 'Request entity too large',
|
||||
'message': 'The request payload is too large. Please reduce the size and try again.',
|
||||
'max_size': self.limits['max_content_length'],
|
||||
'max_size_mb': round(self.limits['max_content_length'] / 1024 / 1024, 1)
|
||||
}), 413
|
||||
|
||||
def update_limits(self, **kwargs):
|
||||
"""Update size limits dynamically"""
|
||||
old_limits = self.limits.copy()
|
||||
self.limits.update(kwargs)
|
||||
|
||||
# Update Flask's MAX_CONTENT_LENGTH if changed
|
||||
if 'max_content_length' in kwargs and current_app:
|
||||
current_app.config['MAX_CONTENT_LENGTH'] = kwargs['max_content_length']
|
||||
|
||||
logger.info(f"Updated size limits: {kwargs}")
|
||||
return old_limits
|
||||
|
||||
def limit_request_size(**limit_kwargs):
|
||||
"""
|
||||
Decorator to apply custom size limits to specific routes
|
||||
|
||||
Usage:
|
||||
@app.route('/upload')
|
||||
@limit_request_size(max_size=10*1024*1024) # 10MB limit
|
||||
def upload():
|
||||
...
|
||||
"""
|
||||
def decorator(f):
|
||||
@wraps(f)
|
||||
def wrapper(*args, **kwargs):
|
||||
# Check content length
|
||||
content_length = request.content_length
|
||||
max_size = limit_kwargs.get('max_size', DEFAULT_LIMITS['max_content_length'])
|
||||
|
||||
if content_length and content_length > max_size:
|
||||
logger.warning(f"Request to {request.endpoint} exceeds custom limit: {content_length} bytes")
|
||||
return jsonify({
|
||||
'error': 'Request too large',
|
||||
'max_size': max_size,
|
||||
'your_size': content_length,
|
||||
'max_size_mb': round(max_size / 1024 / 1024, 1)
|
||||
}), 413
|
||||
|
||||
# Check specific file types if specified
|
||||
if 'max_audio_size' in limit_kwargs and request.files:
|
||||
for file_obj in request.files.values():
|
||||
if file_obj.filename:
|
||||
ext = os.path.splitext(file_obj.filename)[1].lower()
|
||||
if ext in AUDIO_EXTENSIONS:
|
||||
file_obj.seek(0, os.SEEK_END)
|
||||
file_size = file_obj.tell()
|
||||
file_obj.seek(0)
|
||||
|
||||
if file_size > limit_kwargs['max_audio_size']:
|
||||
return jsonify({
|
||||
'error': 'Audio file too large',
|
||||
'max_size': limit_kwargs['max_audio_size'],
|
||||
'your_size': file_size,
|
||||
'max_size_mb': round(limit_kwargs['max_audio_size'] / 1024 / 1024, 1)
|
||||
}), 413
|
||||
|
||||
return f(*args, **kwargs)
|
||||
return wrapper
|
||||
return decorator
|
||||
|
||||
class StreamSizeLimiter:
|
||||
"""
|
||||
Helper class to limit streaming request sizes
|
||||
"""
|
||||
def __init__(self, stream, max_size):
|
||||
self.stream = stream
|
||||
self.max_size = max_size
|
||||
self.bytes_read = 0
|
||||
|
||||
def read(self, size=-1):
|
||||
"""Read from stream with size limit enforcement"""
|
||||
if size == -1:
|
||||
# Read all remaining, but respect limit
|
||||
size = self.max_size - self.bytes_read
|
||||
|
||||
# Check if we would exceed limit
|
||||
if self.bytes_read + size > self.max_size:
|
||||
raise ValueError(f"Stream size exceeds limit of {self.max_size} bytes")
|
||||
|
||||
data = self.stream.read(size)
|
||||
self.bytes_read += len(data)
|
||||
|
||||
return data
|
||||
|
||||
def readline(self, size=-1):
|
||||
"""Read line from stream with size limit enforcement"""
|
||||
if size == -1:
|
||||
size = self.max_size - self.bytes_read
|
||||
|
||||
if self.bytes_read + size > self.max_size:
|
||||
raise ValueError(f"Stream size exceeds limit of {self.max_size} bytes")
|
||||
|
||||
line = self.stream.readline(size)
|
||||
self.bytes_read += len(line)
|
||||
|
||||
return line
|
||||
|
||||
# Utility functions
|
||||
def get_request_size():
|
||||
"""Get the size of the current request"""
|
||||
if request.content_length:
|
||||
return request.content_length
|
||||
|
||||
# For chunked requests, read and measure
|
||||
try:
|
||||
data = request.get_data()
|
||||
return len(data)
|
||||
except Exception:
|
||||
return 0
|
||||
|
||||
def format_size(size_bytes):
|
||||
"""Format size in human-readable format"""
|
||||
for unit in ['B', 'KB', 'MB', 'GB']:
|
||||
if size_bytes < 1024.0:
|
||||
return f"{size_bytes:.1f} {unit}"
|
||||
size_bytes /= 1024.0
|
||||
return f"{size_bytes:.1f} TB"
|
||||
|
||||
# Configuration helper
|
||||
def configure_size_limits(app, **kwargs):
|
||||
"""
|
||||
Configure size limits for the application
|
||||
|
||||
Args:
|
||||
app: Flask application
|
||||
max_content_length: Global maximum request size
|
||||
max_audio_size: Maximum audio file size
|
||||
max_json_size: Maximum JSON payload size
|
||||
max_image_size: Maximum image file size
|
||||
"""
|
||||
config = {
|
||||
'max_content_length': kwargs.get('max_content_length', DEFAULT_LIMITS['max_content_length']),
|
||||
'max_audio_size': kwargs.get('max_audio_size', DEFAULT_LIMITS['max_audio_size']),
|
||||
'max_json_size': kwargs.get('max_json_size', DEFAULT_LIMITS['max_json_size']),
|
||||
'max_image_size': kwargs.get('max_image_size', DEFAULT_LIMITS['max_image_size']),
|
||||
}
|
||||
|
||||
limiter = RequestSizeLimiter(app, config)
|
||||
return limiter
|
27
requirements-prod.txt
Normal file
27
requirements-prod.txt
Normal file
@ -0,0 +1,27 @@
|
||||
# Production requirements for Talk2Me
|
||||
# Includes base requirements plus production WSGI server
|
||||
|
||||
# Include base requirements
|
||||
-r requirements.txt
|
||||
|
||||
# Production WSGI server
|
||||
gunicorn==21.2.0
|
||||
|
||||
# Async workers (optional, for better concurrency)
|
||||
gevent==23.9.1
|
||||
greenlet==3.0.1
|
||||
|
||||
# Production monitoring
|
||||
prometheus-client==0.19.0
|
||||
|
||||
# Production caching (optional)
|
||||
redis==5.0.1
|
||||
hiredis==2.3.2
|
||||
|
||||
# Database for production (optional, for session storage)
|
||||
psycopg2-binary==2.9.9
|
||||
SQLAlchemy==2.0.23
|
||||
|
||||
# Additional production utilities
|
||||
python-json-logger==2.0.7 # JSON logging
|
||||
sentry-sdk[flask]==1.39.1 # Error tracking (optional)
|
@ -1,5 +1,24 @@
|
||||
flask
|
||||
flask-cors
|
||||
flask-sqlalchemy
|
||||
flask-migrate
|
||||
flask-jwt-extended
|
||||
flask-bcrypt
|
||||
flask-login
|
||||
requests
|
||||
openai-whisper
|
||||
torch
|
||||
ollama
|
||||
pywebpush
|
||||
cryptography
|
||||
python-dotenv
|
||||
click
|
||||
colorlog
|
||||
psutil
|
||||
redis
|
||||
psycopg2-binary
|
||||
alembic
|
||||
flask-socketio
|
||||
python-socketio
|
||||
eventlet
|
||||
python-dateutil
|
||||
|
32
run_dev_server.sh
Executable file
32
run_dev_server.sh
Executable file
@ -0,0 +1,32 @@
|
||||
#!/bin/bash
|
||||
# Run Talk2Me development server locally
|
||||
|
||||
echo "Starting Talk2Me development server..."
|
||||
echo "=================================="
|
||||
echo "Admin Dashboard: http://localhost:5005/admin"
|
||||
echo " Token: 4CFvwzmeDWhecfuOHYz7Hyb8nQQ="
|
||||
echo ""
|
||||
echo "User Login: http://localhost:5005/login"
|
||||
echo " Username: admin"
|
||||
echo " Password: talk2me123"
|
||||
echo ""
|
||||
echo "API Authentication:"
|
||||
echo " API Key: 6sy2_m8e89FeC2RmUo0CcgufM9b_0OoIwIa8LSEbNhI"
|
||||
echo "=================================="
|
||||
echo ""
|
||||
|
||||
# Kill any existing process on port 5005
|
||||
lsof -ti:5005 | xargs kill -9 2>/dev/null
|
||||
|
||||
# Set environment variables
|
||||
export FLASK_ENV=development
|
||||
export FLASK_DEBUG=1
|
||||
|
||||
# Run with gunicorn for a more production-like environment
|
||||
gunicorn --bind 0.0.0.0:5005 \
|
||||
--workers 1 \
|
||||
--threads 2 \
|
||||
--timeout 120 \
|
||||
--reload \
|
||||
--log-level debug \
|
||||
wsgi:application
|
411
secrets_manager.py
Normal file
411
secrets_manager.py
Normal file
@ -0,0 +1,411 @@
|
||||
# Secrets management system for secure configuration
|
||||
import os
|
||||
import json
|
||||
import base64
|
||||
import logging
|
||||
from typing import Any, Dict, Optional, List
|
||||
from datetime import datetime, timedelta
|
||||
from cryptography.fernet import Fernet
|
||||
from cryptography.hazmat.primitives import hashes
|
||||
from cryptography.hazmat.primitives.kdf.pbkdf2 import PBKDF2HMAC
|
||||
import hashlib
|
||||
import hmac
|
||||
import secrets
|
||||
from functools import lru_cache
|
||||
from threading import Lock
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
class SecretsManager:
|
||||
"""
|
||||
Secure secrets management with encryption, rotation, and audit logging
|
||||
"""
|
||||
def __init__(self, config_file: str = None):
|
||||
self.config_file = config_file or os.environ.get('SECRETS_CONFIG', '.secrets.json')
|
||||
self.lock = Lock()
|
||||
self._secrets_cache = {}
|
||||
self._encryption_key = None
|
||||
self._master_key = None
|
||||
self._audit_log = []
|
||||
self._rotation_schedule = {}
|
||||
self._validators = {}
|
||||
|
||||
# Initialize encryption
|
||||
self._init_encryption()
|
||||
|
||||
# Load secrets
|
||||
self._load_secrets()
|
||||
|
||||
def _init_encryption(self):
|
||||
"""Initialize encryption key from environment or generate new one"""
|
||||
# Try to get master key from environment
|
||||
master_key = os.environ.get('MASTER_KEY')
|
||||
|
||||
if not master_key:
|
||||
# Try to load from secure file
|
||||
key_file = os.environ.get('MASTER_KEY_FILE', '.master_key')
|
||||
if os.path.exists(key_file):
|
||||
try:
|
||||
with open(key_file, 'rb') as f:
|
||||
master_key = f.read().decode('utf-8').strip()
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to load master key from file: {e}")
|
||||
|
||||
if not master_key:
|
||||
# Generate new master key
|
||||
logger.warning("No master key found. Generating new one.")
|
||||
master_key = Fernet.generate_key().decode('utf-8')
|
||||
|
||||
# Save to secure file (should be protected by OS permissions)
|
||||
key_file = os.environ.get('MASTER_KEY_FILE', '.master_key')
|
||||
try:
|
||||
with open(key_file, 'wb') as f:
|
||||
f.write(master_key.encode('utf-8'))
|
||||
os.chmod(key_file, 0o600) # Owner read/write only
|
||||
logger.info(f"Master key saved to {key_file}")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to save master key: {e}")
|
||||
|
||||
self._master_key = master_key.encode('utf-8')
|
||||
|
||||
# Derive encryption key from master key
|
||||
kdf = PBKDF2HMAC(
|
||||
algorithm=hashes.SHA256(),
|
||||
length=32,
|
||||
salt=b'talk2me-secrets-salt', # In production, use random salt
|
||||
iterations=100000,
|
||||
)
|
||||
key = base64.urlsafe_b64encode(kdf.derive(self._master_key))
|
||||
self._encryption_key = Fernet(key)
|
||||
|
||||
def _load_secrets(self):
|
||||
"""Load encrypted secrets from file"""
|
||||
if not os.path.exists(self.config_file):
|
||||
logger.info(f"No secrets file found at {self.config_file}")
|
||||
return
|
||||
|
||||
try:
|
||||
with open(self.config_file, 'r') as f:
|
||||
data = json.load(f)
|
||||
|
||||
# Decrypt secrets
|
||||
for key, value in data.get('secrets', {}).items():
|
||||
if isinstance(value, dict) and 'encrypted' in value:
|
||||
try:
|
||||
decrypted = self._decrypt(value['encrypted'])
|
||||
self._secrets_cache[key] = {
|
||||
'value': decrypted,
|
||||
'created': value.get('created'),
|
||||
'rotated': value.get('rotated'),
|
||||
'metadata': value.get('metadata', {})
|
||||
}
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to decrypt secret {key}: {e}")
|
||||
else:
|
||||
# Plain text (for migration)
|
||||
self._secrets_cache[key] = {
|
||||
'value': value,
|
||||
'created': datetime.now().isoformat(),
|
||||
'rotated': None,
|
||||
'metadata': {}
|
||||
}
|
||||
|
||||
# Load rotation schedule
|
||||
self._rotation_schedule = data.get('rotation_schedule', {})
|
||||
|
||||
# Load audit log
|
||||
self._audit_log = data.get('audit_log', [])
|
||||
|
||||
logger.info(f"Loaded {len(self._secrets_cache)} secrets")
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to load secrets: {e}")
|
||||
|
||||
def _save_secrets(self):
|
||||
"""Save encrypted secrets to file"""
|
||||
with self.lock:
|
||||
data = {
|
||||
'secrets': {},
|
||||
'rotation_schedule': self._rotation_schedule,
|
||||
'audit_log': self._audit_log[-1000:] # Keep last 1000 entries
|
||||
}
|
||||
|
||||
# Encrypt secrets
|
||||
for key, secret_data in self._secrets_cache.items():
|
||||
data['secrets'][key] = {
|
||||
'encrypted': self._encrypt(secret_data['value']),
|
||||
'created': secret_data.get('created'),
|
||||
'rotated': secret_data.get('rotated'),
|
||||
'metadata': secret_data.get('metadata', {})
|
||||
}
|
||||
|
||||
# Save to file
|
||||
try:
|
||||
# Write to temporary file first
|
||||
temp_file = f"{self.config_file}.tmp"
|
||||
with open(temp_file, 'w') as f:
|
||||
json.dump(data, f, indent=2)
|
||||
|
||||
# Set secure permissions
|
||||
os.chmod(temp_file, 0o600) # Owner read/write only
|
||||
|
||||
# Atomic rename
|
||||
os.rename(temp_file, self.config_file)
|
||||
|
||||
logger.info(f"Saved {len(self._secrets_cache)} secrets")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to save secrets: {e}")
|
||||
raise
|
||||
|
||||
def _encrypt(self, value: str) -> str:
|
||||
"""Encrypt a value"""
|
||||
if not isinstance(value, str):
|
||||
value = str(value)
|
||||
return self._encryption_key.encrypt(value.encode('utf-8')).decode('utf-8')
|
||||
|
||||
def _decrypt(self, encrypted_value: str) -> str:
|
||||
"""Decrypt a value"""
|
||||
return self._encryption_key.decrypt(encrypted_value.encode('utf-8')).decode('utf-8')
|
||||
|
||||
def _audit(self, action: str, key: str, user: str = None, details: dict = None):
|
||||
"""Add entry to audit log"""
|
||||
entry = {
|
||||
'timestamp': datetime.now().isoformat(),
|
||||
'action': action,
|
||||
'key': key,
|
||||
'user': user or 'system',
|
||||
'details': details or {}
|
||||
}
|
||||
self._audit_log.append(entry)
|
||||
logger.info(f"Audit: {action} on {key} by {user or 'system'}")
|
||||
|
||||
def get(self, key: str, default: Any = None) -> Any:
|
||||
"""Get a secret value"""
|
||||
# Try cache first
|
||||
if key in self._secrets_cache:
|
||||
self._audit('access', key)
|
||||
return self._secrets_cache[key]['value']
|
||||
|
||||
# Try environment variable
|
||||
env_key = f"SECRET_{key.upper()}"
|
||||
env_value = os.environ.get(env_key)
|
||||
if env_value:
|
||||
self._audit('access', key, details={'source': 'environment'})
|
||||
return env_value
|
||||
|
||||
# Try regular environment variable
|
||||
env_value = os.environ.get(key)
|
||||
if env_value:
|
||||
self._audit('access', key, details={'source': 'environment'})
|
||||
return env_value
|
||||
|
||||
self._audit('access_failed', key)
|
||||
return default
|
||||
|
||||
def set(self, key: str, value: str, metadata: dict = None, user: str = None):
|
||||
"""Set a secret value"""
|
||||
with self.lock:
|
||||
old_value = self._secrets_cache.get(key, {}).get('value')
|
||||
|
||||
self._secrets_cache[key] = {
|
||||
'value': value,
|
||||
'created': self._secrets_cache.get(key, {}).get('created', datetime.now().isoformat()),
|
||||
'rotated': datetime.now().isoformat() if old_value else None,
|
||||
'metadata': metadata or {}
|
||||
}
|
||||
|
||||
self._audit('set' if not old_value else 'update', key, user)
|
||||
self._save_secrets()
|
||||
|
||||
def delete(self, key: str, user: str = None):
|
||||
"""Delete a secret"""
|
||||
with self.lock:
|
||||
if key in self._secrets_cache:
|
||||
del self._secrets_cache[key]
|
||||
self._audit('delete', key, user)
|
||||
self._save_secrets()
|
||||
return True
|
||||
return False
|
||||
|
||||
def rotate(self, key: str, new_value: str = None, user: str = None):
|
||||
"""Rotate a secret"""
|
||||
with self.lock:
|
||||
if key not in self._secrets_cache:
|
||||
raise KeyError(f"Secret {key} not found")
|
||||
|
||||
old_value = self._secrets_cache[key]['value']
|
||||
|
||||
# Generate new value if not provided
|
||||
if not new_value:
|
||||
if key.endswith('_KEY') or key.endswith('_TOKEN'):
|
||||
new_value = secrets.token_urlsafe(32)
|
||||
elif key.endswith('_PASSWORD'):
|
||||
new_value = secrets.token_urlsafe(24)
|
||||
else:
|
||||
raise ValueError(f"Cannot auto-generate value for {key}")
|
||||
|
||||
# Update secret
|
||||
self._secrets_cache[key]['value'] = new_value
|
||||
self._secrets_cache[key]['rotated'] = datetime.now().isoformat()
|
||||
|
||||
self._audit('rotate', key, user, {'generated': new_value is None})
|
||||
self._save_secrets()
|
||||
|
||||
return old_value, new_value
|
||||
|
||||
def list_secrets(self) -> List[Dict[str, Any]]:
|
||||
"""List all secrets (without values)"""
|
||||
secrets_list = []
|
||||
for key, data in self._secrets_cache.items():
|
||||
secrets_list.append({
|
||||
'key': key,
|
||||
'created': data.get('created'),
|
||||
'rotated': data.get('rotated'),
|
||||
'metadata': data.get('metadata', {}),
|
||||
'has_value': bool(data.get('value'))
|
||||
})
|
||||
return secrets_list
|
||||
|
||||
def add_validator(self, key: str, validator):
|
||||
"""Add a validator function for a secret"""
|
||||
self._validators[key] = validator
|
||||
|
||||
def validate(self, key: str, value: str) -> bool:
|
||||
"""Validate a secret value"""
|
||||
if key in self._validators:
|
||||
try:
|
||||
return self._validators[key](value)
|
||||
except Exception as e:
|
||||
logger.error(f"Validation failed for {key}: {e}")
|
||||
return False
|
||||
return True
|
||||
|
||||
def schedule_rotation(self, key: str, days: int):
|
||||
"""Schedule automatic rotation for a secret"""
|
||||
self._rotation_schedule[key] = {
|
||||
'days': days,
|
||||
'last_rotated': self._secrets_cache.get(key, {}).get('rotated', datetime.now().isoformat())
|
||||
}
|
||||
self._save_secrets()
|
||||
|
||||
def check_rotation_needed(self) -> List[str]:
|
||||
"""Check which secrets need rotation"""
|
||||
needs_rotation = []
|
||||
now = datetime.now()
|
||||
|
||||
for key, schedule in self._rotation_schedule.items():
|
||||
last_rotated = datetime.fromisoformat(schedule['last_rotated'])
|
||||
if now - last_rotated > timedelta(days=schedule['days']):
|
||||
needs_rotation.append(key)
|
||||
|
||||
return needs_rotation
|
||||
|
||||
def get_audit_log(self, key: str = None, limit: int = 100) -> List[Dict]:
|
||||
"""Get audit log entries"""
|
||||
logs = self._audit_log
|
||||
|
||||
if key:
|
||||
logs = [log for log in logs if log['key'] == key]
|
||||
|
||||
return logs[-limit:]
|
||||
|
||||
def export_for_environment(self) -> Dict[str, str]:
|
||||
"""Export secrets as environment variables"""
|
||||
env_vars = {}
|
||||
for key, data in self._secrets_cache.items():
|
||||
env_key = f"SECRET_{key.upper()}"
|
||||
env_vars[env_key] = data['value']
|
||||
return env_vars
|
||||
|
||||
def verify_integrity(self) -> bool:
|
||||
"""Verify integrity of secrets"""
|
||||
try:
|
||||
# Try to decrypt all secrets
|
||||
for key, secret_data in self._secrets_cache.items():
|
||||
if 'value' in secret_data:
|
||||
# Re-encrypt and compare
|
||||
encrypted = self._encrypt(secret_data['value'])
|
||||
decrypted = self._decrypt(encrypted)
|
||||
if decrypted != secret_data['value']:
|
||||
logger.error(f"Integrity check failed for {key}")
|
||||
return False
|
||||
|
||||
logger.info("Integrity check passed")
|
||||
return True
|
||||
except Exception as e:
|
||||
logger.error(f"Integrity check failed: {e}")
|
||||
return False
|
||||
|
||||
# Global instance
|
||||
_secrets_manager = None
|
||||
_secrets_lock = Lock()
|
||||
|
||||
def get_secrets_manager(config_file: str = None) -> SecretsManager:
|
||||
"""Get or create global secrets manager instance"""
|
||||
global _secrets_manager
|
||||
|
||||
with _secrets_lock:
|
||||
if _secrets_manager is None:
|
||||
_secrets_manager = SecretsManager(config_file)
|
||||
return _secrets_manager
|
||||
|
||||
def get_secret(key: str, default: Any = None) -> Any:
|
||||
"""Convenience function to get a secret"""
|
||||
manager = get_secrets_manager()
|
||||
return manager.get(key, default)
|
||||
|
||||
def set_secret(key: str, value: str, metadata: dict = None):
|
||||
"""Convenience function to set a secret"""
|
||||
manager = get_secrets_manager()
|
||||
manager.set(key, value, metadata)
|
||||
|
||||
# Flask integration
|
||||
def init_app(app):
|
||||
"""Initialize secrets management for Flask app"""
|
||||
manager = get_secrets_manager()
|
||||
|
||||
# Load secrets into app config
|
||||
app.config['SECRET_KEY'] = manager.get('FLASK_SECRET_KEY') or app.config.get('SECRET_KEY')
|
||||
app.config['TTS_API_KEY'] = manager.get('TTS_API_KEY') or app.config.get('TTS_API_KEY')
|
||||
|
||||
# Add secret manager to app
|
||||
app.secrets_manager = manager
|
||||
|
||||
# Add CLI commands
|
||||
@app.cli.command('secrets-list')
|
||||
def list_secrets_cmd():
|
||||
"""List all secrets"""
|
||||
secrets = manager.list_secrets()
|
||||
for secret in secrets:
|
||||
print(f"{secret['key']}: created={secret['created']}, rotated={secret['rotated']}")
|
||||
|
||||
@app.cli.command('secrets-set')
|
||||
def set_secret_cmd():
|
||||
"""Set a secret"""
|
||||
import click
|
||||
key = click.prompt('Secret key')
|
||||
value = click.prompt('Secret value', hide_input=True)
|
||||
manager.set(key, value, user='cli')
|
||||
print(f"Secret {key} set successfully")
|
||||
|
||||
@app.cli.command('secrets-rotate')
|
||||
def rotate_secret_cmd():
|
||||
"""Rotate a secret"""
|
||||
import click
|
||||
key = click.prompt('Secret key to rotate')
|
||||
old_value, new_value = manager.rotate(key, user='cli')
|
||||
print(f"Secret {key} rotated successfully")
|
||||
print(f"New value: {new_value}")
|
||||
|
||||
@app.cli.command('secrets-check-rotation')
|
||||
def check_rotation_cmd():
|
||||
"""Check which secrets need rotation"""
|
||||
needs_rotation = manager.check_rotation_needed()
|
||||
if needs_rotation:
|
||||
print("Secrets needing rotation:")
|
||||
for key in needs_rotation:
|
||||
print(f" - {key}")
|
||||
else:
|
||||
print("No secrets need rotation")
|
||||
|
||||
logger.info("Secrets management initialized")
|
607
session_manager.py
Normal file
607
session_manager.py
Normal file
@ -0,0 +1,607 @@
|
||||
# Session management system for preventing resource leaks
|
||||
import time
|
||||
import uuid
|
||||
import logging
|
||||
from datetime import datetime, timedelta
|
||||
from typing import Dict, Any, Optional, List, Tuple
|
||||
from dataclasses import dataclass, field
|
||||
from threading import Lock, Thread
|
||||
import json
|
||||
import os
|
||||
import tempfile
|
||||
import shutil
|
||||
from collections import defaultdict
|
||||
from functools import wraps
|
||||
from flask import session, request, g, current_app
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
@dataclass
|
||||
class SessionResource:
|
||||
"""Represents a resource associated with a session"""
|
||||
resource_id: str
|
||||
resource_type: str # 'audio_file', 'temp_file', 'websocket', 'stream'
|
||||
path: Optional[str] = None
|
||||
created_at: float = field(default_factory=time.time)
|
||||
last_accessed: float = field(default_factory=time.time)
|
||||
size_bytes: int = 0
|
||||
metadata: Dict[str, Any] = field(default_factory=dict)
|
||||
|
||||
@dataclass
|
||||
class UserSession:
|
||||
"""Represents a user session with associated resources"""
|
||||
session_id: str
|
||||
user_id: Optional[str] = None
|
||||
ip_address: Optional[str] = None
|
||||
user_agent: Optional[str] = None
|
||||
created_at: float = field(default_factory=time.time)
|
||||
last_activity: float = field(default_factory=time.time)
|
||||
resources: Dict[str, SessionResource] = field(default_factory=dict)
|
||||
request_count: int = 0
|
||||
total_bytes_used: int = 0
|
||||
active_streams: int = 0
|
||||
metadata: Dict[str, Any] = field(default_factory=dict)
|
||||
|
||||
class SessionManager:
|
||||
"""
|
||||
Manages user sessions and associated resources to prevent leaks
|
||||
"""
|
||||
def __init__(self, config: Dict[str, Any] = None):
|
||||
self.config = config or {}
|
||||
self.sessions: Dict[str, UserSession] = {}
|
||||
self.lock = Lock()
|
||||
|
||||
# Configuration
|
||||
self.max_session_duration = self.config.get('max_session_duration', 3600) # 1 hour
|
||||
self.max_idle_time = self.config.get('max_idle_time', 900) # 15 minutes
|
||||
self.max_resources_per_session = self.config.get('max_resources_per_session', 100)
|
||||
self.max_bytes_per_session = self.config.get('max_bytes_per_session', 100 * 1024 * 1024) # 100MB
|
||||
self.cleanup_interval = self.config.get('cleanup_interval', 60) # 1 minute
|
||||
self.session_storage_path = self.config.get('session_storage_path',
|
||||
os.path.join(tempfile.gettempdir(), 'talk2me_sessions'))
|
||||
|
||||
# Statistics
|
||||
self.stats = {
|
||||
'total_sessions_created': 0,
|
||||
'total_sessions_cleaned': 0,
|
||||
'total_resources_cleaned': 0,
|
||||
'total_bytes_cleaned': 0,
|
||||
'active_sessions': 0,
|
||||
'active_resources': 0,
|
||||
'active_bytes': 0
|
||||
}
|
||||
|
||||
# Resource cleanup handlers
|
||||
self.cleanup_handlers = {
|
||||
'audio_file': self._cleanup_audio_file,
|
||||
'temp_file': self._cleanup_temp_file,
|
||||
'websocket': self._cleanup_websocket,
|
||||
'stream': self._cleanup_stream
|
||||
}
|
||||
|
||||
# Initialize storage
|
||||
self._init_storage()
|
||||
|
||||
# Start cleanup thread
|
||||
self.cleanup_thread = Thread(target=self._cleanup_loop, daemon=True)
|
||||
self.cleanup_thread.start()
|
||||
|
||||
logger.info("Session manager initialized")
|
||||
|
||||
def _init_storage(self):
|
||||
"""Initialize session storage directory"""
|
||||
try:
|
||||
os.makedirs(self.session_storage_path, mode=0o755, exist_ok=True)
|
||||
logger.info(f"Session storage initialized at {self.session_storage_path}")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to create session storage: {e}")
|
||||
|
||||
def create_session(self, session_id: str = None, user_id: str = None,
|
||||
ip_address: str = None, user_agent: str = None) -> UserSession:
|
||||
"""Create a new session"""
|
||||
with self.lock:
|
||||
if not session_id:
|
||||
session_id = str(uuid.uuid4())
|
||||
|
||||
if session_id in self.sessions:
|
||||
logger.warning(f"Session {session_id} already exists")
|
||||
return self.sessions[session_id]
|
||||
|
||||
session = UserSession(
|
||||
session_id=session_id,
|
||||
user_id=user_id,
|
||||
ip_address=ip_address,
|
||||
user_agent=user_agent
|
||||
)
|
||||
|
||||
self.sessions[session_id] = session
|
||||
self.stats['total_sessions_created'] += 1
|
||||
self.stats['active_sessions'] = len(self.sessions)
|
||||
|
||||
# Create session directory
|
||||
session_dir = os.path.join(self.session_storage_path, session_id)
|
||||
try:
|
||||
os.makedirs(session_dir, mode=0o755, exist_ok=True)
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to create session directory: {e}")
|
||||
|
||||
logger.info(f"Created session {session_id}")
|
||||
return session
|
||||
|
||||
def get_session(self, session_id: str) -> Optional[UserSession]:
|
||||
"""Get a session by ID"""
|
||||
with self.lock:
|
||||
session = self.sessions.get(session_id)
|
||||
if session:
|
||||
session.last_activity = time.time()
|
||||
return session
|
||||
|
||||
def add_resource(self, session_id: str, resource_type: str,
|
||||
resource_id: str = None, path: str = None,
|
||||
size_bytes: int = 0, metadata: Dict[str, Any] = None) -> Optional[SessionResource]:
|
||||
"""Add a resource to a session"""
|
||||
with self.lock:
|
||||
session = self.sessions.get(session_id)
|
||||
if not session:
|
||||
logger.warning(f"Session {session_id} not found")
|
||||
return None
|
||||
|
||||
# Check limits
|
||||
if len(session.resources) >= self.max_resources_per_session:
|
||||
logger.warning(f"Session {session_id} reached resource limit")
|
||||
self._cleanup_oldest_resources(session, 1)
|
||||
|
||||
if session.total_bytes_used + size_bytes > self.max_bytes_per_session:
|
||||
logger.warning(f"Session {session_id} reached size limit")
|
||||
bytes_to_free = (session.total_bytes_used + size_bytes) - self.max_bytes_per_session
|
||||
self._cleanup_resources_by_size(session, bytes_to_free)
|
||||
|
||||
# Create resource
|
||||
if not resource_id:
|
||||
resource_id = str(uuid.uuid4())
|
||||
|
||||
resource = SessionResource(
|
||||
resource_id=resource_id,
|
||||
resource_type=resource_type,
|
||||
path=path,
|
||||
size_bytes=size_bytes,
|
||||
metadata=metadata or {}
|
||||
)
|
||||
|
||||
session.resources[resource_id] = resource
|
||||
session.total_bytes_used += size_bytes
|
||||
session.last_activity = time.time()
|
||||
|
||||
# Update stats
|
||||
self.stats['active_resources'] += 1
|
||||
self.stats['active_bytes'] += size_bytes
|
||||
|
||||
logger.debug(f"Added {resource_type} resource {resource_id} to session {session_id}")
|
||||
return resource
|
||||
|
||||
def remove_resource(self, session_id: str, resource_id: str) -> bool:
|
||||
"""Remove a resource from a session"""
|
||||
with self.lock:
|
||||
session = self.sessions.get(session_id)
|
||||
if not session:
|
||||
return False
|
||||
|
||||
resource = session.resources.get(resource_id)
|
||||
if not resource:
|
||||
return False
|
||||
|
||||
# Cleanup resource
|
||||
self._cleanup_resource(resource)
|
||||
|
||||
# Remove from session
|
||||
del session.resources[resource_id]
|
||||
session.total_bytes_used -= resource.size_bytes
|
||||
|
||||
# Update stats
|
||||
self.stats['active_resources'] -= 1
|
||||
self.stats['active_bytes'] -= resource.size_bytes
|
||||
self.stats['total_resources_cleaned'] += 1
|
||||
self.stats['total_bytes_cleaned'] += resource.size_bytes
|
||||
|
||||
logger.debug(f"Removed resource {resource_id} from session {session_id}")
|
||||
return True
|
||||
|
||||
def update_session_activity(self, session_id: str):
|
||||
"""Update session last activity time"""
|
||||
with self.lock:
|
||||
session = self.sessions.get(session_id)
|
||||
if session:
|
||||
session.last_activity = time.time()
|
||||
session.request_count += 1
|
||||
|
||||
def cleanup_session(self, session_id: str) -> bool:
|
||||
"""Clean up a session and all its resources"""
|
||||
with self.lock:
|
||||
session = self.sessions.get(session_id)
|
||||
if not session:
|
||||
return False
|
||||
|
||||
# Cleanup all resources
|
||||
for resource_id in list(session.resources.keys()):
|
||||
self.remove_resource(session_id, resource_id)
|
||||
|
||||
# Remove session directory
|
||||
session_dir = os.path.join(self.session_storage_path, session_id)
|
||||
try:
|
||||
if os.path.exists(session_dir):
|
||||
shutil.rmtree(session_dir)
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to remove session directory: {e}")
|
||||
|
||||
# Remove session
|
||||
del self.sessions[session_id]
|
||||
|
||||
# Update stats
|
||||
self.stats['active_sessions'] = len(self.sessions)
|
||||
self.stats['total_sessions_cleaned'] += 1
|
||||
|
||||
logger.info(f"Cleaned up session {session_id}")
|
||||
return True
|
||||
|
||||
def _cleanup_resource(self, resource: SessionResource):
|
||||
"""Clean up a single resource"""
|
||||
handler = self.cleanup_handlers.get(resource.resource_type)
|
||||
if handler:
|
||||
try:
|
||||
handler(resource)
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to cleanup {resource.resource_type} {resource.resource_id}: {e}")
|
||||
|
||||
def _cleanup_audio_file(self, resource: SessionResource):
|
||||
"""Clean up audio file resource"""
|
||||
if resource.path and os.path.exists(resource.path):
|
||||
try:
|
||||
os.remove(resource.path)
|
||||
logger.debug(f"Removed audio file {resource.path}")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to remove audio file {resource.path}: {e}")
|
||||
|
||||
def _cleanup_temp_file(self, resource: SessionResource):
|
||||
"""Clean up temporary file resource"""
|
||||
if resource.path and os.path.exists(resource.path):
|
||||
try:
|
||||
os.remove(resource.path)
|
||||
logger.debug(f"Removed temp file {resource.path}")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to remove temp file {resource.path}: {e}")
|
||||
|
||||
def _cleanup_websocket(self, resource: SessionResource):
|
||||
"""Clean up websocket resource"""
|
||||
# Implement websocket cleanup if needed
|
||||
pass
|
||||
|
||||
def _cleanup_stream(self, resource: SessionResource):
|
||||
"""Clean up stream resource"""
|
||||
# Implement stream cleanup if needed
|
||||
if resource.metadata.get('stream_id'):
|
||||
# Close any open streams
|
||||
pass
|
||||
|
||||
def _cleanup_oldest_resources(self, session: UserSession, count: int):
|
||||
"""Clean up oldest resources from a session"""
|
||||
# Sort resources by creation time
|
||||
sorted_resources = sorted(
|
||||
session.resources.items(),
|
||||
key=lambda x: x[1].created_at
|
||||
)
|
||||
|
||||
# Remove oldest resources
|
||||
for resource_id, _ in sorted_resources[:count]:
|
||||
self.remove_resource(session.session_id, resource_id)
|
||||
|
||||
def _cleanup_resources_by_size(self, session: UserSession, bytes_to_free: int):
|
||||
"""Clean up resources to free up space"""
|
||||
freed_bytes = 0
|
||||
|
||||
# Sort resources by size (largest first)
|
||||
sorted_resources = sorted(
|
||||
session.resources.items(),
|
||||
key=lambda x: x[1].size_bytes,
|
||||
reverse=True
|
||||
)
|
||||
|
||||
# Remove resources until we've freed enough space
|
||||
for resource_id, resource in sorted_resources:
|
||||
if freed_bytes >= bytes_to_free:
|
||||
break
|
||||
|
||||
freed_bytes += resource.size_bytes
|
||||
self.remove_resource(session.session_id, resource_id)
|
||||
|
||||
def _cleanup_loop(self):
|
||||
"""Background cleanup thread"""
|
||||
while True:
|
||||
try:
|
||||
time.sleep(self.cleanup_interval)
|
||||
self.cleanup_expired_sessions()
|
||||
self.cleanup_idle_sessions()
|
||||
self.cleanup_orphaned_files()
|
||||
except Exception as e:
|
||||
logger.error(f"Error in cleanup loop: {e}")
|
||||
|
||||
def cleanup_expired_sessions(self):
|
||||
"""Clean up sessions that have exceeded max duration"""
|
||||
with self.lock:
|
||||
now = time.time()
|
||||
expired_sessions = []
|
||||
|
||||
for session_id, session in self.sessions.items():
|
||||
if now - session.created_at > self.max_session_duration:
|
||||
expired_sessions.append(session_id)
|
||||
|
||||
for session_id in expired_sessions:
|
||||
logger.info(f"Cleaning up expired session {session_id}")
|
||||
self.cleanup_session(session_id)
|
||||
|
||||
def cleanup_idle_sessions(self):
|
||||
"""Clean up sessions that have been idle too long"""
|
||||
with self.lock:
|
||||
now = time.time()
|
||||
idle_sessions = []
|
||||
|
||||
for session_id, session in self.sessions.items():
|
||||
if now - session.last_activity > self.max_idle_time:
|
||||
idle_sessions.append(session_id)
|
||||
|
||||
for session_id in idle_sessions:
|
||||
logger.info(f"Cleaning up idle session {session_id}")
|
||||
self.cleanup_session(session_id)
|
||||
|
||||
def cleanup_orphaned_files(self):
|
||||
"""Clean up orphaned files in session storage"""
|
||||
try:
|
||||
if not os.path.exists(self.session_storage_path):
|
||||
return
|
||||
|
||||
# Get all session directories
|
||||
session_dirs = set(os.listdir(self.session_storage_path))
|
||||
|
||||
# Get active session IDs
|
||||
with self.lock:
|
||||
active_sessions = set(self.sessions.keys())
|
||||
|
||||
# Find orphaned directories
|
||||
orphaned_dirs = session_dirs - active_sessions
|
||||
|
||||
# Clean up orphaned directories
|
||||
for dir_name in orphaned_dirs:
|
||||
dir_path = os.path.join(self.session_storage_path, dir_name)
|
||||
if os.path.isdir(dir_path):
|
||||
try:
|
||||
shutil.rmtree(dir_path)
|
||||
logger.info(f"Cleaned up orphaned session directory {dir_name}")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to remove orphaned directory {dir_path}: {e}")
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Error cleaning orphaned files: {e}")
|
||||
|
||||
def get_session_info(self, session_id: str) -> Optional[Dict[str, Any]]:
|
||||
"""Get detailed information about a session"""
|
||||
with self.lock:
|
||||
session = self.sessions.get(session_id)
|
||||
if not session:
|
||||
return None
|
||||
|
||||
return {
|
||||
'session_id': session.session_id,
|
||||
'user_id': session.user_id,
|
||||
'ip_address': session.ip_address,
|
||||
'created_at': datetime.fromtimestamp(session.created_at).isoformat(),
|
||||
'last_activity': datetime.fromtimestamp(session.last_activity).isoformat(),
|
||||
'duration_seconds': int(time.time() - session.created_at),
|
||||
'idle_seconds': int(time.time() - session.last_activity),
|
||||
'request_count': session.request_count,
|
||||
'resource_count': len(session.resources),
|
||||
'total_bytes_used': session.total_bytes_used,
|
||||
'active_streams': session.active_streams,
|
||||
'resources': [
|
||||
{
|
||||
'resource_id': r.resource_id,
|
||||
'resource_type': r.resource_type,
|
||||
'size_bytes': r.size_bytes,
|
||||
'created_at': datetime.fromtimestamp(r.created_at).isoformat(),
|
||||
'last_accessed': datetime.fromtimestamp(r.last_accessed).isoformat()
|
||||
}
|
||||
for r in session.resources.values()
|
||||
]
|
||||
}
|
||||
|
||||
def get_all_sessions_info(self) -> List[Dict[str, Any]]:
|
||||
"""Get information about all active sessions"""
|
||||
with self.lock:
|
||||
return [
|
||||
self.get_session_info(session_id)
|
||||
for session_id in self.sessions.keys()
|
||||
]
|
||||
|
||||
def get_stats(self) -> Dict[str, Any]:
|
||||
"""Get session manager statistics"""
|
||||
with self.lock:
|
||||
return {
|
||||
**self.stats,
|
||||
'uptime_seconds': int(time.time() - self.stats.get('start_time', time.time())),
|
||||
'avg_session_duration': self._calculate_avg_session_duration(),
|
||||
'avg_resources_per_session': self._calculate_avg_resources_per_session(),
|
||||
'total_storage_used': self._calculate_total_storage_used()
|
||||
}
|
||||
|
||||
def _calculate_avg_session_duration(self) -> float:
|
||||
"""Calculate average session duration"""
|
||||
if not self.sessions:
|
||||
return 0
|
||||
|
||||
total_duration = sum(
|
||||
time.time() - session.created_at
|
||||
for session in self.sessions.values()
|
||||
)
|
||||
return total_duration / len(self.sessions)
|
||||
|
||||
def _calculate_avg_resources_per_session(self) -> float:
|
||||
"""Calculate average resources per session"""
|
||||
if not self.sessions:
|
||||
return 0
|
||||
|
||||
total_resources = sum(
|
||||
len(session.resources)
|
||||
for session in self.sessions.values()
|
||||
)
|
||||
return total_resources / len(self.sessions)
|
||||
|
||||
def _calculate_total_storage_used(self) -> int:
|
||||
"""Calculate total storage used"""
|
||||
total = 0
|
||||
try:
|
||||
for root, dirs, files in os.walk(self.session_storage_path):
|
||||
for file in files:
|
||||
filepath = os.path.join(root, file)
|
||||
total += os.path.getsize(filepath)
|
||||
except Exception as e:
|
||||
logger.error(f"Error calculating storage used: {e}")
|
||||
return total
|
||||
|
||||
def export_metrics(self) -> Dict[str, Any]:
|
||||
"""Export metrics for monitoring"""
|
||||
with self.lock:
|
||||
return {
|
||||
'sessions': {
|
||||
'active': self.stats['active_sessions'],
|
||||
'total_created': self.stats['total_sessions_created'],
|
||||
'total_cleaned': self.stats['total_sessions_cleaned']
|
||||
},
|
||||
'resources': {
|
||||
'active': self.stats['active_resources'],
|
||||
'total_cleaned': self.stats['total_resources_cleaned'],
|
||||
'active_bytes': self.stats['active_bytes'],
|
||||
'total_bytes_cleaned': self.stats['total_bytes_cleaned']
|
||||
},
|
||||
'limits': {
|
||||
'max_session_duration': self.max_session_duration,
|
||||
'max_idle_time': self.max_idle_time,
|
||||
'max_resources_per_session': self.max_resources_per_session,
|
||||
'max_bytes_per_session': self.max_bytes_per_session
|
||||
}
|
||||
}
|
||||
|
||||
# Global session manager instance
|
||||
_session_manager = None
|
||||
_session_lock = Lock()
|
||||
|
||||
def get_session_manager(config: Dict[str, Any] = None) -> SessionManager:
|
||||
"""Get or create global session manager instance"""
|
||||
global _session_manager
|
||||
|
||||
with _session_lock:
|
||||
if _session_manager is None:
|
||||
_session_manager = SessionManager(config)
|
||||
return _session_manager
|
||||
|
||||
# Flask integration
|
||||
def init_app(app):
|
||||
"""Initialize session management for Flask app"""
|
||||
config = {
|
||||
'max_session_duration': app.config.get('MAX_SESSION_DURATION', 3600),
|
||||
'max_idle_time': app.config.get('MAX_SESSION_IDLE_TIME', 900),
|
||||
'max_resources_per_session': app.config.get('MAX_RESOURCES_PER_SESSION', 100),
|
||||
'max_bytes_per_session': app.config.get('MAX_BYTES_PER_SESSION', 100 * 1024 * 1024),
|
||||
'cleanup_interval': app.config.get('SESSION_CLEANUP_INTERVAL', 60),
|
||||
'session_storage_path': app.config.get('SESSION_STORAGE_PATH',
|
||||
os.path.join(app.config.get('UPLOAD_FOLDER', tempfile.gettempdir()), 'sessions'))
|
||||
}
|
||||
|
||||
manager = get_session_manager(config)
|
||||
app.session_manager = manager
|
||||
|
||||
# Add before_request handler
|
||||
@app.before_request
|
||||
def before_request_session():
|
||||
# Get or create session
|
||||
session_id = session.get('session_id')
|
||||
if not session_id:
|
||||
session_id = str(uuid.uuid4())
|
||||
session['session_id'] = session_id
|
||||
session.permanent = True
|
||||
|
||||
# Get session from manager
|
||||
user_session = manager.get_session(session_id)
|
||||
if not user_session:
|
||||
user_session = manager.create_session(
|
||||
session_id=session_id,
|
||||
ip_address=request.remote_addr,
|
||||
user_agent=request.headers.get('User-Agent')
|
||||
)
|
||||
|
||||
# Update activity
|
||||
manager.update_session_activity(session_id)
|
||||
|
||||
# Store in g for request access
|
||||
g.user_session = user_session
|
||||
g.session_manager = manager
|
||||
|
||||
# Add CLI commands
|
||||
@app.cli.command('sessions-list')
|
||||
def list_sessions_cmd():
|
||||
"""List all active sessions"""
|
||||
sessions = manager.get_all_sessions_info()
|
||||
for session_info in sessions:
|
||||
print(f"\nSession: {session_info['session_id']}")
|
||||
print(f" Created: {session_info['created_at']}")
|
||||
print(f" Last activity: {session_info['last_activity']}")
|
||||
print(f" Resources: {session_info['resource_count']}")
|
||||
print(f" Bytes used: {session_info['total_bytes_used']}")
|
||||
|
||||
@app.cli.command('sessions-cleanup')
|
||||
def cleanup_sessions_cmd():
|
||||
"""Manual session cleanup"""
|
||||
manager.cleanup_expired_sessions()
|
||||
manager.cleanup_idle_sessions()
|
||||
manager.cleanup_orphaned_files()
|
||||
print("Session cleanup completed")
|
||||
|
||||
@app.cli.command('sessions-stats')
|
||||
def session_stats_cmd():
|
||||
"""Show session statistics"""
|
||||
stats = manager.get_stats()
|
||||
print(json.dumps(stats, indent=2))
|
||||
|
||||
logger.info("Session management initialized")
|
||||
|
||||
# Decorator for session resource tracking
|
||||
def track_resource(resource_type: str):
|
||||
"""Decorator to track resources for a session"""
|
||||
def decorator(func):
|
||||
@wraps(func)
|
||||
def wrapper(*args, **kwargs):
|
||||
result = func(*args, **kwargs)
|
||||
|
||||
# Track resource if in request context
|
||||
if hasattr(g, 'user_session') and hasattr(g, 'session_manager'):
|
||||
if isinstance(result, (str, bytes)) or hasattr(result, 'filename'):
|
||||
# Determine path and size
|
||||
path = None
|
||||
size = 0
|
||||
|
||||
if isinstance(result, str) and os.path.exists(result):
|
||||
path = result
|
||||
size = os.path.getsize(result)
|
||||
elif hasattr(result, 'filename'):
|
||||
path = result.filename
|
||||
if os.path.exists(path):
|
||||
size = os.path.getsize(path)
|
||||
|
||||
# Add resource to session
|
||||
g.session_manager.add_resource(
|
||||
session_id=g.user_session.session_id,
|
||||
resource_type=resource_type,
|
||||
path=path,
|
||||
size_bytes=size
|
||||
)
|
||||
|
||||
return result
|
||||
return wrapper
|
||||
return decorator
|
776
setup-script.sh
776
setup-script.sh
@ -1,776 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
# Create necessary directories
|
||||
mkdir -p templates static/{css,js}
|
||||
|
||||
# Move HTML template to templates directory
|
||||
cat > templates/index.html << 'EOL'
|
||||
<!DOCTYPE html>
|
||||
<html lang="en">
|
||||
<head>
|
||||
<meta charset="UTF-8">
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||||
<title>Voice Language Translator</title>
|
||||
<link href="https://cdn.jsdelivr.net/npm/bootstrap@5.3.0-alpha1/dist/css/bootstrap.min.css" rel="stylesheet">
|
||||
<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.0.0/css/all.min.css">
|
||||
<style>
|
||||
body {
|
||||
padding-top: 20px;
|
||||
padding-bottom: 20px;
|
||||
background-color: #f8f9fa;
|
||||
}
|
||||
.record-btn {
|
||||
width: 80px;
|
||||
height: 80px;
|
||||
border-radius: 50%;
|
||||
display: flex;
|
||||
align-items: center;
|
||||
justify-content: center;
|
||||
font-size: 32px;
|
||||
margin: 20px auto;
|
||||
box-shadow: 0 4px 8px rgba(0, 0, 0, 0.1);
|
||||
transition: all 0.3s;
|
||||
}
|
||||
.record-btn:active {
|
||||
transform: scale(0.95);
|
||||
box-shadow: 0 2px 4px rgba(0, 0, 0, 0.1);
|
||||
}
|
||||
.recording {
|
||||
background-color: #dc3545 !important;
|
||||
animation: pulse 1.5s infinite;
|
||||
}
|
||||
@keyframes pulse {
|
||||
0% {
|
||||
transform: scale(1);
|
||||
}
|
||||
50% {
|
||||
transform: scale(1.05);
|
||||
}
|
||||
100% {
|
||||
transform: scale(1);
|
||||
}
|
||||
}
|
||||
.card {
|
||||
border-radius: 15px;
|
||||
box-shadow: 0 4px 6px rgba(0, 0, 0, 0.1);
|
||||
margin-bottom: 20px;
|
||||
}
|
||||
.card-header {
|
||||
border-radius: 15px 15px 0 0 !important;
|
||||
}
|
||||
.language-select {
|
||||
border-radius: 10px;
|
||||
padding: 10px;
|
||||
}
|
||||
.text-display {
|
||||
min-height: 100px;
|
||||
padding: 15px;
|
||||
background-color: #f8f9fa;
|
||||
border-radius: 10px;
|
||||
margin-bottom: 15px;
|
||||
}
|
||||
.btn-action {
|
||||
border-radius: 10px;
|
||||
padding: 8px 15px;
|
||||
margin: 5px;
|
||||
}
|
||||
.spinner-border {
|
||||
width: 1rem;
|
||||
height: 1rem;
|
||||
margin-right: 5px;
|
||||
}
|
||||
.status-indicator {
|
||||
font-size: 0.9rem;
|
||||
font-style: italic;
|
||||
color: #6c757d;
|
||||
}
|
||||
</style>
|
||||
</head>
|
||||
<body>
|
||||
<div class="container">
|
||||
<h1 class="text-center mb-4">Voice Language Translator</h1>
|
||||
<p class="text-center text-muted">Powered by Gemma 3, Whisper & Edge TTS</p>
|
||||
|
||||
<div class="row">
|
||||
<div class="col-md-6 mb-3">
|
||||
<div class="card">
|
||||
<div class="card-header bg-primary text-white">
|
||||
<h5 class="mb-0">Source</h5>
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<select id="sourceLanguage" class="form-select language-select mb-3">
|
||||
{% for language in languages %}
|
||||
<option value="{{ language }}">{{ language }}</option>
|
||||
{% endfor %}
|
||||
</select>
|
||||
<div class="text-display" id="sourceText">
|
||||
<p class="text-muted">Your transcribed text will appear here...</p>
|
||||
</div>
|
||||
<div class="d-flex justify-content-between">
|
||||
<button id="playSource" class="btn btn-outline-primary btn-action" disabled>
|
||||
<i class="fas fa-play"></i> Play
|
||||
</button>
|
||||
<button id="clearSource" class="btn btn-outline-secondary btn-action">
|
||||
<i class="fas fa-trash"></i> Clear
|
||||
</button>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="col-md-6 mb-3">
|
||||
<div class="card">
|
||||
<div class="card-header bg-success text-white">
|
||||
<h5 class="mb-0">Translation</h5>
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<select id="targetLanguage" class="form-select language-select mb-3">
|
||||
{% for language in languages %}
|
||||
<option value="{{ language }}">{{ language }}</option>
|
||||
{% endfor %}
|
||||
</select>
|
||||
<div class="text-display" id="translatedText">
|
||||
<p class="text-muted">Translation will appear here...</p>
|
||||
</div>
|
||||
<div class="d-flex justify-content-between">
|
||||
<button id="playTranslation" class="btn btn-outline-success btn-action" disabled>
|
||||
<i class="fas fa-play"></i> Play
|
||||
</button>
|
||||
<button id="clearTranslation" class="btn btn-outline-secondary btn-action">
|
||||
<i class="fas fa-trash"></i> Clear
|
||||
</button>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="text-center">
|
||||
<button id="recordBtn" class="btn btn-primary record-btn">
|
||||
<i class="fas fa-microphone"></i>
|
||||
</button>
|
||||
<p class="status-indicator" id="statusIndicator">Click to start recording</p>
|
||||
</div>
|
||||
|
||||
<div class="text-center mt-3">
|
||||
<button id="translateBtn" class="btn btn-success" disabled>
|
||||
<i class="fas fa-language"></i> Translate
|
||||
</button>
|
||||
</div>
|
||||
|
||||
<div class="mt-3">
|
||||
<div class="progress d-none" id="progressContainer">
|
||||
<div id="progressBar" class="progress-bar progress-bar-striped progress-bar-animated" role="progressbar" style="width: 0%"></div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<audio id="audioPlayer" style="display: none;"></audio>
|
||||
</div>
|
||||
|
||||
<script src="https://cdn.jsdelivr.net/npm/bootstrap@5.3.0-alpha1/dist/js/bootstrap.bundle.min.js"></script>
|
||||
<script>
|
||||
document.addEventListener('DOMContentLoaded', function() {
|
||||
// DOM elements
|
||||
const recordBtn = document.getElementById('recordBtn');
|
||||
const translateBtn = document.getElementById('translateBtn');
|
||||
const sourceText = document.getElementById('sourceText');
|
||||
const translatedText = document.getElementById('translatedText');
|
||||
const sourceLanguage = document.getElementById('sourceLanguage');
|
||||
const targetLanguage = document.getElementById('targetLanguage');
|
||||
const playSource = document.getElementById('playSource');
|
||||
const playTranslation = document.getElementById('playTranslation');
|
||||
const clearSource = document.getElementById('clearSource');
|
||||
const clearTranslation = document.getElementById('clearTranslation');
|
||||
const statusIndicator = document.getElementById('statusIndicator');
|
||||
const progressContainer = document.getElementById('progressContainer');
|
||||
const progressBar = document.getElementById('progressBar');
|
||||
const audioPlayer = document.getElementById('audioPlayer');
|
||||
|
||||
// Set initial values
|
||||
let isRecording = false;
|
||||
let mediaRecorder = null;
|
||||
let audioChunks = [];
|
||||
let currentSourceText = '';
|
||||
let currentTranslationText = '';
|
||||
|
||||
// Make sure target language is different from source
|
||||
if (targetLanguage.options[0].value === sourceLanguage.value) {
|
||||
targetLanguage.selectedIndex = 1;
|
||||
}
|
||||
|
||||
// Event listeners for language selection
|
||||
sourceLanguage.addEventListener('change', function() {
|
||||
if (targetLanguage.value === sourceLanguage.value) {
|
||||
for (let i = 0; i < targetLanguage.options.length; i++) {
|
||||
if (targetLanguage.options[i].value !== sourceLanguage.value) {
|
||||
targetLanguage.selectedIndex = i;
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
});
|
||||
|
||||
targetLanguage.addEventListener('change', function() {
|
||||
if (targetLanguage.value === sourceLanguage.value) {
|
||||
for (let i = 0; i < sourceLanguage.options.length; i++) {
|
||||
if (sourceLanguage.options[i].value !== targetLanguage.value) {
|
||||
sourceLanguage.selectedIndex = i;
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
});
|
||||
|
||||
// Record button click event
|
||||
recordBtn.addEventListener('click', function() {
|
||||
if (isRecording) {
|
||||
stopRecording();
|
||||
} else {
|
||||
startRecording();
|
||||
}
|
||||
});
|
||||
|
||||
// Function to start recording
|
||||
function startRecording() {
|
||||
navigator.mediaDevices.getUserMedia({ audio: true })
|
||||
.then(stream => {
|
||||
mediaRecorder = new MediaRecorder(stream);
|
||||
audioChunks = [];
|
||||
|
||||
mediaRecorder.addEventListener('dataavailable', event => {
|
||||
audioChunks.push(event.data);
|
||||
});
|
||||
|
||||
mediaRecorder.addEventListener('stop', () => {
|
||||
const audioBlob = new Blob(audioChunks, { type: 'audio/wav' });
|
||||
transcribeAudio(audioBlob);
|
||||
});
|
||||
|
||||
mediaRecorder.start();
|
||||
isRecording = true;
|
||||
recordBtn.classList.add('recording');
|
||||
recordBtn.classList.replace('btn-primary', 'btn-danger');
|
||||
recordBtn.innerHTML = '<i class="fas fa-stop"></i>';
|
||||
statusIndicator.textContent = 'Recording... Click to stop';
|
||||
})
|
||||
.catch(error => {
|
||||
console.error('Error accessing microphone:', error);
|
||||
alert('Error accessing microphone. Please make sure you have given permission for microphone access.');
|
||||
});
|
||||
}
|
||||
|
||||
// Function to stop recording
|
||||
function stopRecording() {
|
||||
mediaRecorder.stop();
|
||||
isRecording = false;
|
||||
recordBtn.classList.remove('recording');
|
||||
recordBtn.classList.replace('btn-danger', 'btn-primary');
|
||||
recordBtn.innerHTML = '<i class="fas fa-microphone"></i>';
|
||||
statusIndicator.textContent = 'Processing audio...';
|
||||
|
||||
// Stop all audio tracks
|
||||
mediaRecorder.stream.getTracks().forEach(track => track.stop());
|
||||
}
|
||||
|
||||
// Function to transcribe audio
|
||||
function transcribeAudio(audioBlob) {
|
||||
const formData = new FormData();
|
||||
formData.append('audio', audioBlob);
|
||||
formData.append('source_lang', sourceLanguage.value);
|
||||
|
||||
showProgress();
|
||||
|
||||
fetch('/transcribe', {
|
||||
method: 'POST',
|
||||
body: formData
|
||||
})
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
hideProgress();
|
||||
|
||||
if (data.success) {
|
||||
currentSourceText = data.text;
|
||||
sourceText.innerHTML = `<p>${data.text}</p>`;
|
||||
playSource.disabled = false;
|
||||
translateBtn.disabled = false;
|
||||
statusIndicator.textContent = 'Transcription complete';
|
||||
} else {
|
||||
sourceText.innerHTML = `<p class="text-danger">Error: ${data.error}</p>`;
|
||||
statusIndicator.textContent = 'Transcription failed';
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
hideProgress();
|
||||
console.error('Transcription error:', error);
|
||||
sourceText.innerHTML = `<p class="text-danger">Failed to transcribe audio. Please try again.</p>`;
|
||||
statusIndicator.textContent = 'Transcription failed';
|
||||
});
|
||||
}
|
||||
|
||||
// Translate button click event
|
||||
translateBtn.addEventListener('click', function() {
|
||||
if (!currentSourceText) {
|
||||
return;
|
||||
}
|
||||
|
||||
statusIndicator.textContent = 'Translating...';
|
||||
showProgress();
|
||||
|
||||
fetch('/translate', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
text: currentSourceText,
|
||||
source_lang: sourceLanguage.value,
|
||||
target_lang: targetLanguage.value
|
||||
})
|
||||
})
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
hideProgress();
|
||||
|
||||
if (data.success) {
|
||||
currentTranslationText = data.translation;
|
||||
translatedText.innerHTML = `<p>${data.translation}</p>`;
|
||||
playTranslation.disabled = false;
|
||||
statusIndicator.textContent = 'Translation complete';
|
||||
} else {
|
||||
translatedText.innerHTML = `<p class="text-danger">Error: ${data.error}</p>`;
|
||||
statusIndicator.textContent = 'Translation failed';
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
hideProgress();
|
||||
console.error('Translation error:', error);
|
||||
translatedText.innerHTML = `<p class="text-danger">Failed to translate. Please try again.</p>`;
|
||||
statusIndicator.textContent = 'Translation failed';
|
||||
});
|
||||
});
|
||||
|
||||
// Play source text
|
||||
playSource.addEventListener('click', function() {
|
||||
if (!currentSourceText) return;
|
||||
|
||||
playAudio(currentSourceText, sourceLanguage.value);
|
||||
statusIndicator.textContent = 'Playing source audio...';
|
||||
});
|
||||
|
||||
// Play translation
|
||||
playTranslation.addEventListener('click', function() {
|
||||
if (!currentTranslationText) return;
|
||||
|
||||
playAudio(currentTranslationText, targetLanguage.value);
|
||||
statusIndicator.textContent = 'Playing translation audio...';
|
||||
});
|
||||
|
||||
// Function to play audio via TTS
|
||||
function playAudio(text, language) {
|
||||
showProgress();
|
||||
|
||||
fetch('/speak', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
text: text,
|
||||
language: language
|
||||
})
|
||||
})
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
hideProgress();
|
||||
|
||||
if (data.success) {
|
||||
audioPlayer.src = data.audio_url;
|
||||
audioPlayer.onended = function() {
|
||||
statusIndicator.textContent = 'Ready';
|
||||
};
|
||||
audioPlayer.play();
|
||||
} else {
|
||||
statusIndicator.textContent = 'TTS failed';
|
||||
alert('Failed to play audio: ' + data.error);
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
hideProgress();
|
||||
console.error('TTS error:', error);
|
||||
statusIndicator.textContent = 'TTS failed';
|
||||
});
|
||||
}
|
||||
|
||||
// Clear buttons
|
||||
clearSource.addEventListener('click', function() {
|
||||
sourceText.innerHTML = '<p class="text-muted">Your transcribed text will appear here...</p>';
|
||||
currentSourceText = '';
|
||||
playSource.disabled = true;
|
||||
translateBtn.disabled = true;
|
||||
});
|
||||
|
||||
clearTranslation.addEventListener('click', function() {
|
||||
translatedText.innerHTML = '<p class="text-muted">Translation will appear here...</p>';
|
||||
currentTranslationText = '';
|
||||
playTranslation.disabled = true;
|
||||
});
|
||||
|
||||
// Progress indicator functions
|
||||
function showProgress() {
|
||||
progressContainer.classList.remove('d-none');
|
||||
let progress = 0;
|
||||
const interval = setInterval(() => {
|
||||
progress += 5;
|
||||
if (progress > 90) {
|
||||
clearInterval(interval);
|
||||
}
|
||||
progressBar.style.width = `${progress}%`;
|
||||
}, 100);
|
||||
progressBar.dataset.interval = interval;
|
||||
}
|
||||
|
||||
function hideProgress() {
|
||||
const interval = progressBar.dataset.interval;
|
||||
if (interval) {
|
||||
clearInterval(Number(interval));
|
||||
}
|
||||
progressBar.style.width = '100%';
|
||||
setTimeout(() => {
|
||||
progressContainer.classList.add('d-none');
|
||||
progressBar.style.width = '0%';
|
||||
}, 500);
|
||||
}
|
||||
});
|
||||
</script>
|
||||
</body>
|
||||
</html>
|
||||
EOL
|
||||
|
||||
# Create app.py
|
||||
cat > app.py << 'EOL'
|
||||
import os
|
||||
import time
|
||||
import tempfile
|
||||
import requests
|
||||
import json
|
||||
from flask import Flask, render_template, request, jsonify, Response, send_file
|
||||
import whisper
|
||||
import torch
|
||||
import ollama
|
||||
import logging
|
||||
|
||||
app = Flask(__name__)
|
||||
app.config['UPLOAD_FOLDER'] = tempfile.mkdtemp()
|
||||
app.config['TTS_SERVER'] = os.environ.get('TTS_SERVER_URL', 'http://localhost:5050/v1/audio/speech')
|
||||
app.config['TTS_API_KEY'] = os.environ.get('TTS_API_KEY', 'your_api_key_here')
|
||||
|
||||
# Add a route to check TTS server status
|
||||
@app.route('/check_tts_server', methods=['GET'])
|
||||
def check_tts_server():
|
||||
try:
|
||||
# Try a simple HTTP request to the TTS server
|
||||
response = requests.get(app.config['TTS_SERVER'].rsplit('/api/generate', 1)[0] + '/status', timeout=5)
|
||||
if response.status_code == 200:
|
||||
return jsonify({
|
||||
'status': 'online',
|
||||
'url': app.config['TTS_SERVER']
|
||||
})
|
||||
else:
|
||||
return jsonify({
|
||||
'status': 'error',
|
||||
'message': f'TTS server returned status code {response.status_code}',
|
||||
'url': app.config['TTS_SERVER']
|
||||
})
|
||||
except requests.exceptions.RequestException as e:
|
||||
return jsonify({
|
||||
'status': 'error',
|
||||
'message': f'Cannot connect to TTS server: {str(e)}',
|
||||
'url': app.config['TTS_SERVER']
|
||||
})
|
||||
|
||||
# Initialize logging
|
||||
logging.basicConfig(level=logging.INFO)
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
# Load Whisper model
|
||||
logger.info("Loading Whisper model...")
|
||||
whisper_model = whisper.load_model("base")
|
||||
logger.info("Whisper model loaded successfully")
|
||||
|
||||
# Supported languages
|
||||
SUPPORTED_LANGUAGES = {
|
||||
"ar": "Arabic",
|
||||
"hy": "Armenian",
|
||||
"az": "Azerbaijani",
|
||||
"en": "English",
|
||||
"fr": "French",
|
||||
"ka": "Georgian",
|
||||
"kk": "Kazakh",
|
||||
"zh": "Mandarin",
|
||||
"fa": "Farsi",
|
||||
"pt": "Portuguese",
|
||||
"ru": "Russian",
|
||||
"es": "Spanish",
|
||||
"tr": "Turkish",
|
||||
"uz": "Uzbek"
|
||||
}
|
||||
|
||||
# Map language names to language codes
|
||||
LANGUAGE_TO_CODE = {v: k for k, v in SUPPORTED_LANGUAGES.items()}
|
||||
|
||||
# Map language names to OpenAI TTS voice options
|
||||
LANGUAGE_TO_VOICE = {
|
||||
"Arabic": "alloy", # Using OpenAI general voices
|
||||
"Armenian": "echo", # as OpenAI doesn't have specific voices
|
||||
"Azerbaijani": "nova", # for all these languages
|
||||
"English": "echo", # We'll use the available voices
|
||||
"French": "alloy", # and rely on the translation being
|
||||
"Georgian": "fable", # in the correct language text
|
||||
"Kazakh": "onyx",
|
||||
"Mandarin": "shimmer",
|
||||
"Farsi": "nova",
|
||||
"Portuguese": "alloy",
|
||||
"Russian": "echo",
|
||||
"Spanish": "nova",
|
||||
"Turkish": "fable",
|
||||
"Uzbek": "onyx"
|
||||
}
|
||||
|
||||
@app.route('/')
|
||||
def index():
|
||||
return render_template('index.html', languages=sorted(SUPPORTED_LANGUAGES.values()))
|
||||
|
||||
@app.route('/transcribe', methods=['POST'])
|
||||
def transcribe():
|
||||
if 'audio' not in request.files:
|
||||
return jsonify({'error': 'No audio file provided'}), 400
|
||||
|
||||
audio_file = request.files['audio']
|
||||
source_lang = request.form.get('source_lang', '')
|
||||
|
||||
# Save the audio file temporarily
|
||||
temp_path = os.path.join(app.config['UPLOAD_FOLDER'], 'input_audio.wav')
|
||||
audio_file.save(temp_path)
|
||||
|
||||
try:
|
||||
# Use Whisper for transcription
|
||||
result = whisper_model.transcribe(
|
||||
temp_path,
|
||||
language=LANGUAGE_TO_CODE.get(source_lang, None)
|
||||
)
|
||||
transcribed_text = result["text"]
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'text': transcribed_text
|
||||
})
|
||||
except Exception as e:
|
||||
logger.error(f"Transcription error: {str(e)}")
|
||||
return jsonify({'error': f'Transcription failed: {str(e)}'}), 500
|
||||
finally:
|
||||
# Clean up the temporary file
|
||||
if os.path.exists(temp_path):
|
||||
os.remove(temp_path)
|
||||
|
||||
@app.route('/translate', methods=['POST'])
|
||||
def translate():
|
||||
try:
|
||||
data = request.json
|
||||
text = data.get('text', '')
|
||||
source_lang = data.get('source_lang', '')
|
||||
target_lang = data.get('target_lang', '')
|
||||
|
||||
if not text or not source_lang or not target_lang:
|
||||
return jsonify({'error': 'Missing required parameters'}), 400
|
||||
|
||||
# Create a prompt for Gemma 3 translation
|
||||
prompt = f"""
|
||||
Translate the following text from {source_lang} to {target_lang}:
|
||||
|
||||
"{text}"
|
||||
|
||||
Provide only the translation without any additional text.
|
||||
"""
|
||||
|
||||
# Use Ollama to interact with Gemma 3
|
||||
response = ollama.chat(
|
||||
model="gemma3",
|
||||
messages=[
|
||||
{
|
||||
"role": "user",
|
||||
"content": prompt
|
||||
}
|
||||
]
|
||||
)
|
||||
|
||||
translated_text = response['message']['content'].strip()
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'translation': translated_text
|
||||
})
|
||||
except Exception as e:
|
||||
logger.error(f"Translation error: {str(e)}")
|
||||
return jsonify({'error': f'Translation failed: {str(e)}'}), 500
|
||||
|
||||
@app.route('/speak', methods=['POST'])
|
||||
def speak():
|
||||
try:
|
||||
data = request.json
|
||||
text = data.get('text', '')
|
||||
language = data.get('language', '')
|
||||
|
||||
if not text or not language:
|
||||
return jsonify({'error': 'Missing required parameters'}), 400
|
||||
|
||||
voice = LANGUAGE_TO_VOICE.get(language)
|
||||
if not voice:
|
||||
return jsonify({'error': 'Unsupported language for TTS'}), 400
|
||||
|
||||
# Get TTS server URL from environment or config
|
||||
tts_server_url = app.config['TTS_SERVER']
|
||||
|
||||
try:
|
||||
# Request TTS from the Edge TTS server
|
||||
logger.info(f"Sending TTS request to {tts_server_url}")
|
||||
tts_response = requests.post(
|
||||
tts_server_url,
|
||||
json={
|
||||
'text': text,
|
||||
'voice': voice,
|
||||
'output_format': 'mp3'
|
||||
},
|
||||
timeout=10 # Add timeout
|
||||
)
|
||||
|
||||
logger.info(f"TTS response status: {tts_response.status_code}")
|
||||
|
||||
if tts_response.status_code != 200:
|
||||
error_msg = f'TTS request failed with status {tts_response.status_code}'
|
||||
logger.error(error_msg)
|
||||
|
||||
# Try to get error details from response if possible
|
||||
try:
|
||||
error_details = tts_response.json()
|
||||
logger.error(f"Error details: {error_details}")
|
||||
except:
|
||||
pass
|
||||
|
||||
return jsonify({'error': error_msg}), 500
|
||||
|
||||
# The response contains the audio data directly
|
||||
temp_audio_path = os.path.join(app.config['UPLOAD_FOLDER'], f'output_{int(time.time())}.mp3')
|
||||
with open(temp_audio_path, 'wb') as f:
|
||||
f.write(tts_response.content)
|
||||
|
||||
return jsonify({
|
||||
'success': True,
|
||||
'audio_url': f'/get_audio/{os.path.basename(temp_audio_path)}'
|
||||
})
|
||||
except requests.exceptions.RequestException as e:
|
||||
error_msg = f'Failed to connect to TTS server: {str(e)}'
|
||||
logger.error(error_msg)
|
||||
return jsonify({'error': error_msg}), 500
|
||||
except Exception as e:
|
||||
logger.error(f"TTS error: {str(e)}")
|
||||
return jsonify({'error': f'TTS failed: {str(e)}'}), 500
|
||||
|
||||
@app.route('/get_audio/<filename>')
|
||||
def get_audio(filename):
|
||||
try:
|
||||
file_path = os.path.join(app.config['UPLOAD_FOLDER'], filename)
|
||||
return send_file(file_path, mimetype='audio/mpeg')
|
||||
except Exception as e:
|
||||
logger.error(f"Audio retrieval error: {str(e)}")
|
||||
return jsonify({'error': f'Audio retrieval failed: {str(e)}'}), 500
|
||||
|
||||
if __name__ == '__main__':
|
||||
app.run(host='0.0.0.0', port=8000, debug=True)
|
||||
EOL
|
||||
|
||||
# Create requirements.txt
|
||||
cat > requirements.txt << 'EOL'
|
||||
flask==2.3.2
|
||||
requests==2.31.0
|
||||
openai-whisper==20231117
|
||||
torch==2.1.0
|
||||
ollama==0.1.5
|
||||
EOL
|
||||
|
||||
# Create README.md
|
||||
cat > README.md << 'EOL'
|
||||
# Voice Language Translator
|
||||
|
||||
A mobile-friendly web application that translates spoken language between multiple languages using:
|
||||
- Gemma 3 open-source LLM via Ollama for translation
|
||||
- OpenAI Whisper for speech-to-text
|
||||
- OpenAI Edge TTS for text-to-speech
|
||||
|
||||
## Supported Languages
|
||||
|
||||
- Arabic
|
||||
- Armenian
|
||||
- Azerbaijani
|
||||
- English
|
||||
- French
|
||||
- Georgian
|
||||
- Kazakh
|
||||
- Mandarin
|
||||
- Farsi
|
||||
- Portuguese
|
||||
- Russian
|
||||
- Spanish
|
||||
- Turkish
|
||||
- Uzbek
|
||||
|
||||
## Setup Instructions
|
||||
|
||||
1. Install the required Python packages:
|
||||
```
|
||||
pip install -r requirements.txt
|
||||
```
|
||||
|
||||
2. Make sure you have Ollama installed and the Gemma 3 model loaded:
|
||||
```
|
||||
ollama pull gemma3
|
||||
```
|
||||
|
||||
3. Ensure your OpenAI Edge TTS server is running on port 5050.
|
||||
|
||||
4. Run the application:
|
||||
```
|
||||
python app.py
|
||||
```
|
||||
|
||||
5. Open your browser and navigate to:
|
||||
```
|
||||
http://localhost:8000
|
||||
```
|
||||
|
||||
## Usage
|
||||
|
||||
1. Select your source language from the dropdown menu
|
||||
2. Press the microphone button and speak
|
||||
3. Press the button again to stop recording
|
||||
4. Wait for the transcription to complete
|
||||
5. Select your target language
|
||||
6. Press the "Translate" button
|
||||
7. Use the play buttons to hear the original or translated text
|
||||
|
||||
## Technical Details
|
||||
|
||||
- The app uses Flask for the web server
|
||||
- Audio is processed client-side using the MediaRecorder API
|
||||
- Whisper for speech recognition with language hints
|
||||
- Ollama provides access to the Gemma 3 model for translation
|
||||
- OpenAI Edge TTS delivers natural-sounding speech output
|
||||
|
||||
## Mobile Support
|
||||
|
||||
The interface is fully responsive and designed to work well on mobile devices.
|
||||
EOL
|
||||
|
||||
# Make the script executable
|
||||
chmod +x app.py
|
||||
|
||||
echo "Setup complete! Run the app with: python app.py"
|
156
setup_databases.sh
Executable file
156
setup_databases.sh
Executable file
@ -0,0 +1,156 @@
|
||||
#!/bin/bash
|
||||
# Setup script for Redis and PostgreSQL
|
||||
|
||||
echo "Talk2Me Database Setup Script"
|
||||
echo "============================="
|
||||
|
||||
# Check if running as root
|
||||
if [ "$EUID" -eq 0 ]; then
|
||||
echo "Please do not run this script as root"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Function to check if command exists
|
||||
command_exists() {
|
||||
command -v "$1" >/dev/null 2>&1
|
||||
}
|
||||
|
||||
# Check for PostgreSQL
|
||||
echo "Checking PostgreSQL installation..."
|
||||
if command_exists psql; then
|
||||
echo "✓ PostgreSQL is installed"
|
||||
psql --version
|
||||
else
|
||||
echo "✗ PostgreSQL is not installed"
|
||||
echo "Please install PostgreSQL first:"
|
||||
echo " Ubuntu/Debian: sudo apt-get install postgresql postgresql-contrib"
|
||||
echo " MacOS: brew install postgresql"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Check for Redis
|
||||
echo ""
|
||||
echo "Checking Redis installation..."
|
||||
if command_exists redis-cli; then
|
||||
echo "✓ Redis is installed"
|
||||
redis-cli --version
|
||||
else
|
||||
echo "✗ Redis is not installed"
|
||||
echo "Please install Redis first:"
|
||||
echo " Ubuntu/Debian: sudo apt-get install redis-server"
|
||||
echo " MacOS: brew install redis"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Check if Redis is running
|
||||
echo ""
|
||||
echo "Checking Redis server status..."
|
||||
if redis-cli ping > /dev/null 2>&1; then
|
||||
echo "✓ Redis server is running"
|
||||
else
|
||||
echo "✗ Redis server is not running"
|
||||
echo "Starting Redis server..."
|
||||
if command_exists systemctl; then
|
||||
sudo systemctl start redis
|
||||
else
|
||||
redis-server --daemonize yes
|
||||
fi
|
||||
sleep 2
|
||||
if redis-cli ping > /dev/null 2>&1; then
|
||||
echo "✓ Redis server started successfully"
|
||||
else
|
||||
echo "✗ Failed to start Redis server"
|
||||
exit 1
|
||||
fi
|
||||
fi
|
||||
|
||||
# Create PostgreSQL database
|
||||
echo ""
|
||||
echo "Setting up PostgreSQL database..."
|
||||
read -p "Enter PostgreSQL username (default: $USER): " PG_USER
|
||||
PG_USER=${PG_USER:-$USER}
|
||||
|
||||
read -p "Enter database name (default: talk2me): " DB_NAME
|
||||
DB_NAME=${DB_NAME:-talk2me}
|
||||
|
||||
# Check if database exists
|
||||
if psql -U "$PG_USER" -lqt | cut -d \| -f 1 | grep -qw "$DB_NAME"; then
|
||||
echo "Database '$DB_NAME' already exists"
|
||||
read -p "Do you want to drop and recreate it? (y/N): " RECREATE
|
||||
if [[ $RECREATE =~ ^[Yy]$ ]]; then
|
||||
echo "Dropping database '$DB_NAME'..."
|
||||
psql -U "$PG_USER" -c "DROP DATABASE $DB_NAME;"
|
||||
echo "Creating database '$DB_NAME'..."
|
||||
psql -U "$PG_USER" -c "CREATE DATABASE $DB_NAME;"
|
||||
fi
|
||||
else
|
||||
echo "Creating database '$DB_NAME'..."
|
||||
psql -U "$PG_USER" -c "CREATE DATABASE $DB_NAME;"
|
||||
fi
|
||||
|
||||
# Create .env file if it doesn't exist
|
||||
if [ ! -f .env ]; then
|
||||
echo ""
|
||||
echo "Creating .env file..."
|
||||
cat > .env << EOF
|
||||
# Database Configuration
|
||||
DATABASE_URL=postgresql://$PG_USER@localhost/$DB_NAME
|
||||
SQLALCHEMY_DATABASE_URI=postgresql://$PG_USER@localhost/$DB_NAME
|
||||
|
||||
# Redis Configuration
|
||||
REDIS_URL=redis://localhost:6379/0
|
||||
REDIS_DECODE_RESPONSES=false
|
||||
REDIS_MAX_CONNECTIONS=50
|
||||
|
||||
# Flask Configuration
|
||||
FLASK_ENV=development
|
||||
SECRET_KEY=$(openssl rand -base64 32)
|
||||
ADMIN_TOKEN=$(openssl rand -base64 24)
|
||||
|
||||
# TTS Configuration
|
||||
TTS_SERVER_URL=http://localhost:5050/v1/audio/speech
|
||||
TTS_API_KEY=your-tts-api-key-here
|
||||
|
||||
# Whisper Configuration
|
||||
WHISPER_MODEL_SIZE=base
|
||||
WHISPER_DEVICE=auto
|
||||
|
||||
# Ollama Configuration
|
||||
OLLAMA_HOST=http://localhost:11434
|
||||
OLLAMA_MODEL=gemma3:27b
|
||||
EOF
|
||||
echo "✓ .env file created"
|
||||
echo "Please update the TTS_API_KEY in .env file"
|
||||
else
|
||||
echo "✓ .env file already exists"
|
||||
fi
|
||||
|
||||
# Install Python dependencies
|
||||
echo ""
|
||||
echo "Installing Python dependencies..."
|
||||
if [ -f "requirements.txt" ]; then
|
||||
pip install -r requirements.txt
|
||||
echo "✓ Python dependencies installed"
|
||||
else
|
||||
echo "✗ requirements.txt not found"
|
||||
fi
|
||||
|
||||
# Initialize database
|
||||
echo ""
|
||||
echo "Initializing database..."
|
||||
python database_init.py
|
||||
|
||||
echo ""
|
||||
echo "Setup complete!"
|
||||
echo ""
|
||||
echo "Next steps:"
|
||||
echo "1. Update the TTS_API_KEY in .env file"
|
||||
echo "2. Run 'python migrations.py init' to initialize migrations"
|
||||
echo "3. Run 'python migrations.py create \"Initial migration\"' to create first migration"
|
||||
echo "4. Run 'python migrations.py run' to apply migrations"
|
||||
echo "5. Backup your current app.py and rename app_with_db.py to app.py"
|
||||
echo "6. Start the application with 'python app.py'"
|
||||
echo ""
|
||||
echo "To run Redis and PostgreSQL as services:"
|
||||
echo " Redis: sudo systemctl enable redis && sudo systemctl start redis"
|
||||
echo " PostgreSQL: sudo systemctl enable postgresql && sudo systemctl start postgresql"
|
218
static/check-pwa-status.html
Normal file
218
static/check-pwa-status.html
Normal file
@ -0,0 +1,218 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en">
|
||||
<head>
|
||||
<meta charset="UTF-8">
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||||
<title>PWA Status Check - Talk2Me</title>
|
||||
<style>
|
||||
body {
|
||||
font-family: Arial, sans-serif;
|
||||
padding: 20px;
|
||||
max-width: 600px;
|
||||
margin: 0 auto;
|
||||
}
|
||||
.status {
|
||||
padding: 10px;
|
||||
margin: 10px 0;
|
||||
border-radius: 5px;
|
||||
}
|
||||
.success {
|
||||
background: #d4edda;
|
||||
color: #155724;
|
||||
}
|
||||
.error {
|
||||
background: #f8d7da;
|
||||
color: #721c24;
|
||||
}
|
||||
.warning {
|
||||
background: #fff3cd;
|
||||
color: #856404;
|
||||
}
|
||||
.info {
|
||||
background: #d1ecf1;
|
||||
color: #0c5460;
|
||||
}
|
||||
button {
|
||||
padding: 10px 20px;
|
||||
margin: 10px 0;
|
||||
font-size: 16px;
|
||||
cursor: pointer;
|
||||
}
|
||||
pre {
|
||||
background: #f5f5f5;
|
||||
padding: 10px;
|
||||
overflow-x: auto;
|
||||
}
|
||||
</style>
|
||||
</head>
|
||||
<body>
|
||||
<h1>Talk2Me PWA Status Check</h1>
|
||||
|
||||
<div id="results"></div>
|
||||
|
||||
<h2>Actions</h2>
|
||||
<button onclick="testInstall()">Test Install Prompt</button>
|
||||
<button onclick="clearPWA()">Clear PWA Data</button>
|
||||
<button onclick="location.reload()">Refresh Page</button>
|
||||
|
||||
<script>
|
||||
const results = document.getElementById('results');
|
||||
|
||||
function addResult(message, type = 'info') {
|
||||
const div = document.createElement('div');
|
||||
div.className = `status ${type}`;
|
||||
div.textContent = message;
|
||||
results.appendChild(div);
|
||||
}
|
||||
|
||||
// Check HTTPS
|
||||
if (location.protocol === 'https:' || location.hostname === 'localhost') {
|
||||
addResult('✓ HTTPS enabled', 'success');
|
||||
} else {
|
||||
addResult('✗ HTTPS required for PWA', 'error');
|
||||
}
|
||||
|
||||
// Check Service Worker support
|
||||
if ('serviceWorker' in navigator) {
|
||||
addResult('✓ Service Worker supported', 'success');
|
||||
|
||||
// Check registration
|
||||
navigator.serviceWorker.getRegistration().then(reg => {
|
||||
if (reg) {
|
||||
addResult(`✓ Service Worker registered (scope: ${reg.scope})`, 'success');
|
||||
if (reg.active) {
|
||||
addResult('✓ Service Worker is active', 'success');
|
||||
} else if (reg.installing) {
|
||||
addResult('⚠ Service Worker is installing', 'warning');
|
||||
} else if (reg.waiting) {
|
||||
addResult('⚠ Service Worker is waiting', 'warning');
|
||||
}
|
||||
} else {
|
||||
addResult('✗ Service Worker not registered', 'error');
|
||||
}
|
||||
});
|
||||
} else {
|
||||
addResult('✗ Service Worker not supported', 'error');
|
||||
}
|
||||
|
||||
// Check manifest
|
||||
const manifestLink = document.querySelector('link[rel="manifest"]');
|
||||
if (manifestLink) {
|
||||
addResult('✓ Manifest link found', 'success');
|
||||
|
||||
fetch(manifestLink.href)
|
||||
.then(r => r.json())
|
||||
.then(manifest => {
|
||||
addResult('✓ Manifest loaded successfully', 'success');
|
||||
|
||||
// Check required fields
|
||||
const required = ['name', 'short_name', 'start_url', 'display', 'icons'];
|
||||
required.forEach(field => {
|
||||
if (manifest[field]) {
|
||||
addResult(`✓ Manifest has ${field}`, 'success');
|
||||
} else {
|
||||
addResult(`✗ Manifest missing ${field}`, 'error');
|
||||
}
|
||||
});
|
||||
|
||||
// Check icons
|
||||
if (manifest.icons && manifest.icons.length > 0) {
|
||||
const has192 = manifest.icons.some(i => i.sizes && i.sizes.includes('192'));
|
||||
const has512 = manifest.icons.some(i => i.sizes && i.sizes.includes('512'));
|
||||
|
||||
if (has192) addResult('✓ Has 192x192 icon', 'success');
|
||||
else addResult('✗ Missing 192x192 icon', 'error');
|
||||
|
||||
if (has512) addResult('✓ Has 512x512 icon', 'success');
|
||||
else addResult('⚠ Missing 512x512 icon (recommended)', 'warning');
|
||||
|
||||
// Check icon purposes
|
||||
manifest.icons.forEach((icon, i) => {
|
||||
if (icon.purpose) {
|
||||
addResult(`Icon ${i+1}: purpose="${icon.purpose}"`, 'info');
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
// Show manifest content
|
||||
const pre = document.createElement('pre');
|
||||
pre.textContent = JSON.stringify(manifest, null, 2);
|
||||
results.appendChild(pre);
|
||||
})
|
||||
.catch(err => {
|
||||
addResult(`✗ Failed to load manifest: ${err}`, 'error');
|
||||
});
|
||||
} else {
|
||||
addResult('✗ No manifest link found', 'error');
|
||||
}
|
||||
|
||||
// Check if already installed
|
||||
if (window.matchMedia('(display-mode: standalone)').matches) {
|
||||
addResult('✓ App is running in standalone mode (already installed)', 'success');
|
||||
} else {
|
||||
addResult('App is running in browser mode', 'info');
|
||||
}
|
||||
|
||||
// Listen for install prompt
|
||||
let deferredPrompt;
|
||||
window.addEventListener('beforeinstallprompt', (e) => {
|
||||
e.preventDefault();
|
||||
deferredPrompt = e;
|
||||
addResult('✓ Install prompt is available!', 'success');
|
||||
addResult('Chrome recognizes this as an installable PWA', 'success');
|
||||
});
|
||||
|
||||
// Check Chrome version
|
||||
const userAgent = navigator.userAgent;
|
||||
const chromeMatch = userAgent.match(/Chrome\/(\d+)/);
|
||||
if (chromeMatch) {
|
||||
const version = parseInt(chromeMatch[1]);
|
||||
addResult(`Chrome version: ${version}`, 'info');
|
||||
if (version < 90) {
|
||||
addResult('⚠ Chrome version is old, consider updating', 'warning');
|
||||
}
|
||||
}
|
||||
|
||||
// Test install
|
||||
function testInstall() {
|
||||
if (deferredPrompt) {
|
||||
deferredPrompt.prompt();
|
||||
deferredPrompt.userChoice.then((choiceResult) => {
|
||||
if (choiceResult.outcome === 'accepted') {
|
||||
addResult('✓ User accepted the install prompt', 'success');
|
||||
} else {
|
||||
addResult('User dismissed the install prompt', 'warning');
|
||||
}
|
||||
deferredPrompt = null;
|
||||
});
|
||||
} else {
|
||||
addResult('No install prompt available. Chrome may not recognize this as installable.', 'error');
|
||||
addResult('Try: Menu (⋮) → Add to Home screen', 'info');
|
||||
}
|
||||
}
|
||||
|
||||
// Clear PWA data
|
||||
function clearPWA() {
|
||||
if ('serviceWorker' in navigator) {
|
||||
navigator.serviceWorker.getRegistrations().then(function(registrations) {
|
||||
for(let registration of registrations) {
|
||||
registration.unregister();
|
||||
}
|
||||
addResult('Service Workers unregistered', 'info');
|
||||
});
|
||||
}
|
||||
|
||||
if ('caches' in window) {
|
||||
caches.keys().then(function(names) {
|
||||
for (let name of names) {
|
||||
caches.delete(name);
|
||||
}
|
||||
addResult('Caches cleared', 'info');
|
||||
});
|
||||
}
|
||||
|
||||
addResult('PWA data cleared. Reload the page to re-register.', 'info');
|
||||
}
|
||||
</script>
|
||||
</body>
|
||||
</html>
|
@ -0,0 +1,594 @@
|
||||
/* Main styles for Talk2Me application */
|
||||
|
||||
/* Loading animations */
|
||||
.loading-dots {
|
||||
display: inline-flex;
|
||||
align-items: center;
|
||||
gap: 4px;
|
||||
}
|
||||
|
||||
.loading-dots span {
|
||||
width: 8px;
|
||||
height: 8px;
|
||||
border-radius: 50%;
|
||||
background-color: #007bff;
|
||||
animation: dotPulse 1.4s infinite ease-in-out both;
|
||||
}
|
||||
|
||||
.loading-dots span:nth-child(1) {
|
||||
animation-delay: -0.32s;
|
||||
}
|
||||
|
||||
.loading-dots span:nth-child(2) {
|
||||
animation-delay: -0.16s;
|
||||
}
|
||||
|
||||
@keyframes dotPulse {
|
||||
0%, 80%, 100% {
|
||||
transform: scale(0);
|
||||
opacity: 0.5;
|
||||
}
|
||||
40% {
|
||||
transform: scale(1);
|
||||
opacity: 1;
|
||||
}
|
||||
}
|
||||
|
||||
/* Wave animation for recording */
|
||||
.recording-wave {
|
||||
position: relative;
|
||||
display: inline-block;
|
||||
width: 40px;
|
||||
height: 40px;
|
||||
}
|
||||
|
||||
.recording-wave span {
|
||||
position: absolute;
|
||||
bottom: 0;
|
||||
width: 4px;
|
||||
height: 100%;
|
||||
background: #fff;
|
||||
border-radius: 2px;
|
||||
animation: wave 1.2s linear infinite;
|
||||
}
|
||||
|
||||
.recording-wave span:nth-child(1) {
|
||||
left: 0;
|
||||
animation-delay: 0s;
|
||||
}
|
||||
|
||||
.recording-wave span:nth-child(2) {
|
||||
left: 8px;
|
||||
animation-delay: -1.1s;
|
||||
}
|
||||
|
||||
.recording-wave span:nth-child(3) {
|
||||
left: 16px;
|
||||
animation-delay: -1s;
|
||||
}
|
||||
|
||||
.recording-wave span:nth-child(4) {
|
||||
left: 24px;
|
||||
animation-delay: -0.9s;
|
||||
}
|
||||
|
||||
.recording-wave span:nth-child(5) {
|
||||
left: 32px;
|
||||
animation-delay: -0.8s;
|
||||
}
|
||||
|
||||
@keyframes wave {
|
||||
0%, 40%, 100% {
|
||||
transform: scaleY(0.4);
|
||||
}
|
||||
20% {
|
||||
transform: scaleY(1);
|
||||
}
|
||||
}
|
||||
|
||||
/* Spinner animation */
|
||||
.spinner-custom {
|
||||
width: 40px;
|
||||
height: 40px;
|
||||
position: relative;
|
||||
display: inline-block;
|
||||
}
|
||||
|
||||
.spinner-custom::before {
|
||||
content: '';
|
||||
position: absolute;
|
||||
width: 100%;
|
||||
height: 100%;
|
||||
border-radius: 50%;
|
||||
border: 3px solid rgba(0, 123, 255, 0.2);
|
||||
}
|
||||
|
||||
.spinner-custom::after {
|
||||
content: '';
|
||||
position: absolute;
|
||||
width: 100%;
|
||||
height: 100%;
|
||||
border-radius: 50%;
|
||||
border: 3px solid transparent;
|
||||
border-top-color: #007bff;
|
||||
animation: spin 0.8s linear infinite;
|
||||
}
|
||||
|
||||
@keyframes spin {
|
||||
to {
|
||||
transform: rotate(360deg);
|
||||
}
|
||||
}
|
||||
|
||||
/* Translation animation */
|
||||
.translation-animation {
|
||||
position: relative;
|
||||
display: inline-flex;
|
||||
align-items: center;
|
||||
gap: 10px;
|
||||
}
|
||||
|
||||
.translation-animation .arrow {
|
||||
width: 30px;
|
||||
height: 2px;
|
||||
background: #28a745;
|
||||
position: relative;
|
||||
animation: moveArrow 1.5s infinite;
|
||||
}
|
||||
|
||||
.translation-animation .arrow::after {
|
||||
content: '';
|
||||
position: absolute;
|
||||
right: -8px;
|
||||
top: -4px;
|
||||
width: 0;
|
||||
height: 0;
|
||||
border-left: 8px solid #28a745;
|
||||
border-top: 5px solid transparent;
|
||||
border-bottom: 5px solid transparent;
|
||||
}
|
||||
|
||||
@keyframes moveArrow {
|
||||
0%, 100% {
|
||||
transform: translateX(0);
|
||||
}
|
||||
50% {
|
||||
transform: translateX(10px);
|
||||
}
|
||||
}
|
||||
|
||||
/* Processing text animation */
|
||||
.processing-text {
|
||||
display: inline-block;
|
||||
position: relative;
|
||||
font-style: italic;
|
||||
color: #6c757d;
|
||||
}
|
||||
|
||||
.processing-text::after {
|
||||
content: '';
|
||||
position: absolute;
|
||||
bottom: -2px;
|
||||
left: 0;
|
||||
width: 100%;
|
||||
height: 2px;
|
||||
background: linear-gradient(90deg,
|
||||
transparent 0%,
|
||||
#007bff 50%,
|
||||
transparent 100%);
|
||||
animation: processLine 2s linear infinite;
|
||||
}
|
||||
|
||||
@keyframes processLine {
|
||||
0% {
|
||||
transform: translateX(-100%);
|
||||
}
|
||||
100% {
|
||||
transform: translateX(100%);
|
||||
}
|
||||
}
|
||||
|
||||
/* Fade in animation for results */
|
||||
.fade-in {
|
||||
animation: fadeIn 0.5s ease-in;
|
||||
}
|
||||
|
||||
@keyframes fadeIn {
|
||||
from {
|
||||
opacity: 0;
|
||||
transform: translateY(10px);
|
||||
}
|
||||
to {
|
||||
opacity: 1;
|
||||
transform: translateY(0);
|
||||
}
|
||||
}
|
||||
|
||||
/* Pulse animation for buttons */
|
||||
.btn-pulse {
|
||||
animation: pulse 2s infinite;
|
||||
}
|
||||
|
||||
@keyframes pulse {
|
||||
0% {
|
||||
box-shadow: 0 0 0 0 rgba(0, 123, 255, 0.7);
|
||||
}
|
||||
70% {
|
||||
box-shadow: 0 0 0 10px rgba(0, 123, 255, 0);
|
||||
}
|
||||
100% {
|
||||
box-shadow: 0 0 0 0 rgba(0, 123, 255, 0);
|
||||
}
|
||||
}
|
||||
|
||||
/* Loading overlay */
|
||||
.loading-overlay {
|
||||
position: fixed;
|
||||
top: 0;
|
||||
left: 0;
|
||||
right: 0;
|
||||
bottom: 0;
|
||||
background: rgba(255, 255, 255, 0.9);
|
||||
display: flex;
|
||||
align-items: center;
|
||||
justify-content: center;
|
||||
z-index: 9999;
|
||||
opacity: 0;
|
||||
pointer-events: none;
|
||||
transition: opacity 0.3s ease;
|
||||
}
|
||||
|
||||
.loading-overlay.active {
|
||||
opacity: 1;
|
||||
pointer-events: all;
|
||||
}
|
||||
|
||||
.loading-content {
|
||||
text-align: center;
|
||||
}
|
||||
|
||||
.loading-content .spinner-custom {
|
||||
margin-bottom: 20px;
|
||||
}
|
||||
|
||||
/* Status indicator animations */
|
||||
.status-indicator {
|
||||
transition: all 0.3s ease;
|
||||
}
|
||||
|
||||
.status-indicator.processing {
|
||||
font-weight: 500;
|
||||
color: #007bff;
|
||||
}
|
||||
|
||||
.status-indicator.success {
|
||||
color: #28a745;
|
||||
}
|
||||
|
||||
.status-indicator.error {
|
||||
color: #dc3545;
|
||||
}
|
||||
|
||||
/* Card loading state */
|
||||
.card-loading {
|
||||
position: relative;
|
||||
overflow: hidden;
|
||||
}
|
||||
|
||||
.card-loading::after {
|
||||
content: '';
|
||||
position: absolute;
|
||||
top: 0;
|
||||
left: -100%;
|
||||
width: 100%;
|
||||
height: 100%;
|
||||
background: linear-gradient(
|
||||
90deg,
|
||||
transparent,
|
||||
rgba(255, 255, 255, 0.4),
|
||||
transparent
|
||||
);
|
||||
animation: shimmer 2s infinite;
|
||||
}
|
||||
|
||||
@keyframes shimmer {
|
||||
100% {
|
||||
left: 100%;
|
||||
}
|
||||
}
|
||||
|
||||
/* Text skeleton loader */
|
||||
.skeleton-loader {
|
||||
background: #eee;
|
||||
background: linear-gradient(90deg, #eee 25%, #f5f5f5 50%, #eee 75%);
|
||||
background-size: 200% 100%;
|
||||
animation: loading 1.5s infinite;
|
||||
border-radius: 4px;
|
||||
height: 20px;
|
||||
margin: 10px 0;
|
||||
}
|
||||
|
||||
@keyframes loading {
|
||||
0% {
|
||||
background-position: 200% 0;
|
||||
}
|
||||
100% {
|
||||
background-position: -200% 0;
|
||||
}
|
||||
}
|
||||
|
||||
/* Audio playing animation */
|
||||
.audio-playing {
|
||||
display: inline-flex;
|
||||
align-items: flex-end;
|
||||
gap: 2px;
|
||||
height: 20px;
|
||||
}
|
||||
|
||||
.audio-playing span {
|
||||
width: 3px;
|
||||
background: #28a745;
|
||||
animation: audioBar 0.5s ease-in-out infinite alternate;
|
||||
}
|
||||
|
||||
.audio-playing span:nth-child(1) {
|
||||
height: 40%;
|
||||
animation-delay: 0s;
|
||||
}
|
||||
|
||||
.audio-playing span:nth-child(2) {
|
||||
height: 60%;
|
||||
animation-delay: 0.1s;
|
||||
}
|
||||
|
||||
.audio-playing span:nth-child(3) {
|
||||
height: 80%;
|
||||
animation-delay: 0.2s;
|
||||
}
|
||||
|
||||
.audio-playing span:nth-child(4) {
|
||||
height: 60%;
|
||||
animation-delay: 0.3s;
|
||||
}
|
||||
|
||||
.audio-playing span:nth-child(5) {
|
||||
height: 40%;
|
||||
animation-delay: 0.4s;
|
||||
}
|
||||
|
||||
@keyframes audioBar {
|
||||
to {
|
||||
height: 100%;
|
||||
}
|
||||
}
|
||||
|
||||
/* Smooth transitions */
|
||||
.btn {
|
||||
transition: all 0.3s ease;
|
||||
}
|
||||
|
||||
.card {
|
||||
transition: transform 0.3s ease, box-shadow 0.3s ease;
|
||||
}
|
||||
|
||||
.card:hover {
|
||||
transform: translateY(-2px);
|
||||
box-shadow: 0 6px 12px rgba(0, 0, 0, 0.15);
|
||||
}
|
||||
|
||||
/* Success notification */
|
||||
.success-notification {
|
||||
position: fixed;
|
||||
top: 20px;
|
||||
left: 50%;
|
||||
transform: translateX(-50%);
|
||||
background-color: #28a745;
|
||||
color: white;
|
||||
padding: 12px 24px;
|
||||
border-radius: 8px;
|
||||
box-shadow: 0 4px 12px rgba(0, 0, 0, 0.15);
|
||||
display: flex;
|
||||
align-items: center;
|
||||
gap: 10px;
|
||||
z-index: 9999;
|
||||
opacity: 0;
|
||||
transition: opacity 0.3s ease, transform 0.3s ease;
|
||||
pointer-events: none;
|
||||
}
|
||||
|
||||
.success-notification.show {
|
||||
opacity: 1;
|
||||
transform: translateX(-50%) translateY(0);
|
||||
pointer-events: all;
|
||||
}
|
||||
|
||||
.success-notification i {
|
||||
font-size: 18px;
|
||||
}
|
||||
|
||||
/* Mobile optimizations */
|
||||
@media (max-width: 768px) {
|
||||
.loading-overlay {
|
||||
background: rgba(255, 255, 255, 0.95);
|
||||
}
|
||||
|
||||
.spinner-custom,
|
||||
.recording-wave {
|
||||
transform: scale(0.8);
|
||||
}
|
||||
|
||||
.success-notification {
|
||||
width: 90%;
|
||||
max-width: 300px;
|
||||
font-size: 14px;
|
||||
}
|
||||
|
||||
/* Make the entire layout more compact on mobile */
|
||||
body {
|
||||
font-size: 14px;
|
||||
}
|
||||
|
||||
/* Reduce spacing on mobile */
|
||||
.mb-3 {
|
||||
margin-bottom: 0.5rem !important;
|
||||
}
|
||||
|
||||
.mb-4 {
|
||||
margin-bottom: 0.75rem !important;
|
||||
}
|
||||
|
||||
/* Compact cards on mobile */
|
||||
.card {
|
||||
margin-bottom: 8px !important;
|
||||
}
|
||||
|
||||
/* Hide less important elements on small screens */
|
||||
.text-muted.small {
|
||||
font-size: 0.75rem;
|
||||
}
|
||||
|
||||
/* Adjust button sizes */
|
||||
.btn {
|
||||
font-size: 0.875rem;
|
||||
}
|
||||
|
||||
/* Make dropdowns more compact */
|
||||
.form-select {
|
||||
font-size: 0.875rem;
|
||||
padding: 0.25rem 0.5rem;
|
||||
}
|
||||
}
|
||||
|
||||
/* Streaming translation styles */
|
||||
.streaming-text {
|
||||
position: relative;
|
||||
min-height: 1.5em;
|
||||
}
|
||||
|
||||
.streaming-active::after {
|
||||
content: '▊';
|
||||
display: inline-block;
|
||||
animation: cursor-blink 1s infinite;
|
||||
color: #007bff;
|
||||
font-weight: bold;
|
||||
}
|
||||
|
||||
@keyframes cursor-blink {
|
||||
0%, 49% {
|
||||
opacity: 1;
|
||||
}
|
||||
50%, 100% {
|
||||
opacity: 0;
|
||||
}
|
||||
}
|
||||
|
||||
/* Smooth text appearance for streaming */
|
||||
.streaming-text {
|
||||
transition: all 0.1s ease-out;
|
||||
}
|
||||
|
||||
/* Multi-speaker styles */
|
||||
.speaker-button {
|
||||
position: relative;
|
||||
padding: 8px 16px;
|
||||
border-radius: 20px;
|
||||
border: 2px solid;
|
||||
background-color: white;
|
||||
font-weight: 500;
|
||||
transition: all 0.3s ease;
|
||||
min-width: 120px;
|
||||
}
|
||||
|
||||
.speaker-button.active {
|
||||
color: white !important;
|
||||
transform: scale(1.05);
|
||||
box-shadow: 0 2px 8px rgba(0,0,0,0.2);
|
||||
}
|
||||
|
||||
.speaker-avatar {
|
||||
display: inline-flex;
|
||||
align-items: center;
|
||||
justify-content: center;
|
||||
width: 30px;
|
||||
height: 30px;
|
||||
border-radius: 50%;
|
||||
background-color: rgba(255,255,255,0.3);
|
||||
color: inherit;
|
||||
font-weight: bold;
|
||||
font-size: 12px;
|
||||
margin-right: 8px;
|
||||
}
|
||||
|
||||
.speaker-button.active .speaker-avatar {
|
||||
background-color: rgba(255,255,255,0.3);
|
||||
}
|
||||
|
||||
.conversation-entry {
|
||||
margin-bottom: 16px;
|
||||
padding: 12px;
|
||||
border-radius: 12px;
|
||||
background-color: #f8f9fa;
|
||||
position: relative;
|
||||
animation: slideIn 0.3s ease-out;
|
||||
}
|
||||
|
||||
@keyframes slideIn {
|
||||
from {
|
||||
opacity: 0;
|
||||
transform: translateY(10px);
|
||||
}
|
||||
to {
|
||||
opacity: 1;
|
||||
transform: translateY(0);
|
||||
}
|
||||
}
|
||||
|
||||
.conversation-speaker {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
margin-bottom: 8px;
|
||||
font-weight: 600;
|
||||
}
|
||||
|
||||
.conversation-speaker-avatar {
|
||||
display: inline-flex;
|
||||
align-items: center;
|
||||
justify-content: center;
|
||||
width: 25px;
|
||||
height: 25px;
|
||||
border-radius: 50%;
|
||||
color: white;
|
||||
font-size: 11px;
|
||||
margin-right: 8px;
|
||||
}
|
||||
|
||||
.conversation-text {
|
||||
margin-left: 33px;
|
||||
line-height: 1.5;
|
||||
}
|
||||
|
||||
.conversation-time {
|
||||
font-size: 0.8rem;
|
||||
color: #6c757d;
|
||||
margin-left: auto;
|
||||
}
|
||||
|
||||
.conversation-translation {
|
||||
font-style: italic;
|
||||
opacity: 0.9;
|
||||
}
|
||||
|
||||
/* Speaker list responsive */
|
||||
@media (max-width: 768px) {
|
||||
.speaker-button {
|
||||
min-width: 100px;
|
||||
padding: 6px 12px;
|
||||
font-size: 0.9rem;
|
||||
}
|
||||
|
||||
.speaker-avatar {
|
||||
width: 25px;
|
||||
height: 25px;
|
||||
font-size: 10px;
|
||||
}
|
||||
}
|
600
static/js/app.js
600
static/js/app.js
@ -1,600 +0,0 @@
|
||||
// Main application JavaScript with PWA support
|
||||
document.addEventListener('DOMContentLoaded', function() {
|
||||
// Register service worker
|
||||
if ('serviceWorker' in navigator) {
|
||||
registerServiceWorker();
|
||||
}
|
||||
|
||||
// Initialize app
|
||||
initApp();
|
||||
|
||||
// Check for PWA installation prompts
|
||||
initInstallPrompt();
|
||||
});
|
||||
|
||||
// Service Worker Registration
|
||||
async function registerServiceWorker() {
|
||||
try {
|
||||
const registration = await navigator.serviceWorker.register('/service-worker.js');
|
||||
console.log('Service Worker registered with scope:', registration.scope);
|
||||
|
||||
// Setup periodic sync if available
|
||||
if ('periodicSync' in registration) {
|
||||
// Request permission for background sync
|
||||
const status = await navigator.permissions.query({
|
||||
name: 'periodic-background-sync',
|
||||
});
|
||||
|
||||
if (status.state === 'granted') {
|
||||
try {
|
||||
// Register for background sync to check for updates
|
||||
await registration.periodicSync.register('translation-updates', {
|
||||
minInterval: 24 * 60 * 60 * 1000, // once per day
|
||||
});
|
||||
console.log('Periodic background sync registered');
|
||||
} catch (error) {
|
||||
console.error('Periodic background sync could not be registered:', error);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Setup push notification if available
|
||||
if ('PushManager' in window) {
|
||||
setupPushNotifications(registration);
|
||||
}
|
||||
} catch (error) {
|
||||
console.error('Service Worker registration failed:', error);
|
||||
}
|
||||
}
|
||||
|
||||
// Initialize the main application
|
||||
function initApp() {
|
||||
// DOM elements
|
||||
const recordBtn = document.getElementById('recordBtn');
|
||||
const translateBtn = document.getElementById('translateBtn');
|
||||
const sourceText = document.getElementById('sourceText');
|
||||
const translatedText = document.getElementById('translatedText');
|
||||
const sourceLanguage = document.getElementById('sourceLanguage');
|
||||
const targetLanguage = document.getElementById('targetLanguage');
|
||||
const playSource = document.getElementById('playSource');
|
||||
const playTranslation = document.getElementById('playTranslation');
|
||||
const clearSource = document.getElementById('clearSource');
|
||||
const clearTranslation = document.getElementById('clearTranslation');
|
||||
const statusIndicator = document.getElementById('statusIndicator');
|
||||
const progressContainer = document.getElementById('progressContainer');
|
||||
const progressBar = document.getElementById('progressBar');
|
||||
const audioPlayer = document.getElementById('audioPlayer');
|
||||
const ttsServerAlert = document.getElementById('ttsServerAlert');
|
||||
const ttsServerMessage = document.getElementById('ttsServerMessage');
|
||||
const ttsServerUrl = document.getElementById('ttsServerUrl');
|
||||
const ttsApiKey = document.getElementById('ttsApiKey');
|
||||
const updateTtsServer = document.getElementById('updateTtsServer');
|
||||
|
||||
// Set initial values
|
||||
let isRecording = false;
|
||||
let mediaRecorder = null;
|
||||
let audioChunks = [];
|
||||
let currentSourceText = '';
|
||||
let currentTranslationText = '';
|
||||
let currentTtsServerUrl = '';
|
||||
|
||||
// Check TTS server status on page load
|
||||
checkTtsServer();
|
||||
|
||||
// Check for saved translations in IndexedDB
|
||||
loadSavedTranslations();
|
||||
|
||||
// Update TTS server URL and API key
|
||||
updateTtsServer.addEventListener('click', function() {
|
||||
const newUrl = ttsServerUrl.value.trim();
|
||||
const newApiKey = ttsApiKey.value.trim();
|
||||
|
||||
if (!newUrl && !newApiKey) {
|
||||
alert('Please provide at least one value to update');
|
||||
return;
|
||||
}
|
||||
|
||||
const updateData = {};
|
||||
if (newUrl) updateData.server_url = newUrl;
|
||||
if (newApiKey) updateData.api_key = newApiKey;
|
||||
|
||||
fetch('/update_tts_config', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify(updateData)
|
||||
})
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
if (data.success) {
|
||||
statusIndicator.textContent = 'TTS configuration updated';
|
||||
// Save URL to localStorage but not the API key for security
|
||||
if (newUrl) localStorage.setItem('ttsServerUrl', newUrl);
|
||||
// Check TTS server with new configuration
|
||||
checkTtsServer();
|
||||
} else {
|
||||
alert('Failed to update TTS configuration: ' + data.error);
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
console.error('Failed to update TTS config:', error);
|
||||
alert('Failed to update TTS configuration. See console for details.');
|
||||
});
|
||||
});
|
||||
|
||||
// Make sure target language is different from source
|
||||
if (targetLanguage.options[0].value === sourceLanguage.value) {
|
||||
targetLanguage.selectedIndex = 1;
|
||||
}
|
||||
|
||||
// Event listeners for language selection
|
||||
sourceLanguage.addEventListener('change', function() {
|
||||
if (targetLanguage.value === sourceLanguage.value) {
|
||||
for (let i = 0; i < targetLanguage.options.length; i++) {
|
||||
if (targetLanguage.options[i].value !== sourceLanguage.value) {
|
||||
targetLanguage.selectedIndex = i;
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
});
|
||||
|
||||
targetLanguage.addEventListener('change', function() {
|
||||
if (targetLanguage.value === sourceLanguage.value) {
|
||||
for (let i = 0; i < sourceLanguage.options.length; i++) {
|
||||
if (sourceLanguage.options[i].value !== targetLanguage.value) {
|
||||
sourceLanguage.selectedIndex = i;
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
});
|
||||
|
||||
// Record button click event
|
||||
recordBtn.addEventListener('click', function() {
|
||||
if (isRecording) {
|
||||
stopRecording();
|
||||
} else {
|
||||
startRecording();
|
||||
}
|
||||
});
|
||||
|
||||
// Function to start recording
|
||||
function startRecording() {
|
||||
navigator.mediaDevices.getUserMedia({ audio: true })
|
||||
.then(stream => {
|
||||
mediaRecorder = new MediaRecorder(stream);
|
||||
audioChunks = [];
|
||||
|
||||
mediaRecorder.addEventListener('dataavailable', event => {
|
||||
audioChunks.push(event.data);
|
||||
});
|
||||
|
||||
mediaRecorder.addEventListener('stop', () => {
|
||||
const audioBlob = new Blob(audioChunks, { type: 'audio/wav' });
|
||||
transcribeAudio(audioBlob);
|
||||
});
|
||||
|
||||
mediaRecorder.start();
|
||||
isRecording = true;
|
||||
recordBtn.classList.add('recording');
|
||||
recordBtn.classList.replace('btn-primary', 'btn-danger');
|
||||
recordBtn.innerHTML = '<i class="fas fa-stop"></i>';
|
||||
statusIndicator.textContent = 'Recording... Click to stop';
|
||||
})
|
||||
.catch(error => {
|
||||
console.error('Error accessing microphone:', error);
|
||||
alert('Error accessing microphone. Please make sure you have given permission for microphone access.');
|
||||
});
|
||||
}
|
||||
|
||||
// Function to stop recording
|
||||
function stopRecording() {
|
||||
mediaRecorder.stop();
|
||||
isRecording = false;
|
||||
recordBtn.classList.remove('recording');
|
||||
recordBtn.classList.replace('btn-danger', 'btn-primary');
|
||||
recordBtn.innerHTML = '<i class="fas fa-microphone"></i>';
|
||||
statusIndicator.textContent = 'Processing audio...';
|
||||
|
||||
// Stop all audio tracks
|
||||
mediaRecorder.stream.getTracks().forEach(track => track.stop());
|
||||
}
|
||||
|
||||
// Function to transcribe audio
|
||||
function transcribeAudio(audioBlob) {
|
||||
const formData = new FormData();
|
||||
formData.append('audio', audioBlob);
|
||||
formData.append('source_lang', sourceLanguage.value);
|
||||
|
||||
showProgress();
|
||||
|
||||
fetch('/transcribe', {
|
||||
method: 'POST',
|
||||
body: formData
|
||||
})
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
hideProgress();
|
||||
|
||||
if (data.success) {
|
||||
currentSourceText = data.text;
|
||||
sourceText.innerHTML = `<p>${data.text}</p>`;
|
||||
playSource.disabled = false;
|
||||
translateBtn.disabled = false;
|
||||
statusIndicator.textContent = 'Transcription complete';
|
||||
|
||||
// Cache the transcription in IndexedDB
|
||||
saveToIndexedDB('transcriptions', {
|
||||
text: data.text,
|
||||
language: sourceLanguage.value,
|
||||
timestamp: new Date().toISOString()
|
||||
});
|
||||
} else {
|
||||
sourceText.innerHTML = `<p class="text-danger">Error: ${data.error}</p>`;
|
||||
statusIndicator.textContent = 'Transcription failed';
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
hideProgress();
|
||||
console.error('Transcription error:', error);
|
||||
sourceText.innerHTML = `<p class="text-danger">Failed to transcribe audio. Please try again.</p>`;
|
||||
statusIndicator.textContent = 'Transcription failed';
|
||||
});
|
||||
}
|
||||
|
||||
// Translate button click event
|
||||
translateBtn.addEventListener('click', function() {
|
||||
if (!currentSourceText) {
|
||||
return;
|
||||
}
|
||||
|
||||
statusIndicator.textContent = 'Translating...';
|
||||
showProgress();
|
||||
|
||||
fetch('/translate', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
text: currentSourceText,
|
||||
source_lang: sourceLanguage.value,
|
||||
target_lang: targetLanguage.value
|
||||
})
|
||||
})
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
hideProgress();
|
||||
|
||||
if (data.success) {
|
||||
currentTranslationText = data.translation;
|
||||
translatedText.innerHTML = `<p>${data.translation}</p>`;
|
||||
playTranslation.disabled = false;
|
||||
statusIndicator.textContent = 'Translation complete';
|
||||
|
||||
// Cache the translation in IndexedDB
|
||||
saveToIndexedDB('translations', {
|
||||
sourceText: currentSourceText,
|
||||
sourceLanguage: sourceLanguage.value,
|
||||
targetText: data.translation,
|
||||
targetLanguage: targetLanguage.value,
|
||||
timestamp: new Date().toISOString()
|
||||
});
|
||||
} else {
|
||||
translatedText.innerHTML = `<p class="text-danger">Error: ${data.error}</p>`;
|
||||
statusIndicator.textContent = 'Translation failed';
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
hideProgress();
|
||||
console.error('Translation error:', error);
|
||||
translatedText.innerHTML = `<p class="text-danger">Failed to translate. Please try again.</p>`;
|
||||
statusIndicator.textContent = 'Translation failed';
|
||||
});
|
||||
});
|
||||
|
||||
// Play source text
|
||||
playSource.addEventListener('click', function() {
|
||||
if (!currentSourceText) return;
|
||||
|
||||
playAudio(currentSourceText, sourceLanguage.value);
|
||||
statusIndicator.textContent = 'Playing source audio...';
|
||||
});
|
||||
|
||||
// Play translation
|
||||
playTranslation.addEventListener('click', function() {
|
||||
if (!currentTranslationText) return;
|
||||
|
||||
playAudio(currentTranslationText, targetLanguage.value);
|
||||
statusIndicator.textContent = 'Playing translation audio...';
|
||||
});
|
||||
|
||||
// Function to play audio via TTS
|
||||
function playAudio(text, language) {
|
||||
showProgress();
|
||||
|
||||
fetch('/speak', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
text: text,
|
||||
language: language
|
||||
})
|
||||
})
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
hideProgress();
|
||||
|
||||
if (data.success) {
|
||||
audioPlayer.src = data.audio_url;
|
||||
audioPlayer.onended = function() {
|
||||
statusIndicator.textContent = 'Ready';
|
||||
};
|
||||
audioPlayer.play();
|
||||
} else {
|
||||
statusIndicator.textContent = 'TTS failed';
|
||||
|
||||
// Show TTS server alert with error message
|
||||
ttsServerAlert.classList.remove('d-none');
|
||||
ttsServerAlert.classList.remove('alert-success');
|
||||
ttsServerAlert.classList.add('alert-warning');
|
||||
ttsServerMessage.textContent = data.error;
|
||||
|
||||
alert('Failed to play audio: ' + data.error);
|
||||
|
||||
// Check TTS server status again
|
||||
checkTtsServer();
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
hideProgress();
|
||||
console.error('TTS error:', error);
|
||||
statusIndicator.textContent = 'TTS failed';
|
||||
|
||||
// Show TTS server alert
|
||||
ttsServerAlert.classList.remove('d-none');
|
||||
ttsServerAlert.classList.remove('alert-success');
|
||||
ttsServerAlert.classList.add('alert-warning');
|
||||
ttsServerMessage.textContent = 'Failed to connect to TTS server';
|
||||
});
|
||||
}
|
||||
|
||||
// Clear buttons
|
||||
clearSource.addEventListener('click', function() {
|
||||
sourceText.innerHTML = '<p class="text-muted">Your transcribed text will appear here...</p>';
|
||||
currentSourceText = '';
|
||||
playSource.disabled = true;
|
||||
translateBtn.disabled = true;
|
||||
});
|
||||
|
||||
clearTranslation.addEventListener('click', function() {
|
||||
translatedText.innerHTML = '<p class="text-muted">Translation will appear here...</p>';
|
||||
currentTranslationText = '';
|
||||
playTranslation.disabled = true;
|
||||
});
|
||||
|
||||
// Function to check TTS server status
|
||||
function checkTtsServer() {
|
||||
fetch('/check_tts_server')
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
currentTtsServerUrl = data.url;
|
||||
ttsServerUrl.value = currentTtsServerUrl;
|
||||
|
||||
// Load saved API key if available
|
||||
const savedApiKey = localStorage.getItem('ttsApiKeySet');
|
||||
if (savedApiKey === 'true') {
|
||||
ttsApiKey.placeholder = '••••••• (API key saved)';
|
||||
}
|
||||
|
||||
if (data.status === 'error' || data.status === 'auth_error') {
|
||||
ttsServerAlert.classList.remove('d-none');
|
||||
ttsServerAlert.classList.remove('alert-success');
|
||||
ttsServerAlert.classList.add('alert-warning');
|
||||
ttsServerMessage.textContent = data.message;
|
||||
|
||||
if (data.status === 'auth_error') {
|
||||
ttsServerMessage.textContent = 'Authentication error with TTS server. Please check your API key.';
|
||||
}
|
||||
} else {
|
||||
ttsServerAlert.classList.remove('d-none');
|
||||
ttsServerAlert.classList.remove('alert-warning');
|
||||
ttsServerAlert.classList.add('alert-success');
|
||||
ttsServerMessage.textContent = 'TTS server is online and ready.';
|
||||
setTimeout(() => {
|
||||
ttsServerAlert.classList.add('d-none');
|
||||
}, 3000);
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
console.error('Failed to check TTS server:', error);
|
||||
ttsServerAlert.classList.remove('d-none');
|
||||
ttsServerAlert.classList.remove('alert-success');
|
||||
ttsServerAlert.classList.add('alert-warning');
|
||||
ttsServerMessage.textContent = 'Failed to check TTS server status.';
|
||||
});
|
||||
}
|
||||
|
||||
// Progress indicator functions
|
||||
function showProgress() {
|
||||
progressContainer.classList.remove('d-none');
|
||||
let progress = 0;
|
||||
const interval = setInterval(() => {
|
||||
progress += 5;
|
||||
if (progress > 90) {
|
||||
clearInterval(interval);
|
||||
}
|
||||
progressBar.style.width = `${progress}%`;
|
||||
}, 100);
|
||||
progressBar.dataset.interval = interval;
|
||||
}
|
||||
|
||||
function hideProgress() {
|
||||
const interval = progressBar.dataset.interval;
|
||||
if (interval) {
|
||||
clearInterval(Number(interval));
|
||||
}
|
||||
progressBar.style.width = '100%';
|
||||
setTimeout(() => {
|
||||
progressContainer.classList.add('d-none');
|
||||
progressBar.style.width = '0%';
|
||||
}, 500);
|
||||
}
|
||||
}
|
||||
|
||||
// IndexedDB functions for offline data storage
|
||||
function openIndexedDB() {
|
||||
return new Promise((resolve, reject) => {
|
||||
const request = indexedDB.open('VoiceTranslatorDB', 1);
|
||||
|
||||
request.onupgradeneeded = (event) => {
|
||||
const db = event.target.result;
|
||||
|
||||
// Create stores for transcriptions and translations
|
||||
if (!db.objectStoreNames.contains('transcriptions')) {
|
||||
db.createObjectStore('transcriptions', { keyPath: 'timestamp' });
|
||||
}
|
||||
|
||||
if (!db.objectStoreNames.contains('translations')) {
|
||||
db.createObjectStore('translations', { keyPath: 'timestamp' });
|
||||
}
|
||||
};
|
||||
|
||||
request.onsuccess = (event) => {
|
||||
resolve(event.target.result);
|
||||
};
|
||||
|
||||
request.onerror = (event) => {
|
||||
reject('IndexedDB error: ' + event.target.errorCode);
|
||||
};
|
||||
});
|
||||
}
|
||||
|
||||
function saveToIndexedDB(storeName, data) {
|
||||
openIndexedDB().then(db => {
|
||||
const transaction = db.transaction([storeName], 'readwrite');
|
||||
const store = transaction.objectStore(storeName);
|
||||
store.add(data);
|
||||
}).catch(error => {
|
||||
console.error('Error saving to IndexedDB:', error);
|
||||
});
|
||||
}
|
||||
|
||||
function loadSavedTranslations() {
|
||||
openIndexedDB().then(db => {
|
||||
const transaction = db.transaction(['translations'], 'readonly');
|
||||
const store = transaction.objectStore('translations');
|
||||
const request = store.getAll();
|
||||
|
||||
request.onsuccess = (event) => {
|
||||
const translations = event.target.result;
|
||||
if (translations && translations.length > 0) {
|
||||
// Could add a history section or recently used translations
|
||||
console.log('Loaded saved translations:', translations.length);
|
||||
}
|
||||
};
|
||||
}).catch(error => {
|
||||
console.error('Error loading from IndexedDB:', error);
|
||||
});
|
||||
}
|
||||
|
||||
// PWA installation prompt
|
||||
function initInstallPrompt() {
|
||||
let deferredPrompt;
|
||||
const installButton = document.createElement('button');
|
||||
installButton.style.display = 'none';
|
||||
installButton.classList.add('btn', 'btn-success', 'fixed-bottom', 'm-3');
|
||||
installButton.innerHTML = 'Install Voice Translator <i class="fas fa-download ml-2"></i>';
|
||||
document.body.appendChild(installButton);
|
||||
|
||||
window.addEventListener('beforeinstallprompt', (e) => {
|
||||
// Prevent Chrome 67 and earlier from automatically showing the prompt
|
||||
e.preventDefault();
|
||||
// Stash the event so it can be triggered later
|
||||
deferredPrompt = e;
|
||||
// Update UI to notify the user they can add to home screen
|
||||
installButton.style.display = 'block';
|
||||
|
||||
installButton.addEventListener('click', (e) => {
|
||||
// Hide our user interface that shows our install button
|
||||
installButton.style.display = 'none';
|
||||
// Show the prompt
|
||||
deferredPrompt.prompt();
|
||||
// Wait for the user to respond to the prompt
|
||||
deferredPrompt.userChoice.then((choiceResult) => {
|
||||
if (choiceResult.outcome === 'accepted') {
|
||||
console.log('User accepted the install prompt');
|
||||
} else {
|
||||
console.log('User dismissed the install prompt');
|
||||
}
|
||||
deferredPrompt = null;
|
||||
});
|
||||
});
|
||||
});
|
||||
}
|
||||
|
||||
// Push notification setup
|
||||
function setupPushNotifications(swRegistration) {
|
||||
// First check if we already have permission
|
||||
if (Notification.permission === 'granted') {
|
||||
console.log('Notification permission already granted');
|
||||
subscribeToPushManager(swRegistration);
|
||||
} else if (Notification.permission !== 'denied') {
|
||||
// Otherwise, ask for permission
|
||||
Notification.requestPermission().then(function(permission) {
|
||||
if (permission === 'granted') {
|
||||
console.log('Notification permission granted');
|
||||
subscribeToPushManager(swRegistration);
|
||||
}
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
async function subscribeToPushManager(swRegistration) {
|
||||
try {
|
||||
// Get the server's public key
|
||||
const response = await fetch('/api/push-public-key');
|
||||
const data = await response.json();
|
||||
|
||||
// Convert the base64 string to Uint8Array
|
||||
function urlBase64ToUint8Array(base64String) {
|
||||
const padding = '='.repeat((4 - base64String.length % 4) % 4);
|
||||
const base64 = (base64String + padding)
|
||||
.replace(/-/g, '+')
|
||||
.replace(/_/g, '/');
|
||||
|
||||
const rawData = window.atob(base64);
|
||||
const outputArray = new Uint8Array(rawData.length);
|
||||
|
||||
for (let i = 0; i < rawData.length; ++i) {
|
||||
outputArray[i] = rawData.charCodeAt(i);
|
||||
}
|
||||
return outputArray;
|
||||
}
|
||||
|
||||
const convertedVapidKey = urlBase64ToUint8Array(data.publicKey);
|
||||
|
||||
// Subscribe to push notifications
|
||||
const subscription = await swRegistration.pushManager.subscribe({
|
||||
userVisibleOnly: true,
|
||||
applicationServerKey: convertedVapidKey
|
||||
});
|
||||
|
||||
// Send the subscription details to the server
|
||||
await fetch('/api/push-subscribe', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json',
|
||||
},
|
||||
body: JSON.stringify(subscription)
|
||||
});
|
||||
|
||||
console.log('User is subscribed to push notifications');
|
||||
} catch (error) {
|
||||
console.error('Failed to subscribe to push notifications:', error);
|
||||
}
|
||||
}
|
155
static/js/src/apiClient.ts
Normal file
155
static/js/src/apiClient.ts
Normal file
@ -0,0 +1,155 @@
|
||||
// API Client with CORS support
|
||||
export interface ApiClientConfig {
|
||||
baseUrl?: string;
|
||||
credentials?: RequestCredentials;
|
||||
headers?: HeadersInit;
|
||||
}
|
||||
|
||||
export class ApiClient {
|
||||
private static instance: ApiClient;
|
||||
private config: ApiClientConfig;
|
||||
|
||||
private constructor() {
|
||||
// Default configuration
|
||||
this.config = {
|
||||
baseUrl: '', // Use same origin by default
|
||||
credentials: 'same-origin', // Change to 'include' for cross-origin requests
|
||||
headers: {
|
||||
'X-Requested-With': 'XMLHttpRequest' // Identify as AJAX request
|
||||
}
|
||||
};
|
||||
|
||||
// Check if we're in a cross-origin context
|
||||
this.detectCrossOrigin();
|
||||
}
|
||||
|
||||
static getInstance(): ApiClient {
|
||||
if (!ApiClient.instance) {
|
||||
ApiClient.instance = new ApiClient();
|
||||
}
|
||||
return ApiClient.instance;
|
||||
}
|
||||
|
||||
// Detect if we're making cross-origin requests
|
||||
private detectCrossOrigin(): void {
|
||||
// Check if the app is loaded from a different origin
|
||||
const currentScript = document.currentScript as HTMLScriptElement | null;
|
||||
const scriptSrc = currentScript?.src || '';
|
||||
if (scriptSrc && !scriptSrc.startsWith(window.location.origin)) {
|
||||
// We're likely in a cross-origin context
|
||||
this.config.credentials = 'include';
|
||||
console.log('Cross-origin context detected, enabling credentials');
|
||||
}
|
||||
|
||||
// Also check for explicit configuration in meta tags
|
||||
const corsOrigin = document.querySelector('meta[name="cors-origin"]');
|
||||
if (corsOrigin) {
|
||||
const origin = corsOrigin.getAttribute('content');
|
||||
if (origin && origin !== window.location.origin) {
|
||||
this.config.baseUrl = origin;
|
||||
this.config.credentials = 'include';
|
||||
console.log(`Using CORS origin: ${origin}`);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Configure the API client
|
||||
configure(config: Partial<ApiClientConfig>): void {
|
||||
this.config = { ...this.config, ...config };
|
||||
}
|
||||
|
||||
// Make a fetch request with CORS support
|
||||
async fetch(url: string, options: RequestInit = {}): Promise<Response> {
|
||||
// Construct full URL
|
||||
const fullUrl = this.config.baseUrl ? `${this.config.baseUrl}${url}` : url;
|
||||
|
||||
// Merge headers
|
||||
const headers = new Headers(options.headers);
|
||||
if (this.config.headers) {
|
||||
const configHeaders = new Headers(this.config.headers);
|
||||
configHeaders.forEach((value, key) => {
|
||||
if (!headers.has(key)) {
|
||||
headers.set(key, value);
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
// Merge options with defaults
|
||||
const fetchOptions: RequestInit = {
|
||||
...options,
|
||||
headers,
|
||||
credentials: options.credentials || this.config.credentials
|
||||
};
|
||||
|
||||
// Add CORS mode if cross-origin
|
||||
if (this.config.baseUrl && this.config.baseUrl !== window.location.origin) {
|
||||
fetchOptions.mode = 'cors';
|
||||
}
|
||||
|
||||
try {
|
||||
const response = await fetch(fullUrl, fetchOptions);
|
||||
|
||||
// Check for CORS errors
|
||||
if (!response.ok && response.type === 'opaque') {
|
||||
throw new Error('CORS request failed - check server CORS configuration');
|
||||
}
|
||||
|
||||
return response;
|
||||
} catch (error) {
|
||||
// Enhanced error handling for CORS issues
|
||||
if (error instanceof TypeError && error.message.includes('Failed to fetch')) {
|
||||
console.error('CORS Error: Failed to fetch. Check that:', {
|
||||
requestedUrl: fullUrl,
|
||||
origin: window.location.origin,
|
||||
credentials: fetchOptions.credentials,
|
||||
mode: fetchOptions.mode
|
||||
});
|
||||
throw new Error('CORS request failed. The server may not allow requests from this origin.');
|
||||
}
|
||||
throw error;
|
||||
}
|
||||
}
|
||||
|
||||
// Convenience methods
|
||||
async get(url: string, options?: RequestInit): Promise<Response> {
|
||||
return this.fetch(url, { ...options, method: 'GET' });
|
||||
}
|
||||
|
||||
async post(url: string, body?: any, options?: RequestInit): Promise<Response> {
|
||||
const init: RequestInit = { ...options, method: 'POST' };
|
||||
|
||||
if (body) {
|
||||
if (body instanceof FormData) {
|
||||
init.body = body;
|
||||
} else {
|
||||
init.headers = {
|
||||
...init.headers,
|
||||
'Content-Type': 'application/json'
|
||||
};
|
||||
init.body = JSON.stringify(body);
|
||||
}
|
||||
}
|
||||
|
||||
return this.fetch(url, init);
|
||||
}
|
||||
|
||||
// JSON convenience methods
|
||||
async getJSON<T>(url: string, options?: RequestInit): Promise<T> {
|
||||
const response = await this.get(url, options);
|
||||
if (!response.ok) {
|
||||
throw new Error(`HTTP error! status: ${response.status}`);
|
||||
}
|
||||
return response.json();
|
||||
}
|
||||
|
||||
async postJSON<T>(url: string, body?: any, options?: RequestInit): Promise<T> {
|
||||
const response = await this.post(url, body, options);
|
||||
if (!response.ok) {
|
||||
throw new Error(`HTTP error! status: ${response.status}`);
|
||||
}
|
||||
return response.json();
|
||||
}
|
||||
}
|
||||
|
||||
// Export a singleton instance
|
||||
export const apiClient = ApiClient.getInstance();
|
1689
static/js/src/app.ts
Normal file
1689
static/js/src/app.ts
Normal file
File diff suppressed because it is too large
Load Diff
321
static/js/src/connectionManager.ts
Normal file
321
static/js/src/connectionManager.ts
Normal file
@ -0,0 +1,321 @@
|
||||
// Connection management with retry logic
|
||||
export interface ConnectionConfig {
|
||||
maxRetries: number;
|
||||
initialDelay: number;
|
||||
maxDelay: number;
|
||||
backoffMultiplier: number;
|
||||
timeout: number;
|
||||
onlineCheckInterval: number;
|
||||
}
|
||||
|
||||
export interface RetryOptions {
|
||||
retries?: number;
|
||||
delay?: number;
|
||||
onRetry?: (attempt: number, error: Error) => void;
|
||||
}
|
||||
|
||||
export type ConnectionStatus = 'online' | 'offline' | 'connecting' | 'error';
|
||||
|
||||
export interface ConnectionState {
|
||||
status: ConnectionStatus;
|
||||
lastError?: Error;
|
||||
retryCount: number;
|
||||
lastOnlineTime?: Date;
|
||||
}
|
||||
|
||||
export class ConnectionManager {
|
||||
private static instance: ConnectionManager;
|
||||
private config: ConnectionConfig;
|
||||
private state: ConnectionState;
|
||||
private listeners: Map<string, (state: ConnectionState) => void> = new Map();
|
||||
private onlineCheckTimer?: number;
|
||||
private reconnectTimer?: number;
|
||||
|
||||
private constructor() {
|
||||
this.config = {
|
||||
maxRetries: 3,
|
||||
initialDelay: 1000, // 1 second
|
||||
maxDelay: 30000, // 30 seconds
|
||||
backoffMultiplier: 2,
|
||||
timeout: 10000, // 10 seconds
|
||||
onlineCheckInterval: 5000 // 5 seconds
|
||||
};
|
||||
|
||||
this.state = {
|
||||
status: navigator.onLine ? 'online' : 'offline',
|
||||
retryCount: 0
|
||||
};
|
||||
|
||||
this.setupEventListeners();
|
||||
this.startOnlineCheck();
|
||||
}
|
||||
|
||||
static getInstance(): ConnectionManager {
|
||||
if (!ConnectionManager.instance) {
|
||||
ConnectionManager.instance = new ConnectionManager();
|
||||
}
|
||||
return ConnectionManager.instance;
|
||||
}
|
||||
|
||||
// Configure connection settings
|
||||
configure(config: Partial<ConnectionConfig>): void {
|
||||
this.config = { ...this.config, ...config };
|
||||
}
|
||||
|
||||
// Setup browser online/offline event listeners
|
||||
private setupEventListeners(): void {
|
||||
window.addEventListener('online', () => {
|
||||
console.log('Browser online event detected');
|
||||
this.updateState({ status: 'online', retryCount: 0 });
|
||||
this.checkServerConnection();
|
||||
});
|
||||
|
||||
window.addEventListener('offline', () => {
|
||||
console.log('Browser offline event detected');
|
||||
this.updateState({ status: 'offline' });
|
||||
});
|
||||
|
||||
// Listen for visibility changes to check connection when tab becomes active
|
||||
document.addEventListener('visibilitychange', () => {
|
||||
if (!document.hidden && this.state.status === 'offline') {
|
||||
this.checkServerConnection();
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
// Start periodic online checking
|
||||
private startOnlineCheck(): void {
|
||||
this.onlineCheckTimer = window.setInterval(() => {
|
||||
if (this.state.status === 'offline' || this.state.status === 'error') {
|
||||
this.checkServerConnection();
|
||||
}
|
||||
}, this.config.onlineCheckInterval);
|
||||
}
|
||||
|
||||
// Check actual server connection
|
||||
async checkServerConnection(): Promise<boolean> {
|
||||
if (!navigator.onLine) {
|
||||
this.updateState({ status: 'offline' });
|
||||
return false;
|
||||
}
|
||||
|
||||
this.updateState({ status: 'connecting' });
|
||||
|
||||
try {
|
||||
const controller = new AbortController();
|
||||
const timeoutId = setTimeout(() => controller.abort(), 5000);
|
||||
|
||||
const response = await fetch('/health', {
|
||||
method: 'GET',
|
||||
signal: controller.signal,
|
||||
cache: 'no-cache'
|
||||
});
|
||||
|
||||
clearTimeout(timeoutId);
|
||||
|
||||
if (response.ok) {
|
||||
this.updateState({
|
||||
status: 'online',
|
||||
retryCount: 0,
|
||||
lastOnlineTime: new Date()
|
||||
});
|
||||
return true;
|
||||
} else {
|
||||
throw new Error(`Server returned status ${response.status}`);
|
||||
}
|
||||
} catch (error) {
|
||||
this.updateState({
|
||||
status: 'error',
|
||||
lastError: error as Error
|
||||
});
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
// Retry a failed request with exponential backoff
|
||||
async retryRequest<T>(
|
||||
request: () => Promise<T>,
|
||||
options: RetryOptions = {}
|
||||
): Promise<T> {
|
||||
const {
|
||||
retries = this.config.maxRetries,
|
||||
delay = this.config.initialDelay,
|
||||
onRetry
|
||||
} = options;
|
||||
|
||||
let lastError: Error;
|
||||
|
||||
for (let attempt = 0; attempt <= retries; attempt++) {
|
||||
try {
|
||||
// Check if we're online before attempting
|
||||
if (!navigator.onLine) {
|
||||
throw new Error('No internet connection');
|
||||
}
|
||||
|
||||
// Add timeout to request
|
||||
const result = await this.withTimeout(request(), this.config.timeout);
|
||||
|
||||
// Success - reset retry count
|
||||
if (this.state.retryCount > 0) {
|
||||
this.updateState({ retryCount: 0 });
|
||||
}
|
||||
|
||||
return result;
|
||||
} catch (error) {
|
||||
lastError = error as Error;
|
||||
|
||||
// Don't retry if offline
|
||||
if (!navigator.onLine) {
|
||||
this.updateState({ status: 'offline' });
|
||||
throw new Error('Request failed: No internet connection');
|
||||
}
|
||||
|
||||
// Don't retry on client errors (4xx)
|
||||
if (this.isClientError(error)) {
|
||||
throw error;
|
||||
}
|
||||
|
||||
// Call retry callback if provided
|
||||
if (onRetry && attempt < retries) {
|
||||
onRetry(attempt + 1, lastError);
|
||||
}
|
||||
|
||||
// If we have retries left, wait and try again
|
||||
if (attempt < retries) {
|
||||
const backoffDelay = Math.min(
|
||||
delay * Math.pow(this.config.backoffMultiplier, attempt),
|
||||
this.config.maxDelay
|
||||
);
|
||||
|
||||
console.log(`Retry attempt ${attempt + 1}/${retries} after ${backoffDelay}ms`);
|
||||
|
||||
// Update retry count in state
|
||||
this.updateState({ retryCount: attempt + 1 });
|
||||
|
||||
await this.delay(backoffDelay);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// All retries exhausted
|
||||
this.updateState({
|
||||
status: 'error',
|
||||
lastError: lastError!
|
||||
});
|
||||
|
||||
throw new Error(`Request failed after ${retries} retries: ${lastError!.message}`);
|
||||
}
|
||||
|
||||
// Add timeout to a promise
|
||||
private withTimeout<T>(promise: Promise<T>, timeout: number): Promise<T> {
|
||||
return Promise.race([
|
||||
promise,
|
||||
new Promise<T>((_, reject) => {
|
||||
setTimeout(() => reject(new Error('Request timeout')), timeout);
|
||||
})
|
||||
]);
|
||||
}
|
||||
|
||||
// Check if error is a client error (4xx)
|
||||
private isClientError(error: any): boolean {
|
||||
if (error.response && error.response.status >= 400 && error.response.status < 500) {
|
||||
return true;
|
||||
}
|
||||
|
||||
// Check for specific error messages that shouldn't be retried
|
||||
const message = error.message?.toLowerCase() || '';
|
||||
const noRetryErrors = ['unauthorized', 'forbidden', 'bad request', 'not found'];
|
||||
|
||||
return noRetryErrors.some(e => message.includes(e));
|
||||
}
|
||||
|
||||
// Delay helper
|
||||
private delay(ms: number): Promise<void> {
|
||||
return new Promise(resolve => setTimeout(resolve, ms));
|
||||
}
|
||||
|
||||
// Update connection state
|
||||
private updateState(updates: Partial<ConnectionState>): void {
|
||||
this.state = { ...this.state, ...updates };
|
||||
this.notifyListeners();
|
||||
}
|
||||
|
||||
// Subscribe to connection state changes
|
||||
subscribe(id: string, callback: (state: ConnectionState) => void): void {
|
||||
this.listeners.set(id, callback);
|
||||
// Immediately call with current state
|
||||
callback(this.state);
|
||||
}
|
||||
|
||||
// Unsubscribe from connection state changes
|
||||
unsubscribe(id: string): void {
|
||||
this.listeners.delete(id);
|
||||
}
|
||||
|
||||
// Notify all listeners of state change
|
||||
private notifyListeners(): void {
|
||||
this.listeners.forEach(callback => callback(this.state));
|
||||
}
|
||||
|
||||
// Get current connection state
|
||||
getState(): ConnectionState {
|
||||
return { ...this.state };
|
||||
}
|
||||
|
||||
// Check if currently online
|
||||
isOnline(): boolean {
|
||||
return this.state.status === 'online';
|
||||
}
|
||||
|
||||
// Manual reconnect attempt
|
||||
async reconnect(): Promise<boolean> {
|
||||
console.log('Manual reconnect requested');
|
||||
return this.checkServerConnection();
|
||||
}
|
||||
|
||||
// Cleanup
|
||||
destroy(): void {
|
||||
if (this.onlineCheckTimer) {
|
||||
clearInterval(this.onlineCheckTimer);
|
||||
}
|
||||
if (this.reconnectTimer) {
|
||||
clearTimeout(this.reconnectTimer);
|
||||
}
|
||||
this.listeners.clear();
|
||||
}
|
||||
}
|
||||
|
||||
// Helper function for retrying fetch requests
|
||||
export async function fetchWithRetry(
|
||||
url: string,
|
||||
options: RequestInit = {},
|
||||
retryOptions: RetryOptions = {}
|
||||
): Promise<Response> {
|
||||
const connectionManager = ConnectionManager.getInstance();
|
||||
|
||||
return connectionManager.retryRequest(async () => {
|
||||
const response = await fetch(url, options);
|
||||
|
||||
if (!response.ok && response.status >= 500) {
|
||||
// Server error - throw to trigger retry
|
||||
throw new Error(`Server error: ${response.status}`);
|
||||
}
|
||||
|
||||
return response;
|
||||
}, retryOptions);
|
||||
}
|
||||
|
||||
// Helper function for retrying JSON requests
|
||||
export async function fetchJSONWithRetry<T>(
|
||||
url: string,
|
||||
options: RequestInit = {},
|
||||
retryOptions: RetryOptions = {}
|
||||
): Promise<T> {
|
||||
const response = await fetchWithRetry(url, options, retryOptions);
|
||||
|
||||
if (!response.ok) {
|
||||
throw new Error(`HTTP error! status: ${response.status}`);
|
||||
}
|
||||
|
||||
return response.json();
|
||||
}
|
325
static/js/src/connectionUI.ts
Normal file
325
static/js/src/connectionUI.ts
Normal file
@ -0,0 +1,325 @@
|
||||
// Connection status UI component
|
||||
import { ConnectionManager, ConnectionState } from './connectionManager';
|
||||
import { RequestQueueManager } from './requestQueue';
|
||||
|
||||
export class ConnectionUI {
|
||||
private static instance: ConnectionUI;
|
||||
private connectionManager: ConnectionManager;
|
||||
private queueManager: RequestQueueManager;
|
||||
private statusElement: HTMLElement | null = null;
|
||||
private retryButton: HTMLButtonElement | null = null;
|
||||
private offlineMessage: HTMLElement | null = null;
|
||||
|
||||
private constructor() {
|
||||
this.connectionManager = ConnectionManager.getInstance();
|
||||
this.queueManager = RequestQueueManager.getInstance();
|
||||
this.createUI();
|
||||
this.subscribeToConnectionChanges();
|
||||
}
|
||||
|
||||
static getInstance(): ConnectionUI {
|
||||
if (!ConnectionUI.instance) {
|
||||
ConnectionUI.instance = new ConnectionUI();
|
||||
}
|
||||
return ConnectionUI.instance;
|
||||
}
|
||||
|
||||
private createUI(): void {
|
||||
// Create connection status indicator
|
||||
this.statusElement = document.createElement('div');
|
||||
this.statusElement.id = 'connectionStatus';
|
||||
this.statusElement.className = 'connection-status';
|
||||
this.statusElement.innerHTML = `
|
||||
<span class="connection-icon"></span>
|
||||
<span class="connection-text">Checking connection...</span>
|
||||
`;
|
||||
|
||||
// Create offline message banner
|
||||
this.offlineMessage = document.createElement('div');
|
||||
this.offlineMessage.id = 'offlineMessage';
|
||||
this.offlineMessage.className = 'offline-message';
|
||||
this.offlineMessage.innerHTML = `
|
||||
<div class="offline-content">
|
||||
<i class="fas fa-wifi-slash"></i>
|
||||
<span class="offline-text">You're offline. Some features may be limited.</span>
|
||||
<button class="btn btn-sm btn-outline-light retry-connection">
|
||||
<i class="fas fa-sync"></i> Retry
|
||||
</button>
|
||||
<div class="queued-info" style="display: none;">
|
||||
<small class="queued-count"></small>
|
||||
</div>
|
||||
</div>
|
||||
`;
|
||||
this.offlineMessage.style.display = 'none';
|
||||
|
||||
// Add to page
|
||||
document.body.appendChild(this.statusElement);
|
||||
document.body.appendChild(this.offlineMessage);
|
||||
|
||||
// Get retry button reference
|
||||
this.retryButton = this.offlineMessage.querySelector('.retry-connection') as HTMLButtonElement;
|
||||
this.retryButton?.addEventListener('click', () => this.handleRetry());
|
||||
|
||||
// Add CSS if not already present
|
||||
if (!document.getElementById('connection-ui-styles')) {
|
||||
const style = document.createElement('style');
|
||||
style.id = 'connection-ui-styles';
|
||||
style.textContent = `
|
||||
.connection-status {
|
||||
position: fixed;
|
||||
bottom: 20px;
|
||||
right: 20px;
|
||||
background: rgba(0, 0, 0, 0.8);
|
||||
color: white;
|
||||
padding: 8px 16px;
|
||||
border-radius: 20px;
|
||||
display: flex;
|
||||
align-items: center;
|
||||
gap: 8px;
|
||||
font-size: 14px;
|
||||
z-index: 1000;
|
||||
transition: all 0.3s ease;
|
||||
opacity: 0;
|
||||
transform: translateY(10px);
|
||||
}
|
||||
|
||||
.connection-status.visible {
|
||||
opacity: 1;
|
||||
transform: translateY(0);
|
||||
}
|
||||
|
||||
.connection-status.online {
|
||||
background: rgba(40, 167, 69, 0.9);
|
||||
}
|
||||
|
||||
.connection-status.offline {
|
||||
background: rgba(220, 53, 69, 0.9);
|
||||
}
|
||||
|
||||
.connection-status.connecting {
|
||||
background: rgba(255, 193, 7, 0.9);
|
||||
}
|
||||
|
||||
.connection-icon::before {
|
||||
content: '●';
|
||||
display: inline-block;
|
||||
animation: pulse 2s infinite;
|
||||
}
|
||||
|
||||
.connection-status.connecting .connection-icon::before {
|
||||
animation: spin 1s linear infinite;
|
||||
content: '↻';
|
||||
}
|
||||
|
||||
@keyframes pulse {
|
||||
0%, 100% { opacity: 1; }
|
||||
50% { opacity: 0.5; }
|
||||
}
|
||||
|
||||
@keyframes spin {
|
||||
from { transform: rotate(0deg); }
|
||||
to { transform: rotate(360deg); }
|
||||
}
|
||||
|
||||
.offline-message {
|
||||
position: fixed;
|
||||
top: 0;
|
||||
left: 0;
|
||||
right: 0;
|
||||
background: #dc3545;
|
||||
color: white;
|
||||
padding: 12px;
|
||||
text-align: center;
|
||||
z-index: 1001;
|
||||
transform: translateY(-100%);
|
||||
transition: transform 0.3s ease;
|
||||
}
|
||||
|
||||
.offline-message.show {
|
||||
transform: translateY(0);
|
||||
}
|
||||
|
||||
.offline-content {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
justify-content: center;
|
||||
gap: 12px;
|
||||
flex-wrap: wrap;
|
||||
}
|
||||
|
||||
.offline-content i {
|
||||
font-size: 20px;
|
||||
}
|
||||
|
||||
.retry-connection {
|
||||
border-color: white;
|
||||
color: white;
|
||||
}
|
||||
|
||||
.retry-connection:hover {
|
||||
background: white;
|
||||
color: #dc3545;
|
||||
}
|
||||
|
||||
.queued-info {
|
||||
margin-left: 12px;
|
||||
}
|
||||
|
||||
.queued-count {
|
||||
opacity: 0.9;
|
||||
}
|
||||
|
||||
@media (max-width: 768px) {
|
||||
.connection-status {
|
||||
bottom: 10px;
|
||||
right: 10px;
|
||||
font-size: 12px;
|
||||
padding: 6px 12px;
|
||||
}
|
||||
|
||||
.offline-content {
|
||||
font-size: 14px;
|
||||
}
|
||||
}
|
||||
`;
|
||||
document.head.appendChild(style);
|
||||
}
|
||||
}
|
||||
|
||||
private subscribeToConnectionChanges(): void {
|
||||
this.connectionManager.subscribe('connection-ui', (state: ConnectionState) => {
|
||||
this.updateUI(state);
|
||||
});
|
||||
}
|
||||
|
||||
private updateUI(state: ConnectionState): void {
|
||||
if (!this.statusElement || !this.offlineMessage) return;
|
||||
|
||||
const statusText = this.statusElement.querySelector('.connection-text') as HTMLElement;
|
||||
|
||||
// Update status element
|
||||
this.statusElement.className = `connection-status visible ${state.status}`;
|
||||
|
||||
switch (state.status) {
|
||||
case 'online':
|
||||
statusText.textContent = 'Connected';
|
||||
this.hideOfflineMessage();
|
||||
// Hide status after 3 seconds when online
|
||||
setTimeout(() => {
|
||||
if (this.connectionManager.getState().status === 'online') {
|
||||
this.statusElement?.classList.remove('visible');
|
||||
}
|
||||
}, 3000);
|
||||
break;
|
||||
|
||||
case 'offline':
|
||||
statusText.textContent = 'Offline';
|
||||
this.showOfflineMessage();
|
||||
this.updateQueuedInfo();
|
||||
break;
|
||||
|
||||
case 'connecting':
|
||||
statusText.textContent = 'Reconnecting...';
|
||||
if (this.retryButton) {
|
||||
this.retryButton.disabled = true;
|
||||
this.retryButton.innerHTML = '<i class="fas fa-spinner fa-spin"></i> Connecting...';
|
||||
}
|
||||
break;
|
||||
|
||||
case 'error':
|
||||
statusText.textContent = `Connection error${state.retryCount > 0 ? ` (Retry ${state.retryCount})` : ''}`;
|
||||
this.showOfflineMessage();
|
||||
this.updateQueuedInfo();
|
||||
if (this.retryButton) {
|
||||
this.retryButton.disabled = false;
|
||||
this.retryButton.innerHTML = '<i class="fas fa-sync"></i> Retry';
|
||||
}
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
private showOfflineMessage(): void {
|
||||
if (this.offlineMessage) {
|
||||
this.offlineMessage.style.display = 'block';
|
||||
setTimeout(() => {
|
||||
this.offlineMessage?.classList.add('show');
|
||||
}, 10);
|
||||
}
|
||||
}
|
||||
|
||||
private hideOfflineMessage(): void {
|
||||
if (this.offlineMessage) {
|
||||
this.offlineMessage.classList.remove('show');
|
||||
setTimeout(() => {
|
||||
if (this.offlineMessage) {
|
||||
this.offlineMessage.style.display = 'none';
|
||||
}
|
||||
}, 300);
|
||||
}
|
||||
}
|
||||
|
||||
private updateQueuedInfo(): void {
|
||||
const queueStatus = this.queueManager.getStatus();
|
||||
const queuedByType = this.queueManager.getQueuedByType();
|
||||
const queuedInfo = this.offlineMessage?.querySelector('.queued-info') as HTMLElement;
|
||||
const queuedCount = this.offlineMessage?.querySelector('.queued-count') as HTMLElement;
|
||||
|
||||
if (queuedInfo && queuedCount) {
|
||||
const totalQueued = queueStatus.queueLength + queueStatus.activeRequests;
|
||||
|
||||
if (totalQueued > 0) {
|
||||
queuedInfo.style.display = 'block';
|
||||
const parts = [];
|
||||
|
||||
if (queuedByType.transcribe > 0) {
|
||||
parts.push(`${queuedByType.transcribe} transcription${queuedByType.transcribe > 1 ? 's' : ''}`);
|
||||
}
|
||||
if (queuedByType.translate > 0) {
|
||||
parts.push(`${queuedByType.translate} translation${queuedByType.translate > 1 ? 's' : ''}`);
|
||||
}
|
||||
if (queuedByType.tts > 0) {
|
||||
parts.push(`${queuedByType.tts} audio generation${queuedByType.tts > 1 ? 's' : ''}`);
|
||||
}
|
||||
|
||||
queuedCount.textContent = `${totalQueued} request${totalQueued > 1 ? 's' : ''} queued${parts.length > 0 ? ': ' + parts.join(', ') : ''}`;
|
||||
} else {
|
||||
queuedInfo.style.display = 'none';
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
private async handleRetry(): Promise<void> {
|
||||
if (this.retryButton) {
|
||||
this.retryButton.disabled = true;
|
||||
this.retryButton.innerHTML = '<i class="fas fa-spinner fa-spin"></i> Connecting...';
|
||||
}
|
||||
|
||||
const success = await this.connectionManager.reconnect();
|
||||
|
||||
if (!success && this.retryButton) {
|
||||
this.retryButton.disabled = false;
|
||||
this.retryButton.innerHTML = '<i class="fas fa-sync"></i> Retry';
|
||||
}
|
||||
}
|
||||
|
||||
// Public method to show temporary connection message
|
||||
showTemporaryMessage(message: string, type: 'success' | 'error' | 'warning' = 'success'): void {
|
||||
if (!this.statusElement) return;
|
||||
|
||||
const statusText = this.statusElement.querySelector('.connection-text') as HTMLElement;
|
||||
const originalClass = this.statusElement.className;
|
||||
const originalText = statusText.textContent;
|
||||
|
||||
// Update appearance based on type
|
||||
this.statusElement.className = `connection-status visible ${type === 'success' ? 'online' : type === 'error' ? 'offline' : 'connecting'}`;
|
||||
statusText.textContent = message;
|
||||
|
||||
// Reset after 3 seconds
|
||||
setTimeout(() => {
|
||||
if (this.statusElement && statusText) {
|
||||
this.statusElement.className = originalClass;
|
||||
statusText.textContent = originalText || '';
|
||||
}
|
||||
}, 3000);
|
||||
}
|
||||
}
|
286
static/js/src/errorBoundary.ts
Normal file
286
static/js/src/errorBoundary.ts
Normal file
@ -0,0 +1,286 @@
|
||||
// Error boundary implementation for better error handling
|
||||
export interface ErrorInfo {
|
||||
message: string;
|
||||
stack?: string;
|
||||
component?: string;
|
||||
timestamp: number;
|
||||
userAgent: string;
|
||||
url: string;
|
||||
}
|
||||
|
||||
export class ErrorBoundary {
|
||||
private static instance: ErrorBoundary;
|
||||
private errorLog: ErrorInfo[] = [];
|
||||
private maxErrorLog = 50;
|
||||
private errorHandlers: Map<string, (error: Error, errorInfo: ErrorInfo) => void> = new Map();
|
||||
private globalErrorHandler: ((error: Error, errorInfo: ErrorInfo) => void) | null = null;
|
||||
|
||||
private constructor() {
|
||||
this.setupGlobalErrorHandlers();
|
||||
}
|
||||
|
||||
static getInstance(): ErrorBoundary {
|
||||
if (!ErrorBoundary.instance) {
|
||||
ErrorBoundary.instance = new ErrorBoundary();
|
||||
}
|
||||
return ErrorBoundary.instance;
|
||||
}
|
||||
|
||||
private setupGlobalErrorHandlers(): void {
|
||||
// Handle unhandled errors
|
||||
window.addEventListener('error', (event: ErrorEvent) => {
|
||||
const errorInfo: ErrorInfo = {
|
||||
message: event.message,
|
||||
stack: event.error?.stack,
|
||||
timestamp: Date.now(),
|
||||
userAgent: navigator.userAgent,
|
||||
url: window.location.href,
|
||||
component: 'global'
|
||||
};
|
||||
|
||||
this.logError(event.error || new Error(event.message), errorInfo);
|
||||
this.handleError(event.error || new Error(event.message), errorInfo);
|
||||
|
||||
// Prevent default error handling
|
||||
event.preventDefault();
|
||||
});
|
||||
|
||||
// Handle unhandled promise rejections
|
||||
window.addEventListener('unhandledrejection', (event: PromiseRejectionEvent) => {
|
||||
const error = new Error(event.reason?.message || 'Unhandled Promise Rejection');
|
||||
const errorInfo: ErrorInfo = {
|
||||
message: error.message,
|
||||
stack: event.reason?.stack,
|
||||
timestamp: Date.now(),
|
||||
userAgent: navigator.userAgent,
|
||||
url: window.location.href,
|
||||
component: 'promise'
|
||||
};
|
||||
|
||||
this.logError(error, errorInfo);
|
||||
this.handleError(error, errorInfo);
|
||||
|
||||
// Prevent default error handling
|
||||
event.preventDefault();
|
||||
});
|
||||
}
|
||||
|
||||
// Wrap a function with error boundary
|
||||
wrap<T extends (...args: any[]) => any>(
|
||||
fn: T,
|
||||
component: string,
|
||||
fallback?: (...args: Parameters<T>) => ReturnType<T>
|
||||
): T {
|
||||
return ((...args: Parameters<T>): ReturnType<T> => {
|
||||
try {
|
||||
const result = fn(...args);
|
||||
|
||||
// Handle async functions
|
||||
if (result instanceof Promise) {
|
||||
return result.catch((error: Error) => {
|
||||
const errorInfo: ErrorInfo = {
|
||||
message: error.message,
|
||||
stack: error.stack,
|
||||
component,
|
||||
timestamp: Date.now(),
|
||||
userAgent: navigator.userAgent,
|
||||
url: window.location.href
|
||||
};
|
||||
|
||||
this.logError(error, errorInfo);
|
||||
this.handleError(error, errorInfo);
|
||||
|
||||
if (fallback) {
|
||||
return fallback(...args) as ReturnType<T>;
|
||||
}
|
||||
throw error;
|
||||
}) as ReturnType<T>;
|
||||
}
|
||||
|
||||
return result;
|
||||
} catch (error: any) {
|
||||
const errorInfo: ErrorInfo = {
|
||||
message: error.message,
|
||||
stack: error.stack,
|
||||
component,
|
||||
timestamp: Date.now(),
|
||||
userAgent: navigator.userAgent,
|
||||
url: window.location.href
|
||||
};
|
||||
|
||||
this.logError(error, errorInfo);
|
||||
this.handleError(error, errorInfo);
|
||||
|
||||
if (fallback) {
|
||||
return fallback(...args);
|
||||
}
|
||||
throw error;
|
||||
}
|
||||
}) as T;
|
||||
}
|
||||
|
||||
// Wrap async functions specifically
|
||||
wrapAsync<T extends (...args: any[]) => Promise<any>>(
|
||||
fn: T,
|
||||
component: string,
|
||||
fallback?: (...args: Parameters<T>) => ReturnType<T>
|
||||
): T {
|
||||
return (async (...args: Parameters<T>) => {
|
||||
try {
|
||||
return await fn(...args);
|
||||
} catch (error: any) {
|
||||
const errorInfo: ErrorInfo = {
|
||||
message: error.message,
|
||||
stack: error.stack,
|
||||
component,
|
||||
timestamp: Date.now(),
|
||||
userAgent: navigator.userAgent,
|
||||
url: window.location.href
|
||||
};
|
||||
|
||||
this.logError(error, errorInfo);
|
||||
this.handleError(error, errorInfo);
|
||||
|
||||
if (fallback) {
|
||||
return fallback(...args);
|
||||
}
|
||||
throw error;
|
||||
}
|
||||
}) as T;
|
||||
}
|
||||
|
||||
// Register component-specific error handler
|
||||
registerErrorHandler(component: string, handler: (error: Error, errorInfo: ErrorInfo) => void): void {
|
||||
this.errorHandlers.set(component, handler);
|
||||
}
|
||||
|
||||
// Set global error handler
|
||||
setGlobalErrorHandler(handler: (error: Error, errorInfo: ErrorInfo) => void): void {
|
||||
this.globalErrorHandler = handler;
|
||||
}
|
||||
|
||||
private logError(error: Error, errorInfo: ErrorInfo): void {
|
||||
// Add to error log
|
||||
this.errorLog.push(errorInfo);
|
||||
|
||||
// Keep only recent errors
|
||||
if (this.errorLog.length > this.maxErrorLog) {
|
||||
this.errorLog.shift();
|
||||
}
|
||||
|
||||
// Log to console in development
|
||||
console.error(`[${errorInfo.component}] Error:`, error);
|
||||
console.error('Error Info:', errorInfo);
|
||||
|
||||
// Send to monitoring service if available
|
||||
this.sendToMonitoring(error, errorInfo);
|
||||
}
|
||||
|
||||
private handleError(error: Error, errorInfo: ErrorInfo): void {
|
||||
// Check for component-specific handler
|
||||
const componentHandler = this.errorHandlers.get(errorInfo.component || '');
|
||||
if (componentHandler) {
|
||||
componentHandler(error, errorInfo);
|
||||
return;
|
||||
}
|
||||
|
||||
// Use global handler if set
|
||||
if (this.globalErrorHandler) {
|
||||
this.globalErrorHandler(error, errorInfo);
|
||||
return;
|
||||
}
|
||||
|
||||
// Default error handling
|
||||
this.showErrorNotification(error, errorInfo);
|
||||
}
|
||||
|
||||
private showErrorNotification(error: Error, errorInfo: ErrorInfo): void {
|
||||
// Create error notification
|
||||
const notification = document.createElement('div');
|
||||
notification.className = 'alert alert-danger alert-dismissible fade show position-fixed bottom-0 end-0 m-3';
|
||||
notification.style.zIndex = '9999';
|
||||
notification.style.maxWidth = '400px';
|
||||
|
||||
const isUserFacing = this.isUserFacingError(error);
|
||||
const message = isUserFacing ? error.message : 'An unexpected error occurred. Please try again.';
|
||||
|
||||
notification.innerHTML = `
|
||||
<strong><i class="fas fa-exclamation-circle"></i> Error${errorInfo.component ? ` in ${errorInfo.component}` : ''}</strong>
|
||||
<p class="mb-0">${message}</p>
|
||||
${!isUserFacing ? '<small class="text-muted">The error has been logged for investigation.</small>' : ''}
|
||||
<button type="button" class="btn-close" data-bs-dismiss="alert"></button>
|
||||
`;
|
||||
|
||||
document.body.appendChild(notification);
|
||||
|
||||
// Auto-dismiss after 10 seconds
|
||||
setTimeout(() => {
|
||||
if (notification.parentNode) {
|
||||
notification.remove();
|
||||
}
|
||||
}, 10000);
|
||||
}
|
||||
|
||||
private isUserFacingError(error: Error): boolean {
|
||||
// Determine if error should be shown to user as-is
|
||||
const userFacingMessages = [
|
||||
'rate limit',
|
||||
'network',
|
||||
'offline',
|
||||
'not found',
|
||||
'unauthorized',
|
||||
'forbidden',
|
||||
'timeout',
|
||||
'invalid'
|
||||
];
|
||||
|
||||
const message = error.message.toLowerCase();
|
||||
return userFacingMessages.some(msg => message.includes(msg));
|
||||
}
|
||||
|
||||
private async sendToMonitoring(error: Error, errorInfo: ErrorInfo): Promise<void> {
|
||||
// Only send errors in production
|
||||
if (window.location.hostname === 'localhost' || window.location.hostname === '127.0.0.1') {
|
||||
return;
|
||||
}
|
||||
|
||||
try {
|
||||
// Send error to backend monitoring endpoint
|
||||
await fetch('/api/log-error', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
error: {
|
||||
message: error.message,
|
||||
stack: error.stack,
|
||||
name: error.name
|
||||
},
|
||||
errorInfo
|
||||
})
|
||||
});
|
||||
} catch (monitoringError) {
|
||||
// Fail silently - don't create error loop
|
||||
console.error('Failed to send error to monitoring:', monitoringError);
|
||||
}
|
||||
}
|
||||
|
||||
// Get error log for debugging
|
||||
getErrorLog(): ErrorInfo[] {
|
||||
return [...this.errorLog];
|
||||
}
|
||||
|
||||
// Clear error log
|
||||
clearErrorLog(): void {
|
||||
this.errorLog = [];
|
||||
}
|
||||
|
||||
// Check if component has recent errors
|
||||
hasRecentErrors(component: string, timeWindow: number = 60000): boolean {
|
||||
const cutoff = Date.now() - timeWindow;
|
||||
return this.errorLog.some(
|
||||
error => error.component === component && error.timestamp > cutoff
|
||||
);
|
||||
}
|
||||
}
|
309
static/js/src/memoryManager.ts
Normal file
309
static/js/src/memoryManager.ts
Normal file
@ -0,0 +1,309 @@
|
||||
/**
|
||||
* Memory management utilities for preventing leaks in audio handling
|
||||
*/
|
||||
|
||||
export class MemoryManager {
|
||||
private static instance: MemoryManager;
|
||||
private audioContexts: Set<AudioContext> = new Set();
|
||||
private objectURLs: Set<string> = new Set();
|
||||
private mediaStreams: Set<MediaStream> = new Set();
|
||||
private intervals: Set<number> = new Set();
|
||||
private timeouts: Set<number> = new Set();
|
||||
|
||||
private constructor() {
|
||||
// Set up periodic cleanup
|
||||
this.startPeriodicCleanup();
|
||||
|
||||
// Clean up on page unload
|
||||
window.addEventListener('beforeunload', () => this.cleanup());
|
||||
}
|
||||
|
||||
static getInstance(): MemoryManager {
|
||||
if (!MemoryManager.instance) {
|
||||
MemoryManager.instance = new MemoryManager();
|
||||
}
|
||||
return MemoryManager.instance;
|
||||
}
|
||||
|
||||
/**
|
||||
* Register an AudioContext for cleanup
|
||||
*/
|
||||
registerAudioContext(context: AudioContext): void {
|
||||
this.audioContexts.add(context);
|
||||
}
|
||||
|
||||
/**
|
||||
* Register an object URL for cleanup
|
||||
*/
|
||||
registerObjectURL(url: string): void {
|
||||
this.objectURLs.add(url);
|
||||
}
|
||||
|
||||
/**
|
||||
* Register a MediaStream for cleanup
|
||||
*/
|
||||
registerMediaStream(stream: MediaStream): void {
|
||||
this.mediaStreams.add(stream);
|
||||
}
|
||||
|
||||
/**
|
||||
* Register an interval for cleanup
|
||||
*/
|
||||
registerInterval(id: number): void {
|
||||
this.intervals.add(id);
|
||||
}
|
||||
|
||||
/**
|
||||
* Register a timeout for cleanup
|
||||
*/
|
||||
registerTimeout(id: number): void {
|
||||
this.timeouts.add(id);
|
||||
}
|
||||
|
||||
/**
|
||||
* Clean up a specific AudioContext
|
||||
*/
|
||||
cleanupAudioContext(context: AudioContext): void {
|
||||
if (context.state !== 'closed') {
|
||||
context.close().catch(console.error);
|
||||
}
|
||||
this.audioContexts.delete(context);
|
||||
}
|
||||
|
||||
/**
|
||||
* Clean up a specific object URL
|
||||
*/
|
||||
cleanupObjectURL(url: string): void {
|
||||
URL.revokeObjectURL(url);
|
||||
this.objectURLs.delete(url);
|
||||
}
|
||||
|
||||
/**
|
||||
* Clean up a specific MediaStream
|
||||
*/
|
||||
cleanupMediaStream(stream: MediaStream): void {
|
||||
stream.getTracks().forEach(track => {
|
||||
track.stop();
|
||||
});
|
||||
this.mediaStreams.delete(stream);
|
||||
}
|
||||
|
||||
/**
|
||||
* Clean up all resources
|
||||
*/
|
||||
cleanup(): void {
|
||||
// Clean up audio contexts
|
||||
this.audioContexts.forEach(context => {
|
||||
if (context.state !== 'closed') {
|
||||
context.close().catch(console.error);
|
||||
}
|
||||
});
|
||||
this.audioContexts.clear();
|
||||
|
||||
// Clean up object URLs
|
||||
this.objectURLs.forEach(url => {
|
||||
URL.revokeObjectURL(url);
|
||||
});
|
||||
this.objectURLs.clear();
|
||||
|
||||
// Clean up media streams
|
||||
this.mediaStreams.forEach(stream => {
|
||||
stream.getTracks().forEach(track => {
|
||||
track.stop();
|
||||
});
|
||||
});
|
||||
this.mediaStreams.clear();
|
||||
|
||||
// Clear intervals and timeouts
|
||||
this.intervals.forEach(id => clearInterval(id));
|
||||
this.intervals.clear();
|
||||
|
||||
this.timeouts.forEach(id => clearTimeout(id));
|
||||
this.timeouts.clear();
|
||||
|
||||
console.log('Memory cleanup completed');
|
||||
}
|
||||
|
||||
/**
|
||||
* Get memory usage statistics
|
||||
*/
|
||||
getStats(): MemoryStats {
|
||||
return {
|
||||
audioContexts: this.audioContexts.size,
|
||||
objectURLs: this.objectURLs.size,
|
||||
mediaStreams: this.mediaStreams.size,
|
||||
intervals: this.intervals.size,
|
||||
timeouts: this.timeouts.size
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* Start periodic cleanup of orphaned resources
|
||||
*/
|
||||
private startPeriodicCleanup(): void {
|
||||
setInterval(() => {
|
||||
// Clean up closed audio contexts
|
||||
this.audioContexts.forEach(context => {
|
||||
if (context.state === 'closed') {
|
||||
this.audioContexts.delete(context);
|
||||
}
|
||||
});
|
||||
|
||||
// Clean up stopped media streams
|
||||
this.mediaStreams.forEach(stream => {
|
||||
const activeTracks = stream.getTracks().filter(track => track.readyState === 'live');
|
||||
if (activeTracks.length === 0) {
|
||||
this.mediaStreams.delete(stream);
|
||||
}
|
||||
});
|
||||
|
||||
// Log stats in development
|
||||
if (process.env.NODE_ENV === 'development') {
|
||||
const stats = this.getStats();
|
||||
if (Object.values(stats).some(v => v > 0)) {
|
||||
console.log('Memory manager stats:', stats);
|
||||
}
|
||||
}
|
||||
}, 30000); // Every 30 seconds
|
||||
|
||||
// Don't track this interval to avoid self-reference
|
||||
// It will be cleared on page unload
|
||||
}
|
||||
}
|
||||
|
||||
interface MemoryStats {
|
||||
audioContexts: number;
|
||||
objectURLs: number;
|
||||
mediaStreams: number;
|
||||
intervals: number;
|
||||
timeouts: number;
|
||||
}
|
||||
|
||||
/**
|
||||
* Wrapper for safe audio blob handling
|
||||
*/
|
||||
export class AudioBlobHandler {
|
||||
private blob: Blob;
|
||||
private objectURL?: string;
|
||||
private memoryManager: MemoryManager;
|
||||
|
||||
constructor(blob: Blob) {
|
||||
this.blob = blob;
|
||||
this.memoryManager = MemoryManager.getInstance();
|
||||
}
|
||||
|
||||
/**
|
||||
* Get object URL (creates one if needed)
|
||||
*/
|
||||
getObjectURL(): string {
|
||||
if (!this.objectURL) {
|
||||
this.objectURL = URL.createObjectURL(this.blob);
|
||||
this.memoryManager.registerObjectURL(this.objectURL);
|
||||
}
|
||||
return this.objectURL;
|
||||
}
|
||||
|
||||
/**
|
||||
* Get the blob
|
||||
*/
|
||||
getBlob(): Blob {
|
||||
return this.blob;
|
||||
}
|
||||
|
||||
/**
|
||||
* Clean up resources
|
||||
*/
|
||||
cleanup(): void {
|
||||
if (this.objectURL) {
|
||||
this.memoryManager.cleanupObjectURL(this.objectURL);
|
||||
this.objectURL = undefined;
|
||||
}
|
||||
// Help garbage collection
|
||||
(this.blob as any) = null;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Safe MediaRecorder wrapper
|
||||
*/
|
||||
export class SafeMediaRecorder {
|
||||
private mediaRecorder?: MediaRecorder;
|
||||
private stream?: MediaStream;
|
||||
private chunks: Blob[] = [];
|
||||
private memoryManager: MemoryManager;
|
||||
|
||||
constructor() {
|
||||
this.memoryManager = MemoryManager.getInstance();
|
||||
}
|
||||
|
||||
async start(constraints: MediaStreamConstraints = { audio: true }): Promise<void> {
|
||||
// Clean up any existing recorder
|
||||
this.cleanup();
|
||||
|
||||
this.stream = await navigator.mediaDevices.getUserMedia(constraints);
|
||||
this.memoryManager.registerMediaStream(this.stream);
|
||||
|
||||
const options = {
|
||||
mimeType: MediaRecorder.isTypeSupported('audio/webm;codecs=opus')
|
||||
? 'audio/webm;codecs=opus'
|
||||
: 'audio/webm'
|
||||
};
|
||||
|
||||
this.mediaRecorder = new MediaRecorder(this.stream, options);
|
||||
this.chunks = [];
|
||||
|
||||
this.mediaRecorder.ondataavailable = (event) => {
|
||||
if (event.data.size > 0) {
|
||||
this.chunks.push(event.data);
|
||||
}
|
||||
};
|
||||
|
||||
this.mediaRecorder.start();
|
||||
}
|
||||
|
||||
stop(): Promise<Blob> {
|
||||
return new Promise((resolve, reject) => {
|
||||
if (!this.mediaRecorder) {
|
||||
reject(new Error('MediaRecorder not initialized'));
|
||||
return;
|
||||
}
|
||||
|
||||
this.mediaRecorder.onstop = () => {
|
||||
const blob = new Blob(this.chunks, {
|
||||
type: this.mediaRecorder?.mimeType || 'audio/webm'
|
||||
});
|
||||
resolve(blob);
|
||||
|
||||
// Clean up after delivering the blob
|
||||
setTimeout(() => this.cleanup(), 100);
|
||||
};
|
||||
|
||||
this.mediaRecorder.stop();
|
||||
});
|
||||
}
|
||||
|
||||
cleanup(): void {
|
||||
if (this.stream) {
|
||||
this.memoryManager.cleanupMediaStream(this.stream);
|
||||
this.stream = undefined;
|
||||
}
|
||||
|
||||
if (this.mediaRecorder) {
|
||||
if (this.mediaRecorder.state !== 'inactive') {
|
||||
try {
|
||||
this.mediaRecorder.stop();
|
||||
} catch (e) {
|
||||
// Ignore errors
|
||||
}
|
||||
}
|
||||
this.mediaRecorder = undefined;
|
||||
}
|
||||
|
||||
// Clear chunks
|
||||
this.chunks = [];
|
||||
}
|
||||
|
||||
isRecording(): boolean {
|
||||
return this.mediaRecorder?.state === 'recording';
|
||||
}
|
||||
}
|
147
static/js/src/performanceMonitor.ts
Normal file
147
static/js/src/performanceMonitor.ts
Normal file
@ -0,0 +1,147 @@
|
||||
// Performance monitoring for translation latency
|
||||
export class PerformanceMonitor {
|
||||
private static instance: PerformanceMonitor;
|
||||
private metrics: Map<string, number[]> = new Map();
|
||||
private timers: Map<string, number> = new Map();
|
||||
|
||||
private constructor() {}
|
||||
|
||||
static getInstance(): PerformanceMonitor {
|
||||
if (!PerformanceMonitor.instance) {
|
||||
PerformanceMonitor.instance = new PerformanceMonitor();
|
||||
}
|
||||
return PerformanceMonitor.instance;
|
||||
}
|
||||
|
||||
// Start timing an operation
|
||||
startTimer(operation: string): void {
|
||||
this.timers.set(operation, performance.now());
|
||||
}
|
||||
|
||||
// End timing and record the duration
|
||||
endTimer(operation: string): number {
|
||||
const startTime = this.timers.get(operation);
|
||||
if (!startTime) {
|
||||
console.warn(`No start time found for operation: ${operation}`);
|
||||
return 0;
|
||||
}
|
||||
|
||||
const duration = performance.now() - startTime;
|
||||
this.recordMetric(operation, duration);
|
||||
this.timers.delete(operation);
|
||||
|
||||
return duration;
|
||||
}
|
||||
|
||||
// Record a metric value
|
||||
recordMetric(name: string, value: number): void {
|
||||
if (!this.metrics.has(name)) {
|
||||
this.metrics.set(name, []);
|
||||
}
|
||||
|
||||
const values = this.metrics.get(name)!;
|
||||
values.push(value);
|
||||
|
||||
// Keep only last 100 values
|
||||
if (values.length > 100) {
|
||||
values.shift();
|
||||
}
|
||||
}
|
||||
|
||||
// Get average metric value
|
||||
getAverageMetric(name: string): number {
|
||||
const values = this.metrics.get(name);
|
||||
if (!values || values.length === 0) {
|
||||
return 0;
|
||||
}
|
||||
|
||||
const sum = values.reduce((a, b) => a + b, 0);
|
||||
return sum / values.length;
|
||||
}
|
||||
|
||||
// Get time to first byte (TTFB) for streaming
|
||||
measureTTFB(operation: string, firstByteTime: number): number {
|
||||
const startTime = this.timers.get(operation);
|
||||
if (!startTime) {
|
||||
return 0;
|
||||
}
|
||||
|
||||
const ttfb = firstByteTime - startTime;
|
||||
this.recordMetric(`${operation}_ttfb`, ttfb);
|
||||
return ttfb;
|
||||
}
|
||||
|
||||
// Get performance summary
|
||||
getPerformanceSummary(): {
|
||||
streaming: {
|
||||
avgTotalTime: number;
|
||||
avgTTFB: number;
|
||||
count: number;
|
||||
};
|
||||
regular: {
|
||||
avgTotalTime: number;
|
||||
count: number;
|
||||
};
|
||||
improvement: {
|
||||
ttfbReduction: number;
|
||||
perceivedLatencyReduction: number;
|
||||
};
|
||||
} {
|
||||
const streamingTotal = this.getAverageMetric('streaming_translation');
|
||||
const streamingTTFB = this.getAverageMetric('streaming_translation_ttfb');
|
||||
const streamingCount = this.metrics.get('streaming_translation')?.length || 0;
|
||||
|
||||
const regularTotal = this.getAverageMetric('regular_translation');
|
||||
const regularCount = this.metrics.get('regular_translation')?.length || 0;
|
||||
|
||||
// Calculate improvements
|
||||
const ttfbReduction = regularTotal > 0 && streamingTTFB > 0
|
||||
? ((regularTotal - streamingTTFB) / regularTotal) * 100
|
||||
: 0;
|
||||
|
||||
// Perceived latency is based on TTFB for streaming vs total time for regular
|
||||
const perceivedLatencyReduction = ttfbReduction;
|
||||
|
||||
return {
|
||||
streaming: {
|
||||
avgTotalTime: streamingTotal,
|
||||
avgTTFB: streamingTTFB,
|
||||
count: streamingCount
|
||||
},
|
||||
regular: {
|
||||
avgTotalTime: regularTotal,
|
||||
count: regularCount
|
||||
},
|
||||
improvement: {
|
||||
ttfbReduction: Math.round(ttfbReduction),
|
||||
perceivedLatencyReduction: Math.round(perceivedLatencyReduction)
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
// Log performance stats to console
|
||||
logPerformanceStats(): void {
|
||||
const summary = this.getPerformanceSummary();
|
||||
|
||||
console.group('Translation Performance Stats');
|
||||
console.log('Streaming Translation:');
|
||||
console.log(` Average Total Time: ${summary.streaming.avgTotalTime.toFixed(2)}ms`);
|
||||
console.log(` Average TTFB: ${summary.streaming.avgTTFB.toFixed(2)}ms`);
|
||||
console.log(` Sample Count: ${summary.streaming.count}`);
|
||||
|
||||
console.log('Regular Translation:');
|
||||
console.log(` Average Total Time: ${summary.regular.avgTotalTime.toFixed(2)}ms`);
|
||||
console.log(` Sample Count: ${summary.regular.count}`);
|
||||
|
||||
console.log('Improvements:');
|
||||
console.log(` TTFB Reduction: ${summary.improvement.ttfbReduction}%`);
|
||||
console.log(` Perceived Latency Reduction: ${summary.improvement.perceivedLatencyReduction}%`);
|
||||
console.groupEnd();
|
||||
}
|
||||
|
||||
// Clear all metrics
|
||||
clearMetrics(): void {
|
||||
this.metrics.clear();
|
||||
this.timers.clear();
|
||||
}
|
||||
}
|
333
static/js/src/requestQueue.ts
Normal file
333
static/js/src/requestQueue.ts
Normal file
@ -0,0 +1,333 @@
|
||||
// Request queue and throttling manager
|
||||
import { ConnectionManager, ConnectionState } from './connectionManager';
|
||||
|
||||
export interface QueuedRequest {
|
||||
id: string;
|
||||
type: 'transcribe' | 'translate' | 'tts';
|
||||
request: () => Promise<any>;
|
||||
resolve: (value: any) => void;
|
||||
reject: (reason?: any) => void;
|
||||
retryCount: number;
|
||||
priority: number;
|
||||
timestamp: number;
|
||||
}
|
||||
|
||||
export class RequestQueueManager {
|
||||
private static instance: RequestQueueManager;
|
||||
private queue: QueuedRequest[] = [];
|
||||
private activeRequests: Map<string, QueuedRequest> = new Map();
|
||||
private maxConcurrent = 2; // Maximum concurrent requests
|
||||
private maxRetries = 3;
|
||||
private retryDelay = 1000; // Base retry delay in ms
|
||||
private isProcessing = false;
|
||||
private connectionManager: ConnectionManager;
|
||||
private isPaused = false;
|
||||
|
||||
// Rate limiting
|
||||
private requestHistory: number[] = [];
|
||||
private maxRequestsPerMinute = 30;
|
||||
private maxRequestsPerSecond = 2;
|
||||
|
||||
private constructor() {
|
||||
this.connectionManager = ConnectionManager.getInstance();
|
||||
|
||||
// Subscribe to connection state changes
|
||||
this.connectionManager.subscribe('request-queue', (state: ConnectionState) => {
|
||||
this.handleConnectionStateChange(state);
|
||||
});
|
||||
|
||||
// Start processing queue
|
||||
this.startProcessing();
|
||||
}
|
||||
|
||||
static getInstance(): RequestQueueManager {
|
||||
if (!RequestQueueManager.instance) {
|
||||
RequestQueueManager.instance = new RequestQueueManager();
|
||||
}
|
||||
return RequestQueueManager.instance;
|
||||
}
|
||||
|
||||
// Add request to queue
|
||||
async enqueue<T>(
|
||||
type: 'transcribe' | 'translate' | 'tts',
|
||||
request: () => Promise<T>,
|
||||
priority: number = 5
|
||||
): Promise<T> {
|
||||
// Check rate limits
|
||||
if (!this.checkRateLimits()) {
|
||||
throw new Error('Rate limit exceeded. Please slow down.');
|
||||
}
|
||||
|
||||
return new Promise((resolve, reject) => {
|
||||
const id = this.generateId();
|
||||
const queuedRequest: QueuedRequest = {
|
||||
id,
|
||||
type,
|
||||
request,
|
||||
resolve,
|
||||
reject,
|
||||
retryCount: 0,
|
||||
priority,
|
||||
timestamp: Date.now()
|
||||
};
|
||||
|
||||
// Add to queue based on priority
|
||||
this.addToQueue(queuedRequest);
|
||||
|
||||
// Log queue status
|
||||
console.log(`Request queued: ${type}, Queue size: ${this.queue.length}, Active: ${this.activeRequests.size}`);
|
||||
});
|
||||
}
|
||||
|
||||
private addToQueue(request: QueuedRequest): void {
|
||||
// Insert based on priority (higher priority first)
|
||||
const insertIndex = this.queue.findIndex(item => item.priority < request.priority);
|
||||
if (insertIndex === -1) {
|
||||
this.queue.push(request);
|
||||
} else {
|
||||
this.queue.splice(insertIndex, 0, request);
|
||||
}
|
||||
}
|
||||
|
||||
private checkRateLimits(): boolean {
|
||||
const now = Date.now();
|
||||
|
||||
// Clean old entries
|
||||
this.requestHistory = this.requestHistory.filter(
|
||||
time => now - time < 60000 // Keep last minute
|
||||
);
|
||||
|
||||
// Check per-second limit
|
||||
const lastSecond = this.requestHistory.filter(
|
||||
time => now - time < 1000
|
||||
).length;
|
||||
|
||||
if (lastSecond >= this.maxRequestsPerSecond) {
|
||||
console.warn('Per-second rate limit reached');
|
||||
return false;
|
||||
}
|
||||
|
||||
// Check per-minute limit
|
||||
if (this.requestHistory.length >= this.maxRequestsPerMinute) {
|
||||
console.warn('Per-minute rate limit reached');
|
||||
return false;
|
||||
}
|
||||
|
||||
// Record this request
|
||||
this.requestHistory.push(now);
|
||||
return true;
|
||||
}
|
||||
|
||||
private async startProcessing(): Promise<void> {
|
||||
if (this.isProcessing) return;
|
||||
this.isProcessing = true;
|
||||
|
||||
while (true) {
|
||||
await this.processQueue();
|
||||
await this.delay(100); // Check queue every 100ms
|
||||
}
|
||||
}
|
||||
|
||||
private async processQueue(): Promise<void> {
|
||||
// Check if we're paused or can't process more requests
|
||||
if (this.isPaused || this.activeRequests.size >= this.maxConcurrent || this.queue.length === 0) {
|
||||
return;
|
||||
}
|
||||
|
||||
// Check if we're online
|
||||
if (!this.connectionManager.isOnline()) {
|
||||
console.log('Queue processing paused - offline');
|
||||
return;
|
||||
}
|
||||
|
||||
// Get next request
|
||||
const request = this.queue.shift();
|
||||
if (!request) return;
|
||||
|
||||
// Mark as active
|
||||
this.activeRequests.set(request.id, request);
|
||||
|
||||
try {
|
||||
// Execute request with connection manager retry logic
|
||||
const result = await this.connectionManager.retryRequest(
|
||||
request.request,
|
||||
{
|
||||
retries: this.maxRetries - request.retryCount,
|
||||
delay: this.calculateRetryDelay(request.retryCount + 1),
|
||||
onRetry: (attempt, error) => {
|
||||
console.log(`Retry ${attempt} for ${request.type}: ${error.message}`);
|
||||
}
|
||||
}
|
||||
);
|
||||
|
||||
request.resolve(result);
|
||||
console.log(`Request completed: ${request.type}`);
|
||||
} catch (error) {
|
||||
console.error(`Request failed after retries: ${request.type}`, error);
|
||||
|
||||
// Check if it's a connection error and we should queue for later
|
||||
if (this.isConnectionError(error) && request.retryCount < this.maxRetries) {
|
||||
request.retryCount++;
|
||||
console.log(`Re-queuing ${request.type} due to connection error`);
|
||||
|
||||
// Re-queue with higher priority
|
||||
request.priority = Math.max(request.priority + 1, 10);
|
||||
this.addToQueue(request);
|
||||
} else {
|
||||
// Non-recoverable error or max retries reached
|
||||
request.reject(error);
|
||||
}
|
||||
} finally {
|
||||
// Remove from active
|
||||
this.activeRequests.delete(request.id);
|
||||
}
|
||||
}
|
||||
|
||||
// Note: shouldRetry logic is now handled by ConnectionManager
|
||||
// Keeping for reference but not used directly
|
||||
|
||||
private calculateRetryDelay(retryCount: number): number {
|
||||
// Exponential backoff with jitter
|
||||
const baseDelay = this.retryDelay * Math.pow(2, retryCount - 1);
|
||||
const jitter = Math.random() * 0.3 * baseDelay; // 30% jitter
|
||||
return Math.min(baseDelay + jitter, 30000); // Max 30 seconds
|
||||
}
|
||||
|
||||
private generateId(): string {
|
||||
return `${Date.now()}-${Math.random().toString(36).substr(2, 9)}`;
|
||||
}
|
||||
|
||||
private delay(ms: number): Promise<void> {
|
||||
return new Promise(resolve => setTimeout(resolve, ms));
|
||||
}
|
||||
|
||||
// Get queue status
|
||||
getStatus(): {
|
||||
queueLength: number;
|
||||
activeRequests: number;
|
||||
requestsPerMinute: number;
|
||||
} {
|
||||
const now = Date.now();
|
||||
const recentRequests = this.requestHistory.filter(
|
||||
time => now - time < 60000
|
||||
).length;
|
||||
|
||||
return {
|
||||
queueLength: this.queue.length,
|
||||
activeRequests: this.activeRequests.size,
|
||||
requestsPerMinute: recentRequests
|
||||
};
|
||||
}
|
||||
|
||||
// Clear queue (for emergency use)
|
||||
clearQueue(): void {
|
||||
this.queue.forEach(request => {
|
||||
request.reject(new Error('Queue cleared'));
|
||||
});
|
||||
this.queue = [];
|
||||
}
|
||||
|
||||
// Clear stuck requests (requests older than 60 seconds)
|
||||
clearStuckRequests(): void {
|
||||
const now = Date.now();
|
||||
const stuckThreshold = 60000; // 60 seconds
|
||||
|
||||
// Clear stuck active requests
|
||||
this.activeRequests.forEach((request, id) => {
|
||||
if (now - request.timestamp > stuckThreshold) {
|
||||
console.warn(`Clearing stuck active request: ${request.type}`);
|
||||
request.reject(new Error('Request timeout - cleared by recovery'));
|
||||
this.activeRequests.delete(id);
|
||||
}
|
||||
});
|
||||
|
||||
// Clear old queued requests
|
||||
this.queue = this.queue.filter(request => {
|
||||
if (now - request.timestamp > stuckThreshold) {
|
||||
console.warn(`Clearing stuck queued request: ${request.type}`);
|
||||
request.reject(new Error('Request timeout - cleared by recovery'));
|
||||
return false;
|
||||
}
|
||||
return true;
|
||||
});
|
||||
}
|
||||
|
||||
// Update settings
|
||||
updateSettings(settings: {
|
||||
maxConcurrent?: number;
|
||||
maxRequestsPerMinute?: number;
|
||||
maxRequestsPerSecond?: number;
|
||||
}): void {
|
||||
if (settings.maxConcurrent !== undefined) {
|
||||
this.maxConcurrent = settings.maxConcurrent;
|
||||
}
|
||||
if (settings.maxRequestsPerMinute !== undefined) {
|
||||
this.maxRequestsPerMinute = settings.maxRequestsPerMinute;
|
||||
}
|
||||
if (settings.maxRequestsPerSecond !== undefined) {
|
||||
this.maxRequestsPerSecond = settings.maxRequestsPerSecond;
|
||||
}
|
||||
}
|
||||
|
||||
// Handle connection state changes
|
||||
private handleConnectionStateChange(state: ConnectionState): void {
|
||||
console.log(`Connection state changed: ${state.status}`);
|
||||
|
||||
if (state.status === 'offline' || state.status === 'error') {
|
||||
// Pause processing when offline
|
||||
this.isPaused = true;
|
||||
|
||||
// Notify queued requests about offline status
|
||||
if (this.queue.length > 0) {
|
||||
console.log(`${this.queue.length} requests queued while offline`);
|
||||
}
|
||||
} else if (state.status === 'online') {
|
||||
// Resume processing when back online
|
||||
this.isPaused = false;
|
||||
console.log('Connection restored, resuming queue processing');
|
||||
|
||||
// Process any queued requests
|
||||
if (this.queue.length > 0) {
|
||||
console.log(`Processing ${this.queue.length} queued requests`);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Check if error is connection-related
|
||||
private isConnectionError(error: any): boolean {
|
||||
const errorMessage = error.message?.toLowerCase() || '';
|
||||
const connectionErrors = [
|
||||
'network',
|
||||
'fetch',
|
||||
'connection',
|
||||
'timeout',
|
||||
'offline',
|
||||
'cors'
|
||||
];
|
||||
|
||||
return connectionErrors.some(e => errorMessage.includes(e));
|
||||
}
|
||||
|
||||
// Pause queue processing
|
||||
pause(): void {
|
||||
this.isPaused = true;
|
||||
console.log('Request queue paused');
|
||||
}
|
||||
|
||||
// Resume queue processing
|
||||
resume(): void {
|
||||
this.isPaused = false;
|
||||
console.log('Request queue resumed');
|
||||
}
|
||||
|
||||
// Get number of queued requests by type
|
||||
getQueuedByType(): { transcribe: number; translate: number; tts: number } {
|
||||
const counts = { transcribe: 0, translate: 0, tts: 0 };
|
||||
|
||||
this.queue.forEach(request => {
|
||||
counts[request.type]++;
|
||||
});
|
||||
|
||||
return counts;
|
||||
}
|
||||
}
|
270
static/js/src/speakerManager.ts
Normal file
270
static/js/src/speakerManager.ts
Normal file
@ -0,0 +1,270 @@
|
||||
// Speaker management for multi-speaker support
|
||||
export interface Speaker {
|
||||
id: string;
|
||||
name: string;
|
||||
language: string;
|
||||
color: string;
|
||||
avatar?: string;
|
||||
isActive: boolean;
|
||||
lastActiveTime?: number;
|
||||
}
|
||||
|
||||
export interface SpeakerTranscription {
|
||||
speakerId: string;
|
||||
text: string;
|
||||
language: string;
|
||||
timestamp: number;
|
||||
}
|
||||
|
||||
export interface ConversationEntry {
|
||||
id: string;
|
||||
speakerId: string;
|
||||
originalText: string;
|
||||
originalLanguage: string;
|
||||
translations: Map<string, string>; // languageCode -> translatedText
|
||||
timestamp: number;
|
||||
audioUrl?: string;
|
||||
}
|
||||
|
||||
export class SpeakerManager {
|
||||
private static instance: SpeakerManager;
|
||||
private speakers: Map<string, Speaker> = new Map();
|
||||
private conversation: ConversationEntry[] = [];
|
||||
private activeSpeakerId: string | null = null;
|
||||
private maxConversationLength = 100;
|
||||
|
||||
// Predefined colors for speakers
|
||||
private speakerColors = [
|
||||
'#007bff', '#28a745', '#dc3545', '#ffc107',
|
||||
'#17a2b8', '#6f42c1', '#e83e8c', '#fd7e14'
|
||||
];
|
||||
|
||||
private constructor() {
|
||||
this.loadFromLocalStorage();
|
||||
}
|
||||
|
||||
static getInstance(): SpeakerManager {
|
||||
if (!SpeakerManager.instance) {
|
||||
SpeakerManager.instance = new SpeakerManager();
|
||||
}
|
||||
return SpeakerManager.instance;
|
||||
}
|
||||
|
||||
// Add a new speaker
|
||||
addSpeaker(name: string, language: string): Speaker {
|
||||
const id = this.generateSpeakerId();
|
||||
const colorIndex = this.speakers.size % this.speakerColors.length;
|
||||
|
||||
const speaker: Speaker = {
|
||||
id,
|
||||
name,
|
||||
language,
|
||||
color: this.speakerColors[colorIndex],
|
||||
isActive: false,
|
||||
avatar: this.generateAvatar(name)
|
||||
};
|
||||
|
||||
this.speakers.set(id, speaker);
|
||||
this.saveToLocalStorage();
|
||||
|
||||
return speaker;
|
||||
}
|
||||
|
||||
// Update speaker
|
||||
updateSpeaker(id: string, updates: Partial<Speaker>): void {
|
||||
const speaker = this.speakers.get(id);
|
||||
if (speaker) {
|
||||
Object.assign(speaker, updates);
|
||||
this.saveToLocalStorage();
|
||||
}
|
||||
}
|
||||
|
||||
// Remove speaker
|
||||
removeSpeaker(id: string): void {
|
||||
this.speakers.delete(id);
|
||||
if (this.activeSpeakerId === id) {
|
||||
this.activeSpeakerId = null;
|
||||
}
|
||||
this.saveToLocalStorage();
|
||||
}
|
||||
|
||||
// Get all speakers
|
||||
getAllSpeakers(): Speaker[] {
|
||||
return Array.from(this.speakers.values());
|
||||
}
|
||||
|
||||
// Get speaker by ID
|
||||
getSpeaker(id: string): Speaker | undefined {
|
||||
return this.speakers.get(id);
|
||||
}
|
||||
|
||||
// Set active speaker
|
||||
setActiveSpeaker(id: string | null): void {
|
||||
// Deactivate all speakers
|
||||
this.speakers.forEach(speaker => {
|
||||
speaker.isActive = false;
|
||||
});
|
||||
|
||||
// Activate selected speaker
|
||||
if (id && this.speakers.has(id)) {
|
||||
const speaker = this.speakers.get(id)!;
|
||||
speaker.isActive = true;
|
||||
speaker.lastActiveTime = Date.now();
|
||||
this.activeSpeakerId = id;
|
||||
} else {
|
||||
this.activeSpeakerId = null;
|
||||
}
|
||||
|
||||
this.saveToLocalStorage();
|
||||
}
|
||||
|
||||
// Get active speaker
|
||||
getActiveSpeaker(): Speaker | null {
|
||||
return this.activeSpeakerId ? this.speakers.get(this.activeSpeakerId) || null : null;
|
||||
}
|
||||
|
||||
// Add conversation entry
|
||||
addConversationEntry(
|
||||
speakerId: string,
|
||||
originalText: string,
|
||||
originalLanguage: string
|
||||
): ConversationEntry {
|
||||
const entry: ConversationEntry = {
|
||||
id: this.generateEntryId(),
|
||||
speakerId,
|
||||
originalText,
|
||||
originalLanguage,
|
||||
translations: new Map(),
|
||||
timestamp: Date.now()
|
||||
};
|
||||
|
||||
this.conversation.push(entry);
|
||||
|
||||
// Limit conversation length
|
||||
if (this.conversation.length > this.maxConversationLength) {
|
||||
this.conversation.shift();
|
||||
}
|
||||
|
||||
this.saveToLocalStorage();
|
||||
return entry;
|
||||
}
|
||||
|
||||
// Add translation to conversation entry
|
||||
addTranslation(entryId: string, language: string, translatedText: string): void {
|
||||
const entry = this.conversation.find(e => e.id === entryId);
|
||||
if (entry) {
|
||||
entry.translations.set(language, translatedText);
|
||||
this.saveToLocalStorage();
|
||||
}
|
||||
}
|
||||
|
||||
// Get conversation for a specific language
|
||||
getConversationInLanguage(language: string): Array<{
|
||||
speakerId: string;
|
||||
speakerName: string;
|
||||
speakerColor: string;
|
||||
text: string;
|
||||
timestamp: number;
|
||||
isOriginal: boolean;
|
||||
}> {
|
||||
return this.conversation.map(entry => {
|
||||
const speaker = this.speakers.get(entry.speakerId);
|
||||
const isOriginal = entry.originalLanguage === language;
|
||||
const text = isOriginal ?
|
||||
entry.originalText :
|
||||
entry.translations.get(language) || `[Translating from ${entry.originalLanguage}...]`;
|
||||
|
||||
return {
|
||||
speakerId: entry.speakerId,
|
||||
speakerName: speaker?.name || 'Unknown',
|
||||
speakerColor: speaker?.color || '#666',
|
||||
text,
|
||||
timestamp: entry.timestamp,
|
||||
isOriginal
|
||||
};
|
||||
});
|
||||
}
|
||||
|
||||
// Get full conversation history
|
||||
getFullConversation(): ConversationEntry[] {
|
||||
return [...this.conversation];
|
||||
}
|
||||
|
||||
// Clear conversation
|
||||
clearConversation(): void {
|
||||
this.conversation = [];
|
||||
this.saveToLocalStorage();
|
||||
}
|
||||
|
||||
// Generate unique speaker ID
|
||||
private generateSpeakerId(): string {
|
||||
return `speaker_${Date.now()}_${Math.random().toString(36).substr(2, 9)}`;
|
||||
}
|
||||
|
||||
// Generate unique entry ID
|
||||
private generateEntryId(): string {
|
||||
return `entry_${Date.now()}_${Math.random().toString(36).substr(2, 9)}`;
|
||||
}
|
||||
|
||||
// Generate avatar initials
|
||||
private generateAvatar(name: string): string {
|
||||
const parts = name.trim().split(' ');
|
||||
if (parts.length >= 2) {
|
||||
return parts[0][0].toUpperCase() + parts[1][0].toUpperCase();
|
||||
}
|
||||
return name.substr(0, 2).toUpperCase();
|
||||
}
|
||||
|
||||
// Save to localStorage
|
||||
private saveToLocalStorage(): void {
|
||||
try {
|
||||
const data = {
|
||||
speakers: Array.from(this.speakers.entries()),
|
||||
conversation: this.conversation.map(entry => ({
|
||||
...entry,
|
||||
translations: Array.from(entry.translations.entries())
|
||||
})),
|
||||
activeSpeakerId: this.activeSpeakerId
|
||||
};
|
||||
localStorage.setItem('speakerData', JSON.stringify(data));
|
||||
} catch (error) {
|
||||
console.error('Failed to save speaker data:', error);
|
||||
}
|
||||
}
|
||||
|
||||
// Load from localStorage
|
||||
private loadFromLocalStorage(): void {
|
||||
try {
|
||||
const saved = localStorage.getItem('speakerData');
|
||||
if (saved) {
|
||||
const data = JSON.parse(saved);
|
||||
|
||||
// Restore speakers
|
||||
if (data.speakers) {
|
||||
this.speakers = new Map(data.speakers);
|
||||
}
|
||||
|
||||
// Restore conversation with Map translations
|
||||
if (data.conversation) {
|
||||
this.conversation = data.conversation.map((entry: any) => ({
|
||||
...entry,
|
||||
translations: new Map(entry.translations || [])
|
||||
}));
|
||||
}
|
||||
|
||||
// Restore active speaker
|
||||
this.activeSpeakerId = data.activeSpeakerId || null;
|
||||
}
|
||||
} catch (error) {
|
||||
console.error('Failed to load speaker data:', error);
|
||||
}
|
||||
}
|
||||
|
||||
// Export conversation as text
|
||||
exportConversation(language: string): string {
|
||||
const entries = this.getConversationInLanguage(language);
|
||||
return entries.map(entry =>
|
||||
`[${new Date(entry.timestamp).toLocaleTimeString()}] ${entry.speakerName}: ${entry.text}`
|
||||
).join('\n');
|
||||
}
|
||||
}
|
250
static/js/src/streamingTranslation.ts
Normal file
250
static/js/src/streamingTranslation.ts
Normal file
@ -0,0 +1,250 @@
|
||||
// Streaming translation implementation for reduced latency
|
||||
import { Validator } from './validator';
|
||||
import { PerformanceMonitor } from './performanceMonitor';
|
||||
|
||||
export interface StreamChunk {
|
||||
type: 'start' | 'chunk' | 'complete' | 'error';
|
||||
text?: string;
|
||||
full_text?: string;
|
||||
error?: string;
|
||||
source_lang?: string;
|
||||
target_lang?: string;
|
||||
}
|
||||
|
||||
export class StreamingTranslation {
|
||||
private eventSource: EventSource | null = null;
|
||||
private abortController: AbortController | null = null;
|
||||
private performanceMonitor = PerformanceMonitor.getInstance();
|
||||
private firstChunkReceived = false;
|
||||
|
||||
constructor(
|
||||
private onChunk: (text: string) => void,
|
||||
private onComplete: (fullText: string) => void,
|
||||
private onError: (error: string) => void,
|
||||
private onStart?: () => void
|
||||
) {}
|
||||
|
||||
async startStreaming(
|
||||
text: string,
|
||||
sourceLang: string,
|
||||
targetLang: string,
|
||||
useStreaming: boolean = true
|
||||
): Promise<void> {
|
||||
// Cancel any existing stream
|
||||
this.cancel();
|
||||
|
||||
// Validate inputs
|
||||
const sanitizedText = Validator.sanitizeText(text);
|
||||
if (!sanitizedText) {
|
||||
this.onError('No text to translate');
|
||||
return;
|
||||
}
|
||||
|
||||
if (!useStreaming) {
|
||||
// Fall back to regular translation
|
||||
await this.fallbackToRegularTranslation(sanitizedText, sourceLang, targetLang);
|
||||
return;
|
||||
}
|
||||
|
||||
try {
|
||||
// Check if browser supports EventSource
|
||||
if (!window.EventSource) {
|
||||
console.warn('EventSource not supported, falling back to regular translation');
|
||||
await this.fallbackToRegularTranslation(sanitizedText, sourceLang, targetLang);
|
||||
return;
|
||||
}
|
||||
|
||||
// Notify start
|
||||
if (this.onStart) {
|
||||
this.onStart();
|
||||
}
|
||||
|
||||
// Start performance timing
|
||||
this.performanceMonitor.startTimer('streaming_translation');
|
||||
this.firstChunkReceived = false;
|
||||
|
||||
// Create abort controller for cleanup
|
||||
this.abortController = new AbortController();
|
||||
|
||||
// Start streaming request
|
||||
const response = await fetch('/translate/stream', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json',
|
||||
},
|
||||
body: JSON.stringify({
|
||||
text: sanitizedText,
|
||||
source_lang: sourceLang,
|
||||
target_lang: targetLang
|
||||
}),
|
||||
signal: this.abortController.signal
|
||||
});
|
||||
|
||||
if (!response.ok) {
|
||||
throw new Error(`HTTP error! status: ${response.status}`);
|
||||
}
|
||||
|
||||
// Check if response is event-stream
|
||||
const contentType = response.headers.get('content-type');
|
||||
if (!contentType || !contentType.includes('text/event-stream')) {
|
||||
throw new Error('Server does not support streaming');
|
||||
}
|
||||
|
||||
// Process the stream
|
||||
await this.processStream(response);
|
||||
|
||||
} catch (error: any) {
|
||||
if (error.name === 'AbortError') {
|
||||
console.log('Stream cancelled');
|
||||
return;
|
||||
}
|
||||
|
||||
console.error('Streaming error:', error);
|
||||
|
||||
// Fall back to regular translation on error
|
||||
await this.fallbackToRegularTranslation(sanitizedText, sourceLang, targetLang);
|
||||
}
|
||||
}
|
||||
|
||||
private async processStream(response: Response): Promise<void> {
|
||||
const reader = response.body?.getReader();
|
||||
if (!reader) {
|
||||
throw new Error('No response body');
|
||||
}
|
||||
|
||||
const decoder = new TextDecoder();
|
||||
let buffer = '';
|
||||
|
||||
try {
|
||||
while (true) {
|
||||
const { done, value } = await reader.read();
|
||||
|
||||
if (done) {
|
||||
break;
|
||||
}
|
||||
|
||||
buffer += decoder.decode(value, { stream: true });
|
||||
|
||||
// Process complete SSE messages
|
||||
const lines = buffer.split('\n');
|
||||
buffer = lines.pop() || ''; // Keep incomplete line in buffer
|
||||
|
||||
for (const line of lines) {
|
||||
if (line.startsWith('data: ')) {
|
||||
try {
|
||||
const data = JSON.parse(line.slice(6)) as StreamChunk;
|
||||
this.handleStreamChunk(data);
|
||||
} catch (e) {
|
||||
console.error('Failed to parse SSE data:', e);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
} finally {
|
||||
reader.releaseLock();
|
||||
}
|
||||
}
|
||||
|
||||
private handleStreamChunk(chunk: StreamChunk): void {
|
||||
switch (chunk.type) {
|
||||
case 'start':
|
||||
console.log('Translation started:', chunk.source_lang, '->', chunk.target_lang);
|
||||
break;
|
||||
|
||||
case 'chunk':
|
||||
if (chunk.text) {
|
||||
// Record time to first byte
|
||||
if (!this.firstChunkReceived) {
|
||||
this.firstChunkReceived = true;
|
||||
this.performanceMonitor.measureTTFB('streaming_translation', performance.now());
|
||||
}
|
||||
this.onChunk(chunk.text);
|
||||
}
|
||||
break;
|
||||
|
||||
case 'complete':
|
||||
if (chunk.full_text) {
|
||||
// End performance timing
|
||||
this.performanceMonitor.endTimer('streaming_translation');
|
||||
this.onComplete(chunk.full_text);
|
||||
|
||||
// Log performance stats periodically
|
||||
if (Math.random() < 0.1) { // 10% of the time
|
||||
this.performanceMonitor.logPerformanceStats();
|
||||
}
|
||||
}
|
||||
break;
|
||||
|
||||
case 'error':
|
||||
this.onError(chunk.error || 'Unknown streaming error');
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
private async fallbackToRegularTranslation(
|
||||
text: string,
|
||||
sourceLang: string,
|
||||
targetLang: string
|
||||
): Promise<void> {
|
||||
try {
|
||||
const response = await fetch('/translate', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json',
|
||||
},
|
||||
body: JSON.stringify({
|
||||
text: text,
|
||||
source_lang: sourceLang,
|
||||
target_lang: targetLang
|
||||
})
|
||||
});
|
||||
|
||||
if (!response.ok) {
|
||||
throw new Error(`HTTP error! status: ${response.status}`);
|
||||
}
|
||||
|
||||
const data = await response.json();
|
||||
|
||||
if (data.success && data.translation) {
|
||||
// Simulate streaming by showing text progressively
|
||||
this.simulateStreaming(data.translation);
|
||||
} else {
|
||||
this.onError(data.error || 'Translation failed');
|
||||
}
|
||||
} catch (error: any) {
|
||||
this.onError(error.message || 'Translation failed');
|
||||
}
|
||||
}
|
||||
|
||||
private simulateStreaming(text: string): void {
|
||||
// Simulate streaming for better UX even with non-streaming response
|
||||
const words = text.split(' ');
|
||||
let index = 0;
|
||||
let accumulated = '';
|
||||
|
||||
const interval = setInterval(() => {
|
||||
if (index >= words.length) {
|
||||
clearInterval(interval);
|
||||
this.onComplete(accumulated.trim());
|
||||
return;
|
||||
}
|
||||
|
||||
const chunk = words[index] + (index < words.length - 1 ? ' ' : '');
|
||||
accumulated += chunk;
|
||||
this.onChunk(chunk);
|
||||
index++;
|
||||
}, 50); // 50ms between words for smooth appearance
|
||||
}
|
||||
|
||||
cancel(): void {
|
||||
if (this.abortController) {
|
||||
this.abortController.abort();
|
||||
this.abortController = null;
|
||||
}
|
||||
|
||||
if (this.eventSource) {
|
||||
this.eventSource.close();
|
||||
this.eventSource = null;
|
||||
}
|
||||
}
|
||||
}
|
243
static/js/src/translationCache.ts
Normal file
243
static/js/src/translationCache.ts
Normal file
@ -0,0 +1,243 @@
|
||||
// Translation cache management for offline support
|
||||
import { TranslationCacheEntry, CacheStats } from './types';
|
||||
import { Validator } from './validator';
|
||||
|
||||
export class TranslationCache {
|
||||
private static DB_NAME = 'VoiceTranslatorDB';
|
||||
private static DB_VERSION = 2; // Increment version for cache store
|
||||
private static CACHE_STORE = 'translationCache';
|
||||
// private static MAX_CACHE_SIZE = 50 * 1024 * 1024; // 50MB limit - Reserved for future use
|
||||
private static MAX_ENTRIES = 1000; // Maximum number of cached translations
|
||||
private static CACHE_EXPIRY_DAYS = 30; // Expire entries after 30 days
|
||||
|
||||
// Generate cache key from input parameters
|
||||
static generateCacheKey(text: string, sourceLang: string, targetLang: string): string {
|
||||
// Normalize and sanitize text to create a consistent key
|
||||
const normalizedText = text.trim().toLowerCase();
|
||||
const sanitized = Validator.sanitizeCacheKey(normalizedText);
|
||||
return `${sourceLang}:${targetLang}:${sanitized}`;
|
||||
}
|
||||
|
||||
// Open or create the cache database
|
||||
static async openDB(): Promise<IDBDatabase> {
|
||||
return new Promise((resolve, reject) => {
|
||||
const request = indexedDB.open(this.DB_NAME, this.DB_VERSION);
|
||||
|
||||
request.onupgradeneeded = (event: IDBVersionChangeEvent) => {
|
||||
const db = (event.target as IDBOpenDBRequest).result;
|
||||
|
||||
// Create cache store if it doesn't exist
|
||||
if (!db.objectStoreNames.contains(this.CACHE_STORE)) {
|
||||
const store = db.createObjectStore(this.CACHE_STORE, { keyPath: 'key' });
|
||||
store.createIndex('timestamp', 'timestamp', { unique: false });
|
||||
store.createIndex('lastAccessed', 'lastAccessed', { unique: false });
|
||||
store.createIndex('sourceLanguage', 'sourceLanguage', { unique: false });
|
||||
store.createIndex('targetLanguage', 'targetLanguage', { unique: false });
|
||||
}
|
||||
};
|
||||
|
||||
request.onsuccess = (event: Event) => {
|
||||
resolve((event.target as IDBOpenDBRequest).result);
|
||||
};
|
||||
|
||||
request.onerror = () => {
|
||||
reject('Failed to open translation cache database');
|
||||
};
|
||||
});
|
||||
}
|
||||
|
||||
// Get cached translation
|
||||
static async getCachedTranslation(
|
||||
text: string,
|
||||
sourceLang: string,
|
||||
targetLang: string
|
||||
): Promise<string | null> {
|
||||
try {
|
||||
const db = await this.openDB();
|
||||
const transaction = db.transaction([this.CACHE_STORE], 'readwrite');
|
||||
const store = transaction.objectStore(this.CACHE_STORE);
|
||||
|
||||
const key = this.generateCacheKey(text, sourceLang, targetLang);
|
||||
const request = store.get(key);
|
||||
|
||||
return new Promise((resolve) => {
|
||||
request.onsuccess = (event: Event) => {
|
||||
const entry = (event.target as IDBRequest).result as TranslationCacheEntry;
|
||||
|
||||
if (entry) {
|
||||
// Check if entry is not expired
|
||||
const expiryTime = entry.timestamp + (this.CACHE_EXPIRY_DAYS * 24 * 60 * 60 * 1000);
|
||||
if (Date.now() < expiryTime) {
|
||||
// Update access count and last accessed time
|
||||
entry.accessCount++;
|
||||
entry.lastAccessed = Date.now();
|
||||
store.put(entry);
|
||||
|
||||
console.log(`Cache hit for translation: ${sourceLang} -> ${targetLang}`);
|
||||
resolve(entry.targetText);
|
||||
} else {
|
||||
// Entry expired, delete it
|
||||
store.delete(key);
|
||||
resolve(null);
|
||||
}
|
||||
} else {
|
||||
resolve(null);
|
||||
}
|
||||
};
|
||||
|
||||
request.onerror = () => {
|
||||
console.error('Failed to get cached translation');
|
||||
resolve(null);
|
||||
};
|
||||
});
|
||||
} catch (error) {
|
||||
console.error('Cache lookup error:', error);
|
||||
return null;
|
||||
}
|
||||
}
|
||||
|
||||
// Save translation to cache
|
||||
static async cacheTranslation(
|
||||
sourceText: string,
|
||||
sourceLang: string,
|
||||
targetText: string,
|
||||
targetLang: string
|
||||
): Promise<void> {
|
||||
try {
|
||||
const db = await this.openDB();
|
||||
const transaction = db.transaction([this.CACHE_STORE], 'readwrite');
|
||||
const store = transaction.objectStore(this.CACHE_STORE);
|
||||
|
||||
const key = this.generateCacheKey(sourceText, sourceLang, targetLang);
|
||||
const entry: TranslationCacheEntry = {
|
||||
key,
|
||||
sourceText,
|
||||
sourceLanguage: sourceLang,
|
||||
targetText,
|
||||
targetLanguage: targetLang,
|
||||
timestamp: Date.now(),
|
||||
accessCount: 1,
|
||||
lastAccessed: Date.now()
|
||||
};
|
||||
|
||||
// Check cache size before adding
|
||||
await this.ensureCacheSize(db);
|
||||
|
||||
store.put(entry);
|
||||
console.log(`Cached translation: ${sourceLang} -> ${targetLang}`);
|
||||
} catch (error) {
|
||||
console.error('Failed to cache translation:', error);
|
||||
}
|
||||
}
|
||||
|
||||
// Ensure cache doesn't exceed size limits
|
||||
static async ensureCacheSize(db: IDBDatabase): Promise<void> {
|
||||
const transaction = db.transaction([this.CACHE_STORE], 'readwrite');
|
||||
const store = transaction.objectStore(this.CACHE_STORE);
|
||||
|
||||
// Count entries
|
||||
const countRequest = store.count();
|
||||
|
||||
countRequest.onsuccess = async () => {
|
||||
const count = countRequest.result;
|
||||
|
||||
if (count >= this.MAX_ENTRIES) {
|
||||
// Delete least recently accessed entries
|
||||
const index = store.index('lastAccessed');
|
||||
const cursor = index.openCursor();
|
||||
let deleted = 0;
|
||||
const toDelete = Math.floor(count * 0.2); // Delete 20% of entries
|
||||
|
||||
cursor.onsuccess = (event: Event) => {
|
||||
const cursor = (event.target as IDBRequest).result;
|
||||
if (cursor && deleted < toDelete) {
|
||||
cursor.delete();
|
||||
deleted++;
|
||||
cursor.continue();
|
||||
}
|
||||
};
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
// Get cache statistics
|
||||
static async getCacheStats(): Promise<CacheStats> {
|
||||
try {
|
||||
const db = await this.openDB();
|
||||
const transaction = db.transaction([this.CACHE_STORE], 'readonly');
|
||||
const store = transaction.objectStore(this.CACHE_STORE);
|
||||
|
||||
return new Promise((resolve) => {
|
||||
const stats: CacheStats = {
|
||||
totalEntries: 0,
|
||||
totalSize: 0,
|
||||
oldestEntry: Date.now(),
|
||||
newestEntry: 0
|
||||
};
|
||||
|
||||
const countRequest = store.count();
|
||||
countRequest.onsuccess = () => {
|
||||
stats.totalEntries = countRequest.result;
|
||||
};
|
||||
|
||||
const cursor = store.openCursor();
|
||||
cursor.onsuccess = (event: Event) => {
|
||||
const cursor = (event.target as IDBRequest).result;
|
||||
if (cursor) {
|
||||
const entry = cursor.value as TranslationCacheEntry;
|
||||
// Estimate size (rough calculation)
|
||||
stats.totalSize += (entry.sourceText.length + entry.targetText.length) * 2;
|
||||
stats.oldestEntry = Math.min(stats.oldestEntry, entry.timestamp);
|
||||
stats.newestEntry = Math.max(stats.newestEntry, entry.timestamp);
|
||||
cursor.continue();
|
||||
} else {
|
||||
resolve(stats);
|
||||
}
|
||||
};
|
||||
});
|
||||
} catch (error) {
|
||||
console.error('Failed to get cache stats:', error);
|
||||
return {
|
||||
totalEntries: 0,
|
||||
totalSize: 0,
|
||||
oldestEntry: 0,
|
||||
newestEntry: 0
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
// Clear all cache
|
||||
static async clearCache(): Promise<void> {
|
||||
try {
|
||||
const db = await this.openDB();
|
||||
const transaction = db.transaction([this.CACHE_STORE], 'readwrite');
|
||||
const store = transaction.objectStore(this.CACHE_STORE);
|
||||
store.clear();
|
||||
console.log('Translation cache cleared');
|
||||
} catch (error) {
|
||||
console.error('Failed to clear cache:', error);
|
||||
}
|
||||
}
|
||||
|
||||
// Export cache for backup
|
||||
static async exportCache(): Promise<TranslationCacheEntry[]> {
|
||||
try {
|
||||
const db = await this.openDB();
|
||||
const transaction = db.transaction([this.CACHE_STORE], 'readonly');
|
||||
const store = transaction.objectStore(this.CACHE_STORE);
|
||||
const request = store.getAll();
|
||||
|
||||
return new Promise((resolve) => {
|
||||
request.onsuccess = () => {
|
||||
resolve(request.result);
|
||||
};
|
||||
request.onerror = () => {
|
||||
resolve([]);
|
||||
};
|
||||
});
|
||||
} catch (error) {
|
||||
console.error('Failed to export cache:', error);
|
||||
return [];
|
||||
}
|
||||
}
|
||||
}
|
109
static/js/src/types.ts
Normal file
109
static/js/src/types.ts
Normal file
@ -0,0 +1,109 @@
|
||||
// Type definitions for Talk2Me application
|
||||
|
||||
export interface TranscriptionResponse {
|
||||
success: boolean;
|
||||
text?: string;
|
||||
error?: string;
|
||||
detected_language?: string;
|
||||
}
|
||||
|
||||
export interface TranslationResponse {
|
||||
success: boolean;
|
||||
translation?: string;
|
||||
error?: string;
|
||||
}
|
||||
|
||||
export interface TTSResponse {
|
||||
success: boolean;
|
||||
audio_url?: string;
|
||||
error?: string;
|
||||
}
|
||||
|
||||
export interface TTSServerStatus {
|
||||
status: 'online' | 'error' | 'auth_error';
|
||||
message: string;
|
||||
url: string;
|
||||
code?: number;
|
||||
}
|
||||
|
||||
export interface TTSConfigUpdate {
|
||||
server_url?: string;
|
||||
api_key?: string;
|
||||
}
|
||||
|
||||
export interface TTSConfigResponse {
|
||||
success: boolean;
|
||||
message?: string;
|
||||
url?: string;
|
||||
error?: string;
|
||||
}
|
||||
|
||||
export interface TranslationRequest {
|
||||
text: string;
|
||||
source_lang: string;
|
||||
target_lang: string;
|
||||
}
|
||||
|
||||
export interface TTSRequest {
|
||||
text: string;
|
||||
language: string;
|
||||
}
|
||||
|
||||
export interface PushPublicKeyResponse {
|
||||
publicKey: string;
|
||||
}
|
||||
|
||||
export interface IndexedDBRecord {
|
||||
timestamp: string;
|
||||
}
|
||||
|
||||
export interface TranscriptionRecord extends IndexedDBRecord {
|
||||
text: string;
|
||||
language: string;
|
||||
}
|
||||
|
||||
export interface TranslationRecord extends IndexedDBRecord {
|
||||
sourceText: string;
|
||||
sourceLanguage: string;
|
||||
targetText: string;
|
||||
targetLanguage: string;
|
||||
}
|
||||
|
||||
export interface TranslationCacheEntry {
|
||||
key: string;
|
||||
sourceText: string;
|
||||
sourceLanguage: string;
|
||||
targetText: string;
|
||||
targetLanguage: string;
|
||||
timestamp: number;
|
||||
accessCount: number;
|
||||
lastAccessed: number;
|
||||
}
|
||||
|
||||
export interface CacheStats {
|
||||
totalEntries: number;
|
||||
totalSize: number;
|
||||
oldestEntry: number;
|
||||
newestEntry: number;
|
||||
}
|
||||
|
||||
// Service Worker types
|
||||
export interface PeriodicSyncManager {
|
||||
register(tag: string, options?: { minInterval: number }): Promise<void>;
|
||||
}
|
||||
|
||||
export interface ServiceWorkerRegistrationExtended extends ServiceWorkerRegistration {
|
||||
periodicSync?: PeriodicSyncManager;
|
||||
}
|
||||
|
||||
// Extend window interface for PWA features
|
||||
declare global {
|
||||
interface Window {
|
||||
deferredPrompt?: BeforeInstallPromptEvent;
|
||||
}
|
||||
}
|
||||
|
||||
export interface BeforeInstallPromptEvent extends Event {
|
||||
prompt(): Promise<void>;
|
||||
userChoice: Promise<{ outcome: 'accepted' | 'dismissed' }>;
|
||||
}
|
259
static/js/src/validator.ts
Normal file
259
static/js/src/validator.ts
Normal file
@ -0,0 +1,259 @@
|
||||
// Input validation and sanitization utilities
|
||||
export class Validator {
|
||||
// Sanitize HTML to prevent XSS attacks
|
||||
static sanitizeHTML(input: string): string {
|
||||
// Create a temporary div element
|
||||
const temp = document.createElement('div');
|
||||
temp.textContent = input;
|
||||
return temp.innerHTML;
|
||||
}
|
||||
|
||||
// Validate and sanitize text input
|
||||
static sanitizeText(input: string, maxLength: number = 10000): string {
|
||||
if (typeof input !== 'string') {
|
||||
return '';
|
||||
}
|
||||
|
||||
// Trim and limit length
|
||||
let sanitized = input.trim().substring(0, maxLength);
|
||||
|
||||
// Remove null bytes
|
||||
sanitized = sanitized.replace(/\0/g, '');
|
||||
|
||||
// Remove control characters except newlines and tabs
|
||||
sanitized = sanitized.replace(/[\x00-\x08\x0B\x0C\x0E-\x1F\x7F]/g, '');
|
||||
|
||||
return sanitized;
|
||||
}
|
||||
|
||||
// Validate language code
|
||||
static validateLanguageCode(code: string, allowedLanguages: string[]): string | null {
|
||||
if (!code || typeof code !== 'string') {
|
||||
return null;
|
||||
}
|
||||
|
||||
const sanitized = code.trim().toLowerCase();
|
||||
|
||||
// Check if it's in the allowed list
|
||||
if (allowedLanguages.includes(sanitized) || sanitized === 'auto') {
|
||||
return sanitized;
|
||||
}
|
||||
|
||||
return null;
|
||||
}
|
||||
|
||||
// Validate file upload
|
||||
static validateAudioFile(file: File): { valid: boolean; error?: string } {
|
||||
// Check if file exists
|
||||
if (!file) {
|
||||
return { valid: false, error: 'No file provided' };
|
||||
}
|
||||
|
||||
// Check file size (max 25MB)
|
||||
const maxSize = 25 * 1024 * 1024;
|
||||
if (file.size > maxSize) {
|
||||
return { valid: false, error: 'File size exceeds 25MB limit' };
|
||||
}
|
||||
|
||||
// Check file type
|
||||
const allowedTypes = [
|
||||
'audio/webm',
|
||||
'audio/ogg',
|
||||
'audio/wav',
|
||||
'audio/mp3',
|
||||
'audio/mpeg',
|
||||
'audio/mp4',
|
||||
'audio/x-m4a',
|
||||
'audio/x-wav'
|
||||
];
|
||||
|
||||
if (!allowedTypes.includes(file.type)) {
|
||||
// Check by extension as fallback
|
||||
const ext = file.name.toLowerCase().split('.').pop();
|
||||
const allowedExtensions = ['webm', 'ogg', 'wav', 'mp3', 'mp4', 'm4a'];
|
||||
|
||||
if (!ext || !allowedExtensions.includes(ext)) {
|
||||
return { valid: false, error: 'Invalid audio file type' };
|
||||
}
|
||||
}
|
||||
|
||||
return { valid: true };
|
||||
}
|
||||
|
||||
// Validate URL
|
||||
static validateURL(url: string): string | null {
|
||||
if (!url || typeof url !== 'string') {
|
||||
return null;
|
||||
}
|
||||
|
||||
try {
|
||||
const parsed = new URL(url);
|
||||
|
||||
// Only allow http and https
|
||||
if (!['http:', 'https:'].includes(parsed.protocol)) {
|
||||
return null;
|
||||
}
|
||||
|
||||
// Prevent localhost in production
|
||||
if (window.location.hostname !== 'localhost' &&
|
||||
(parsed.hostname === 'localhost' || parsed.hostname === '127.0.0.1')) {
|
||||
return null;
|
||||
}
|
||||
|
||||
return parsed.toString();
|
||||
} catch (e) {
|
||||
return null;
|
||||
}
|
||||
}
|
||||
|
||||
// Validate API key (basic format check)
|
||||
static validateAPIKey(key: string): string | null {
|
||||
if (!key || typeof key !== 'string') {
|
||||
return null;
|
||||
}
|
||||
|
||||
// Trim whitespace
|
||||
const trimmed = key.trim();
|
||||
|
||||
// Check length (most API keys are 20-128 characters)
|
||||
if (trimmed.length < 20 || trimmed.length > 128) {
|
||||
return null;
|
||||
}
|
||||
|
||||
// Only allow alphanumeric, dash, and underscore
|
||||
if (!/^[a-zA-Z0-9\-_]+$/.test(trimmed)) {
|
||||
return null;
|
||||
}
|
||||
|
||||
return trimmed;
|
||||
}
|
||||
|
||||
// Validate request body size
|
||||
static validateRequestSize(data: any, maxSizeKB: number = 1024): boolean {
|
||||
try {
|
||||
const jsonString = JSON.stringify(data);
|
||||
const sizeInBytes = new Blob([jsonString]).size;
|
||||
return sizeInBytes <= maxSizeKB * 1024;
|
||||
} catch (e) {
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
// Sanitize filename
|
||||
static sanitizeFilename(filename: string): string {
|
||||
if (!filename || typeof filename !== 'string') {
|
||||
return 'file';
|
||||
}
|
||||
|
||||
// Remove path components
|
||||
let name = filename.split(/[/\\]/).pop() || 'file';
|
||||
|
||||
// Remove dangerous characters
|
||||
name = name.replace(/[^a-zA-Z0-9.\-_]/g, '_');
|
||||
|
||||
// Limit length
|
||||
if (name.length > 255) {
|
||||
const ext = name.split('.').pop();
|
||||
const base = name.substring(0, 250 - (ext ? ext.length + 1 : 0));
|
||||
name = ext ? `${base}.${ext}` : base;
|
||||
}
|
||||
|
||||
return name;
|
||||
}
|
||||
|
||||
// Validate settings object
|
||||
static validateSettings(settings: any): { valid: boolean; sanitized?: any; errors?: string[] } {
|
||||
const errors: string[] = [];
|
||||
const sanitized: any = {};
|
||||
|
||||
// Validate notification settings
|
||||
if (settings.notificationsEnabled !== undefined) {
|
||||
sanitized.notificationsEnabled = Boolean(settings.notificationsEnabled);
|
||||
}
|
||||
|
||||
if (settings.notifyTranscription !== undefined) {
|
||||
sanitized.notifyTranscription = Boolean(settings.notifyTranscription);
|
||||
}
|
||||
|
||||
if (settings.notifyTranslation !== undefined) {
|
||||
sanitized.notifyTranslation = Boolean(settings.notifyTranslation);
|
||||
}
|
||||
|
||||
if (settings.notifyErrors !== undefined) {
|
||||
sanitized.notifyErrors = Boolean(settings.notifyErrors);
|
||||
}
|
||||
|
||||
// Validate offline mode
|
||||
if (settings.offlineMode !== undefined) {
|
||||
sanitized.offlineMode = Boolean(settings.offlineMode);
|
||||
}
|
||||
|
||||
// Validate TTS settings
|
||||
if (settings.ttsServerUrl !== undefined) {
|
||||
const url = this.validateURL(settings.ttsServerUrl);
|
||||
if (settings.ttsServerUrl && !url) {
|
||||
errors.push('Invalid TTS server URL');
|
||||
} else {
|
||||
sanitized.ttsServerUrl = url;
|
||||
}
|
||||
}
|
||||
|
||||
if (settings.ttsApiKey !== undefined) {
|
||||
const key = this.validateAPIKey(settings.ttsApiKey);
|
||||
if (settings.ttsApiKey && !key) {
|
||||
errors.push('Invalid API key format');
|
||||
} else {
|
||||
sanitized.ttsApiKey = key;
|
||||
}
|
||||
}
|
||||
|
||||
return {
|
||||
valid: errors.length === 0,
|
||||
sanitized: errors.length === 0 ? sanitized : undefined,
|
||||
errors: errors.length > 0 ? errors : undefined
|
||||
};
|
||||
}
|
||||
|
||||
// Rate limiting check
|
||||
private static requestCounts: Map<string, number[]> = new Map();
|
||||
|
||||
static checkRateLimit(
|
||||
action: string,
|
||||
maxRequests: number = 10,
|
||||
windowMs: number = 60000
|
||||
): boolean {
|
||||
const now = Date.now();
|
||||
const key = action;
|
||||
|
||||
if (!this.requestCounts.has(key)) {
|
||||
this.requestCounts.set(key, []);
|
||||
}
|
||||
|
||||
const timestamps = this.requestCounts.get(key)!;
|
||||
|
||||
// Remove old timestamps
|
||||
const cutoff = now - windowMs;
|
||||
const recent = timestamps.filter(t => t > cutoff);
|
||||
|
||||
// Check if limit exceeded
|
||||
if (recent.length >= maxRequests) {
|
||||
return false;
|
||||
}
|
||||
|
||||
// Add current timestamp
|
||||
recent.push(now);
|
||||
this.requestCounts.set(key, recent);
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
// Validate translation cache key
|
||||
static sanitizeCacheKey(key: string): string {
|
||||
if (!key || typeof key !== 'string') {
|
||||
return '';
|
||||
}
|
||||
|
||||
// Remove special characters that might cause issues
|
||||
return key.replace(/[^\w\s-]/gi, '').substring(0, 500);
|
||||
}
|
||||
}
|
@ -1,30 +1,42 @@
|
||||
{
|
||||
"name": "Voice Language Translator",
|
||||
"short_name": "Translator",
|
||||
"description": "Translate spoken language between multiple languages with speech input and output",
|
||||
"name": "Talk2Me",
|
||||
"short_name": "Talk2Me",
|
||||
"description": "Real-time voice translation app - translate spoken language instantly",
|
||||
"start_url": "/",
|
||||
"scope": "/",
|
||||
"display": "standalone",
|
||||
"orientation": "portrait",
|
||||
"background_color": "#ffffff",
|
||||
"theme_color": "#007bff",
|
||||
"icons": [
|
||||
{
|
||||
"src": "./static/icons/icon-192x192.png",
|
||||
"src": "/static/icons/icon-192x192.png",
|
||||
"sizes": "192x192",
|
||||
"type": "image/png",
|
||||
"purpose": "any maskable"
|
||||
},
|
||||
{
|
||||
"src": "./static/icons/icon-512x512.png",
|
||||
"src": "/static/icons/icon-512x512.png",
|
||||
"sizes": "512x512",
|
||||
"type": "image/png",
|
||||
"purpose": "any maskable"
|
||||
}
|
||||
],
|
||||
"screenshots": [
|
||||
"shortcuts": [
|
||||
{
|
||||
"src": "./static/screenshots/screenshot1.png",
|
||||
"sizes": "1280x720",
|
||||
"type": "image/png"
|
||||
"name": "Start Recording",
|
||||
"short_name": "Record",
|
||||
"description": "Start voice recording for translation",
|
||||
"url": "/?action=record",
|
||||
"icons": [
|
||||
{
|
||||
"src": "/static/icons/icon-192x192.png",
|
||||
"sizes": "192x192"
|
||||
}
|
||||
]
|
||||
}
|
||||
]
|
||||
],
|
||||
"categories": ["productivity", "utilities", "education"],
|
||||
"prefer_related_applications": false,
|
||||
"related_applications": []
|
||||
}
|
||||
|
41
static/pwa-update.js
Normal file
41
static/pwa-update.js
Normal file
@ -0,0 +1,41 @@
|
||||
// PWA Update Helper
|
||||
// This script helps force PWA updates on clients
|
||||
|
||||
// Force service worker update
|
||||
if ('serviceWorker' in navigator) {
|
||||
navigator.serviceWorker.getRegistrations().then(function(registrations) {
|
||||
for(let registration of registrations) {
|
||||
registration.unregister().then(function() {
|
||||
console.log('Service worker unregistered');
|
||||
});
|
||||
}
|
||||
});
|
||||
|
||||
// Re-register after a short delay
|
||||
setTimeout(function() {
|
||||
navigator.serviceWorker.register('/service-worker.js')
|
||||
.then(function(registration) {
|
||||
console.log('Service worker re-registered');
|
||||
registration.update();
|
||||
});
|
||||
}, 1000);
|
||||
}
|
||||
|
||||
// Clear all caches
|
||||
if ('caches' in window) {
|
||||
caches.keys().then(function(names) {
|
||||
for (let name of names) {
|
||||
caches.delete(name);
|
||||
console.log('Cache cleared:', name);
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
// Reload manifest
|
||||
var link = document.querySelector('link[rel="manifest"]');
|
||||
if (link) {
|
||||
link.href = link.href + '?v=' + Date.now();
|
||||
console.log('Manifest reloaded');
|
||||
}
|
||||
|
||||
console.log('PWA update complete. Please reload the page.');
|
@ -1,13 +1,16 @@
|
||||
// Service Worker for Voice Language Translator PWA
|
||||
// Service Worker for Talk2Me PWA
|
||||
|
||||
const CACHE_NAME = 'voice-translator-v1';
|
||||
const CACHE_NAME = 'talk2me-v4'; // Increment version to force update
|
||||
const ASSETS_TO_CACHE = [
|
||||
'/',
|
||||
'/static/manifest.json',
|
||||
'/static/css/styles.css',
|
||||
'/static/js/app.js',
|
||||
'/static/js/dist/app.bundle.js',
|
||||
'/static/icons/icon-192x192.png',
|
||||
'/static/icons/icon-512x512.png',
|
||||
'/static/icons/favicon.ico'
|
||||
'/static/icons/favicon.ico',
|
||||
'https://cdn.jsdelivr.net/npm/bootstrap@5.3.0-alpha1/dist/css/bootstrap.min.css',
|
||||
'https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.0.0/css/all.min.css'
|
||||
];
|
||||
|
||||
// Install event - cache essential assets
|
||||
@ -90,15 +93,34 @@ self.addEventListener('fetch', (event) => {
|
||||
|
||||
// Handle push notifications
|
||||
self.addEventListener('push', (event) => {
|
||||
if (!event.data) {
|
||||
return;
|
||||
}
|
||||
|
||||
const data = event.data.json();
|
||||
const options = {
|
||||
body: data.body || 'New translation available',
|
||||
icon: '/static/icons/icon-192x192.png',
|
||||
badge: '/static/icons/badge-72x72.png',
|
||||
icon: data.icon || '/static/icons/icon-192x192.png',
|
||||
badge: data.badge || '/static/icons/icon-192x192.png',
|
||||
vibrate: [100, 50, 100],
|
||||
tag: data.tag || 'talk2me-notification',
|
||||
requireInteraction: false,
|
||||
silent: false,
|
||||
data: {
|
||||
url: data.url || '/'
|
||||
}
|
||||
url: data.url || '/',
|
||||
...data.data
|
||||
},
|
||||
actions: [
|
||||
{
|
||||
action: 'view',
|
||||
title: 'View',
|
||||
icon: '/static/icons/icon-192x192.png'
|
||||
},
|
||||
{
|
||||
action: 'close',
|
||||
title: 'Close'
|
||||
}
|
||||
]
|
||||
};
|
||||
|
||||
event.waitUntil(
|
||||
@ -109,7 +131,55 @@ self.addEventListener('push', (event) => {
|
||||
// Handle notification click
|
||||
self.addEventListener('notificationclick', (event) => {
|
||||
event.notification.close();
|
||||
|
||||
if (event.action === 'close') {
|
||||
return;
|
||||
}
|
||||
|
||||
const urlToOpen = event.notification.data.url || '/';
|
||||
|
||||
event.waitUntil(
|
||||
clients.openWindow(event.notification.data.url)
|
||||
clients.matchAll({
|
||||
type: 'window',
|
||||
includeUncontrolled: true
|
||||
}).then((windowClients) => {
|
||||
// Check if there's already a window/tab with the app open
|
||||
for (let client of windowClients) {
|
||||
if (client.url === urlToOpen && 'focus' in client) {
|
||||
return client.focus();
|
||||
}
|
||||
}
|
||||
// If not, open a new window/tab
|
||||
if (clients.openWindow) {
|
||||
return clients.openWindow(urlToOpen);
|
||||
}
|
||||
})
|
||||
);
|
||||
});
|
||||
|
||||
// Handle periodic background sync
|
||||
self.addEventListener('periodicsync', (event) => {
|
||||
if (event.tag === 'translation-updates') {
|
||||
event.waitUntil(checkForUpdates());
|
||||
}
|
||||
});
|
||||
|
||||
async function checkForUpdates() {
|
||||
// Check for app updates or send usage statistics
|
||||
try {
|
||||
const response = await fetch('/api/check-updates');
|
||||
if (response.ok) {
|
||||
const data = await response.json();
|
||||
if (data.hasUpdate) {
|
||||
self.registration.showNotification('Update Available', {
|
||||
body: 'A new version of Voice Translator is available!',
|
||||
icon: '/static/icons/icon-192x192.png',
|
||||
badge: '/static/icons/icon-192x192.png',
|
||||
tag: 'update-notification'
|
||||
});
|
||||
}
|
||||
}
|
||||
} catch (error) {
|
||||
console.error('Failed to check for updates:', error);
|
||||
}
|
||||
}
|
||||
|
66
talk2me.service
Normal file
66
talk2me.service
Normal file
@ -0,0 +1,66 @@
|
||||
[Unit]
|
||||
Description=Talk2Me Real-time Translation Service
|
||||
Documentation=https://github.com/your-repo/talk2me
|
||||
After=network.target
|
||||
|
||||
[Service]
|
||||
Type=notify
|
||||
User=talk2me
|
||||
Group=talk2me
|
||||
WorkingDirectory=/opt/talk2me
|
||||
Environment="PATH=/opt/talk2me/venv/bin"
|
||||
Environment="FLASK_ENV=production"
|
||||
Environment="PYTHONUNBUFFERED=1"
|
||||
|
||||
# Production environment variables
|
||||
EnvironmentFile=-/opt/talk2me/.env
|
||||
|
||||
# Gunicorn command with production settings
|
||||
ExecStart=/opt/talk2me/venv/bin/gunicorn \
|
||||
--config /opt/talk2me/gunicorn_config.py \
|
||||
--error-logfile /var/log/talk2me/gunicorn-error.log \
|
||||
--access-logfile /var/log/talk2me/gunicorn-access.log \
|
||||
--log-level info \
|
||||
wsgi:application
|
||||
|
||||
# Reload via SIGHUP
|
||||
ExecReload=/bin/kill -s HUP $MAINPID
|
||||
|
||||
# Graceful stop
|
||||
KillMode=mixed
|
||||
TimeoutStopSec=30
|
||||
|
||||
# Restart policy
|
||||
Restart=always
|
||||
RestartSec=10
|
||||
StartLimitBurst=3
|
||||
StartLimitInterval=60
|
||||
|
||||
# Security settings
|
||||
NoNewPrivileges=true
|
||||
PrivateTmp=true
|
||||
ProtectSystem=strict
|
||||
ProtectHome=true
|
||||
ProtectKernelTunables=true
|
||||
ProtectKernelModules=true
|
||||
ProtectControlGroups=true
|
||||
RestrictRealtime=true
|
||||
RestrictSUIDSGID=true
|
||||
LockPersonality=true
|
||||
|
||||
# Allow writing to specific directories
|
||||
ReadWritePaths=/var/log/talk2me /tmp/talk2me_uploads
|
||||
|
||||
# Resource limits
|
||||
LimitNOFILE=65536
|
||||
LimitNPROC=4096
|
||||
|
||||
# Memory limits (adjust based on your system)
|
||||
MemoryLimit=4G
|
||||
MemoryHigh=3G
|
||||
|
||||
# CPU limits (optional)
|
||||
# CPUQuota=200%
|
||||
|
||||
[Install]
|
||||
WantedBy=multi-user.target
|
712
templates/admin_users.html
Normal file
712
templates/admin_users.html
Normal file
@ -0,0 +1,712 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en">
|
||||
<head>
|
||||
<meta charset="UTF-8">
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||||
<title>User Management - Talk2Me Admin</title>
|
||||
<link href="https://cdn.jsdelivr.net/npm/bootstrap@5.1.3/dist/css/bootstrap.min.css" rel="stylesheet">
|
||||
<link rel="stylesheet" href="https://cdn.jsdelivr.net/npm/bootstrap-icons@1.7.2/font/bootstrap-icons.css">
|
||||
<style>
|
||||
.user-avatar {
|
||||
width: 40px;
|
||||
height: 40px;
|
||||
border-radius: 50%;
|
||||
object-fit: cover;
|
||||
}
|
||||
.status-badge {
|
||||
font-size: 0.875rem;
|
||||
}
|
||||
.action-buttons .btn {
|
||||
padding: 0.25rem 0.5rem;
|
||||
font-size: 0.875rem;
|
||||
}
|
||||
.stats-card {
|
||||
border: none;
|
||||
border-radius: 10px;
|
||||
box-shadow: 0 2px 4px rgba(0,0,0,0.1);
|
||||
}
|
||||
.stats-card .card-body {
|
||||
padding: 1.5rem;
|
||||
}
|
||||
.stats-number {
|
||||
font-size: 2rem;
|
||||
font-weight: bold;
|
||||
color: #4a5568;
|
||||
}
|
||||
.table-responsive {
|
||||
border-radius: 8px;
|
||||
box-shadow: 0 1px 3px rgba(0,0,0,0.1);
|
||||
}
|
||||
.modal-header {
|
||||
background-color: #f8f9fa;
|
||||
}
|
||||
.search-filters {
|
||||
background-color: #f8f9fa;
|
||||
padding: 1rem;
|
||||
border-radius: 8px;
|
||||
margin-bottom: 1rem;
|
||||
}
|
||||
</style>
|
||||
</head>
|
||||
<body>
|
||||
<nav class="navbar navbar-expand-lg navbar-dark bg-dark">
|
||||
<div class="container-fluid">
|
||||
<a class="navbar-brand" href="/admin">Talk2Me Admin</a>
|
||||
<div class="navbar-nav ms-auto">
|
||||
<a class="nav-link" href="/" target="_blank">
|
||||
<i class="bi bi-box-arrow-up-right"></i> Main App
|
||||
</a>
|
||||
<a class="nav-link" href="#" onclick="logout()">
|
||||
<i class="bi bi-box-arrow-right"></i> Logout
|
||||
</a>
|
||||
</div>
|
||||
</div>
|
||||
</nav>
|
||||
|
||||
<div class="container-fluid mt-4">
|
||||
<div class="row">
|
||||
<div class="col-12">
|
||||
<h1 class="mb-4">User Management</h1>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Statistics Cards -->
|
||||
<div class="row mb-4">
|
||||
<div class="col-md-3">
|
||||
<div class="card stats-card">
|
||||
<div class="card-body">
|
||||
<h6 class="text-muted mb-2">Total Users</h6>
|
||||
<div class="stats-number" id="stat-total">0</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
<div class="col-md-3">
|
||||
<div class="card stats-card">
|
||||
<div class="card-body">
|
||||
<h6 class="text-muted mb-2">Active Users</h6>
|
||||
<div class="stats-number text-success" id="stat-active">0</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
<div class="col-md-3">
|
||||
<div class="card stats-card">
|
||||
<div class="card-body">
|
||||
<h6 class="text-muted mb-2">Suspended Users</h6>
|
||||
<div class="stats-number text-warning" id="stat-suspended">0</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
<div class="col-md-3">
|
||||
<div class="card stats-card">
|
||||
<div class="card-body">
|
||||
<h6 class="text-muted mb-2">Admin Users</h6>
|
||||
<div class="stats-number text-primary" id="stat-admins">0</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Search and Filters -->
|
||||
<div class="search-filters">
|
||||
<div class="row">
|
||||
<div class="col-md-4">
|
||||
<input type="text" class="form-control" id="searchInput" placeholder="Search by email, username, or name...">
|
||||
</div>
|
||||
<div class="col-md-2">
|
||||
<select class="form-select" id="roleFilter">
|
||||
<option value="">All Roles</option>
|
||||
<option value="admin">Admin</option>
|
||||
<option value="user">User</option>
|
||||
</select>
|
||||
</div>
|
||||
<div class="col-md-2">
|
||||
<select class="form-select" id="statusFilter">
|
||||
<option value="">All Status</option>
|
||||
<option value="active">Active</option>
|
||||
<option value="suspended">Suspended</option>
|
||||
<option value="inactive">Inactive</option>
|
||||
</select>
|
||||
</div>
|
||||
<div class="col-md-2">
|
||||
<select class="form-select" id="sortBy">
|
||||
<option value="created_at">Created Date</option>
|
||||
<option value="last_login_at">Last Login</option>
|
||||
<option value="total_requests">Total Requests</option>
|
||||
<option value="username">Username</option>
|
||||
</select>
|
||||
</div>
|
||||
<div class="col-md-2">
|
||||
<button class="btn btn-primary w-100" onclick="createUser()">
|
||||
<i class="bi bi-plus-circle"></i> Create User
|
||||
</button>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Users Table -->
|
||||
<div class="table-responsive">
|
||||
<table class="table table-hover">
|
||||
<thead>
|
||||
<tr>
|
||||
<th>User</th>
|
||||
<th>Role</th>
|
||||
<th>Status</th>
|
||||
<th>Usage</th>
|
||||
<th>Last Login</th>
|
||||
<th>Created</th>
|
||||
<th>Actions</th>
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody id="usersTableBody">
|
||||
<!-- Users will be loaded here -->
|
||||
</tbody>
|
||||
</table>
|
||||
</div>
|
||||
|
||||
<!-- Pagination -->
|
||||
<nav aria-label="Page navigation">
|
||||
<ul class="pagination justify-content-center" id="pagination">
|
||||
<!-- Pagination will be loaded here -->
|
||||
</ul>
|
||||
</nav>
|
||||
</div>
|
||||
|
||||
<!-- User Details Modal -->
|
||||
<div class="modal fade" id="userModal" tabindex="-1">
|
||||
<div class="modal-dialog modal-lg">
|
||||
<div class="modal-content">
|
||||
<div class="modal-header">
|
||||
<h5 class="modal-title">User Details</h5>
|
||||
<button type="button" class="btn-close" data-bs-dismiss="modal"></button>
|
||||
</div>
|
||||
<div class="modal-body" id="userModalBody">
|
||||
<!-- User details will be loaded here -->
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Create/Edit User Modal -->
|
||||
<div class="modal fade" id="createUserModal" tabindex="-1">
|
||||
<div class="modal-dialog">
|
||||
<div class="modal-content">
|
||||
<div class="modal-header">
|
||||
<h5 class="modal-title" id="createUserModalTitle">Create User</h5>
|
||||
<button type="button" class="btn-close" data-bs-dismiss="modal"></button>
|
||||
</div>
|
||||
<div class="modal-body">
|
||||
<form id="userForm">
|
||||
<div class="mb-3">
|
||||
<label for="userEmail" class="form-label">Email</label>
|
||||
<input type="email" class="form-control" id="userEmail" required>
|
||||
</div>
|
||||
<div class="mb-3">
|
||||
<label for="userUsername" class="form-label">Username</label>
|
||||
<input type="text" class="form-control" id="userUsername" required>
|
||||
</div>
|
||||
<div class="mb-3">
|
||||
<label for="userPassword" class="form-label">Password</label>
|
||||
<input type="password" class="form-control" id="userPassword" minlength="8">
|
||||
<small class="text-muted">Leave blank to keep existing password (edit mode)</small>
|
||||
</div>
|
||||
<div class="mb-3">
|
||||
<label for="userFullName" class="form-label">Full Name</label>
|
||||
<input type="text" class="form-control" id="userFullName">
|
||||
</div>
|
||||
<div class="mb-3">
|
||||
<label for="userRole" class="form-label">Role</label>
|
||||
<select class="form-select" id="userRole">
|
||||
<option value="user">User</option>
|
||||
<option value="admin">Admin</option>
|
||||
</select>
|
||||
</div>
|
||||
<div class="mb-3">
|
||||
<div class="form-check">
|
||||
<input class="form-check-input" type="checkbox" id="userVerified">
|
||||
<label class="form-check-label" for="userVerified">
|
||||
Email Verified
|
||||
</label>
|
||||
</div>
|
||||
</div>
|
||||
<div class="mb-3">
|
||||
<label for="userRateLimit" class="form-label">Rate Limit (per minute)</label>
|
||||
<input type="number" class="form-control" id="userRateLimit" value="30">
|
||||
</div>
|
||||
</form>
|
||||
</div>
|
||||
<div class="modal-footer">
|
||||
<button type="button" class="btn btn-secondary" data-bs-dismiss="modal">Cancel</button>
|
||||
<button type="button" class="btn btn-primary" onclick="saveUser()">Save User</button>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<script src="https://cdn.jsdelivr.net/npm/bootstrap@5.1.3/dist/js/bootstrap.bundle.min.js"></script>
|
||||
<script src="https://cdn.jsdelivr.net/npm/axios/dist/axios.min.js"></script>
|
||||
<script>
|
||||
let currentPage = 1;
|
||||
let editingUserId = null;
|
||||
|
||||
// Configure axios defaults for session-based auth
|
||||
axios.defaults.withCredentials = true;
|
||||
|
||||
// Also check if there's a JWT token (for API users)
|
||||
let authToken = localStorage.getItem('auth_token');
|
||||
if (authToken) {
|
||||
axios.defaults.headers.common['Authorization'] = `Bearer ${authToken}`;
|
||||
}
|
||||
|
||||
// For admin token auth, add the admin token header
|
||||
const adminToken = '{{ session.get("admin_token", "") }}';
|
||||
if (adminToken) {
|
||||
axios.defaults.headers.common['X-Admin-Token'] = adminToken;
|
||||
}
|
||||
|
||||
// Load users on page load
|
||||
document.addEventListener('DOMContentLoaded', function() {
|
||||
loadStats();
|
||||
loadUsers();
|
||||
|
||||
// Add event listeners
|
||||
document.getElementById('searchInput').addEventListener('input', debounce(loadUsers, 300));
|
||||
document.getElementById('roleFilter').addEventListener('change', loadUsers);
|
||||
document.getElementById('statusFilter').addEventListener('change', loadUsers);
|
||||
document.getElementById('sortBy').addEventListener('change', loadUsers);
|
||||
});
|
||||
|
||||
function debounce(func, wait) {
|
||||
let timeout;
|
||||
return function executedFunction(...args) {
|
||||
const later = () => {
|
||||
clearTimeout(timeout);
|
||||
func(...args);
|
||||
};
|
||||
clearTimeout(timeout);
|
||||
timeout = setTimeout(later, wait);
|
||||
};
|
||||
}
|
||||
|
||||
async function loadStats() {
|
||||
try {
|
||||
const response = await axios.get('/api/auth/admin/stats/users');
|
||||
const stats = response.data.stats;
|
||||
|
||||
document.getElementById('stat-total').textContent = stats.total_users;
|
||||
document.getElementById('stat-active').textContent = stats.active_users;
|
||||
document.getElementById('stat-suspended').textContent = stats.suspended_users;
|
||||
document.getElementById('stat-admins').textContent = stats.admin_users;
|
||||
} catch (error) {
|
||||
console.error('Failed to load stats:', error);
|
||||
}
|
||||
}
|
||||
|
||||
async function loadUsers(page = 1) {
|
||||
try {
|
||||
currentPage = page;
|
||||
const params = {
|
||||
page: page,
|
||||
per_page: 20,
|
||||
search: document.getElementById('searchInput').value,
|
||||
role: document.getElementById('roleFilter').value,
|
||||
status: document.getElementById('statusFilter').value,
|
||||
sort_by: document.getElementById('sortBy').value,
|
||||
sort_order: 'desc'
|
||||
};
|
||||
|
||||
console.log('Loading users with params:', params);
|
||||
const response = await axios.get('/api/auth/admin/users', { params });
|
||||
const data = response.data;
|
||||
console.log('Received data:', data);
|
||||
|
||||
// Also try the debug endpoint to see all users
|
||||
try {
|
||||
const debugResponse = await axios.get('/api/debug-users');
|
||||
console.log('Debug endpoint shows:', debugResponse.data);
|
||||
} catch (debugError) {
|
||||
console.error('Debug endpoint error:', debugError);
|
||||
}
|
||||
|
||||
displayUsers(data.users);
|
||||
displayPagination(data.pagination);
|
||||
} catch (error) {
|
||||
console.error('Failed to load users:', error);
|
||||
console.error('Response:', error.response);
|
||||
showAlert('Failed to load users: ' + (error.response?.data?.error || error.message), 'danger');
|
||||
}
|
||||
}
|
||||
|
||||
function displayUsers(users) {
|
||||
const tbody = document.getElementById('usersTableBody');
|
||||
tbody.innerHTML = '';
|
||||
|
||||
console.log('displayUsers called with:', users);
|
||||
console.log('Number of users to display:', users ? users.length : 0);
|
||||
|
||||
if (!users || users.length === 0) {
|
||||
tbody.innerHTML = '<tr><td colspan="7" class="text-center">No users found</td></tr>';
|
||||
return;
|
||||
}
|
||||
|
||||
users.forEach(user => {
|
||||
const tr = document.createElement('tr');
|
||||
tr.innerHTML = `
|
||||
<td>
|
||||
<div class="d-flex align-items-center">
|
||||
${user.avatar_url ?
|
||||
`<img src="${user.avatar_url}" class="user-avatar me-2" alt="${user.username}">` :
|
||||
`<div class="user-avatar me-2 bg-secondary d-flex align-items-center justify-content-center text-white">
|
||||
${user.username.charAt(0).toUpperCase()}
|
||||
</div>`
|
||||
}
|
||||
<div>
|
||||
<div class="fw-bold">${user.username}</div>
|
||||
<small class="text-muted">${user.email}</small>
|
||||
</div>
|
||||
</div>
|
||||
</td>
|
||||
<td>
|
||||
<span class="badge ${user.role === 'admin' ? 'bg-primary' : 'bg-secondary'}">
|
||||
${user.role}
|
||||
</span>
|
||||
</td>
|
||||
<td>
|
||||
${getStatusBadge(user)}
|
||||
</td>
|
||||
<td>
|
||||
<small>
|
||||
<i class="bi bi-translate"></i> ${user.total_translations}<br>
|
||||
<i class="bi bi-mic"></i> ${user.total_transcriptions}<br>
|
||||
<i class="bi bi-volume-up"></i> ${user.total_tts_requests}
|
||||
</small>
|
||||
</td>
|
||||
<td>
|
||||
<small>${user.last_login_at ? new Date(user.last_login_at).toLocaleString() : 'Never'}</small>
|
||||
</td>
|
||||
<td>
|
||||
<small>${new Date(user.created_at).toLocaleDateString()}</small>
|
||||
</td>
|
||||
<td class="action-buttons">
|
||||
<button class="btn btn-sm btn-info" onclick="viewUser('${user.id}')" title="View Details">
|
||||
<i class="bi bi-eye"></i>
|
||||
</button>
|
||||
<button class="btn btn-sm btn-warning" onclick="editUser('${user.id}')" title="Edit">
|
||||
<i class="bi bi-pencil"></i>
|
||||
</button>
|
||||
${user.is_suspended ?
|
||||
`<button class="btn btn-sm btn-success" onclick="unsuspendUser('${user.id}')" title="Unsuspend">
|
||||
<i class="bi bi-play-circle"></i>
|
||||
</button>` :
|
||||
`<button class="btn btn-sm btn-warning" onclick="suspendUser('${user.id}')" title="Suspend">
|
||||
<i class="bi bi-pause-circle"></i>
|
||||
</button>`
|
||||
}
|
||||
${user.role !== 'admin' ?
|
||||
`<button class="btn btn-sm btn-danger" onclick="deleteUser('${user.id}')" title="Delete">
|
||||
<i class="bi bi-trash"></i>
|
||||
</button>` : ''
|
||||
}
|
||||
</td>
|
||||
`;
|
||||
tbody.appendChild(tr);
|
||||
});
|
||||
}
|
||||
|
||||
function getStatusBadge(user) {
|
||||
if (user.is_suspended) {
|
||||
return '<span class="badge bg-warning status-badge">Suspended</span>';
|
||||
} else if (!user.is_active) {
|
||||
return '<span class="badge bg-secondary status-badge">Inactive</span>';
|
||||
} else if (!user.is_verified) {
|
||||
return '<span class="badge bg-info status-badge">Unverified</span>';
|
||||
} else {
|
||||
return '<span class="badge bg-success status-badge">Active</span>';
|
||||
}
|
||||
}
|
||||
|
||||
function displayPagination(pagination) {
|
||||
const nav = document.getElementById('pagination');
|
||||
nav.innerHTML = '';
|
||||
|
||||
const totalPages = pagination.pages;
|
||||
const currentPage = pagination.page;
|
||||
|
||||
// Previous button
|
||||
const prevLi = document.createElement('li');
|
||||
prevLi.className = `page-item ${currentPage === 1 ? 'disabled' : ''}`;
|
||||
prevLi.innerHTML = `<a class="page-link" href="#" onclick="loadUsers(${currentPage - 1})">Previous</a>`;
|
||||
nav.appendChild(prevLi);
|
||||
|
||||
// Page numbers
|
||||
for (let i = 1; i <= totalPages; i++) {
|
||||
if (i === 1 || i === totalPages || (i >= currentPage - 2 && i <= currentPage + 2)) {
|
||||
const li = document.createElement('li');
|
||||
li.className = `page-item ${i === currentPage ? 'active' : ''}`;
|
||||
li.innerHTML = `<a class="page-link" href="#" onclick="loadUsers(${i})">${i}</a>`;
|
||||
nav.appendChild(li);
|
||||
} else if (i === currentPage - 3 || i === currentPage + 3) {
|
||||
const li = document.createElement('li');
|
||||
li.className = 'page-item disabled';
|
||||
li.innerHTML = '<span class="page-link">...</span>';
|
||||
nav.appendChild(li);
|
||||
}
|
||||
}
|
||||
|
||||
// Next button
|
||||
const nextLi = document.createElement('li');
|
||||
nextLi.className = `page-item ${currentPage === totalPages ? 'disabled' : ''}`;
|
||||
nextLi.innerHTML = `<a class="page-link" href="#" onclick="loadUsers(${currentPage + 1})">Next</a>`;
|
||||
nav.appendChild(nextLi);
|
||||
}
|
||||
|
||||
async function viewUser(userId) {
|
||||
try {
|
||||
const response = await axios.get(`/api/auth/admin/users/${userId}`);
|
||||
const data = response.data;
|
||||
|
||||
const modalBody = document.getElementById('userModalBody');
|
||||
modalBody.innerHTML = `
|
||||
<div class="row">
|
||||
<div class="col-md-6">
|
||||
<h6>User Information</h6>
|
||||
<dl class="row">
|
||||
<dt class="col-sm-4">Username:</dt>
|
||||
<dd class="col-sm-8">${data.user.username}</dd>
|
||||
<dt class="col-sm-4">Email:</dt>
|
||||
<dd class="col-sm-8">${data.user.email}</dd>
|
||||
<dt class="col-sm-4">Full Name:</dt>
|
||||
<dd class="col-sm-8">${data.user.full_name || 'N/A'}</dd>
|
||||
<dt class="col-sm-4">Role:</dt>
|
||||
<dd class="col-sm-8">${data.user.role}</dd>
|
||||
<dt class="col-sm-4">Status:</dt>
|
||||
<dd class="col-sm-8">${getStatusBadge(data.user)}</dd>
|
||||
<dt class="col-sm-4">API Key:</dt>
|
||||
<dd class="col-sm-8">
|
||||
<code>${data.user.api_key}</code>
|
||||
<button class="btn btn-sm btn-secondary ms-2" onclick="copyToClipboard('${data.user.api_key}')">
|
||||
<i class="bi bi-clipboard"></i>
|
||||
</button>
|
||||
</dd>
|
||||
</dl>
|
||||
</div>
|
||||
<div class="col-md-6">
|
||||
<h6>Usage Statistics</h6>
|
||||
<dl class="row">
|
||||
<dt class="col-sm-6">Total Requests:</dt>
|
||||
<dd class="col-sm-6">${data.user.total_requests}</dd>
|
||||
<dt class="col-sm-6">Translations:</dt>
|
||||
<dd class="col-sm-6">${data.user.total_translations}</dd>
|
||||
<dt class="col-sm-6">Transcriptions:</dt>
|
||||
<dd class="col-sm-6">${data.user.total_transcriptions}</dd>
|
||||
<dt class="col-sm-6">TTS Requests:</dt>
|
||||
<dd class="col-sm-6">${data.user.total_tts_requests}</dd>
|
||||
<dt class="col-sm-6">Rate Limits:</dt>
|
||||
<dd class="col-sm-6">
|
||||
${data.user.rate_limit_per_minute}/min<br>
|
||||
${data.user.rate_limit_per_hour}/hour<br>
|
||||
${data.user.rate_limit_per_day}/day
|
||||
</dd>
|
||||
</dl>
|
||||
</div>
|
||||
</div>
|
||||
<hr>
|
||||
<h6>Login History</h6>
|
||||
<div class="table-responsive">
|
||||
<table class="table table-sm">
|
||||
<thead>
|
||||
<tr>
|
||||
<th>Date</th>
|
||||
<th>IP Address</th>
|
||||
<th>Method</th>
|
||||
<th>Status</th>
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody>
|
||||
${data.login_history.map(login => `
|
||||
<tr>
|
||||
<td>${new Date(login.login_at).toLocaleString()}</td>
|
||||
<td>${login.ip_address}</td>
|
||||
<td>${login.login_method}</td>
|
||||
<td>${login.success ?
|
||||
'<span class="badge bg-success">Success</span>' :
|
||||
`<span class="badge bg-danger">Failed</span>`
|
||||
}</td>
|
||||
</tr>
|
||||
`).join('')}
|
||||
</tbody>
|
||||
</table>
|
||||
</div>
|
||||
<hr>
|
||||
<h6>Active Sessions (${data.active_sessions.length})</h6>
|
||||
<div class="table-responsive">
|
||||
<table class="table table-sm">
|
||||
<thead>
|
||||
<tr>
|
||||
<th>Session ID</th>
|
||||
<th>Created</th>
|
||||
<th>Last Active</th>
|
||||
<th>IP Address</th>
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody>
|
||||
${data.active_sessions.map(session => `
|
||||
<tr>
|
||||
<td><code>${session.session_id.substr(0, 8)}...</code></td>
|
||||
<td>${new Date(session.created_at).toLocaleString()}</td>
|
||||
<td>${new Date(session.last_active_at).toLocaleString()}</td>
|
||||
<td>${session.ip_address}</td>
|
||||
</tr>
|
||||
`).join('')}
|
||||
</tbody>
|
||||
</table>
|
||||
</div>
|
||||
`;
|
||||
|
||||
const modal = new bootstrap.Modal(document.getElementById('userModal'));
|
||||
modal.show();
|
||||
} catch (error) {
|
||||
console.error('Failed to load user details:', error);
|
||||
showAlert('Failed to load user details', 'danger');
|
||||
}
|
||||
}
|
||||
|
||||
function createUser() {
|
||||
editingUserId = null;
|
||||
document.getElementById('createUserModalTitle').textContent = 'Create User';
|
||||
document.getElementById('userForm').reset();
|
||||
document.getElementById('userPassword').required = true;
|
||||
const modal = new bootstrap.Modal(document.getElementById('createUserModal'));
|
||||
modal.show();
|
||||
}
|
||||
|
||||
async function editUser(userId) {
|
||||
try {
|
||||
const response = await axios.get(`/api/auth/admin/users/${userId}`);
|
||||
const user = response.data.user;
|
||||
|
||||
editingUserId = userId;
|
||||
document.getElementById('createUserModalTitle').textContent = 'Edit User';
|
||||
document.getElementById('userEmail').value = user.email;
|
||||
document.getElementById('userUsername').value = user.username;
|
||||
document.getElementById('userPassword').value = '';
|
||||
document.getElementById('userPassword').required = false;
|
||||
document.getElementById('userFullName').value = user.full_name || '';
|
||||
document.getElementById('userRole').value = user.role;
|
||||
document.getElementById('userVerified').checked = user.is_verified;
|
||||
document.getElementById('userRateLimit').value = user.rate_limit_per_minute;
|
||||
|
||||
const modal = new bootstrap.Modal(document.getElementById('createUserModal'));
|
||||
modal.show();
|
||||
} catch (error) {
|
||||
console.error('Failed to load user for editing:', error);
|
||||
showAlert('Failed to load user', 'danger');
|
||||
}
|
||||
}
|
||||
|
||||
async function saveUser() {
|
||||
try {
|
||||
const data = {
|
||||
email: document.getElementById('userEmail').value,
|
||||
username: document.getElementById('userUsername').value,
|
||||
full_name: document.getElementById('userFullName').value,
|
||||
role: document.getElementById('userRole').value,
|
||||
is_verified: document.getElementById('userVerified').checked,
|
||||
rate_limit_per_minute: parseInt(document.getElementById('userRateLimit').value)
|
||||
};
|
||||
|
||||
if (editingUserId) {
|
||||
// Update existing user
|
||||
if (document.getElementById('userPassword').value) {
|
||||
data.password = document.getElementById('userPassword').value;
|
||||
}
|
||||
await axios.put(`/api/auth/admin/users/${editingUserId}`, data);
|
||||
showAlert('User updated successfully', 'success');
|
||||
} else {
|
||||
// Create new user
|
||||
data.password = document.getElementById('userPassword').value;
|
||||
await axios.post('/api/auth/admin/users', data);
|
||||
showAlert('User created successfully', 'success');
|
||||
}
|
||||
|
||||
bootstrap.Modal.getInstance(document.getElementById('createUserModal')).hide();
|
||||
loadUsers(currentPage);
|
||||
loadStats();
|
||||
} catch (error) {
|
||||
console.error('Failed to save user:', error);
|
||||
showAlert(error.response?.data?.error || 'Failed to save user', 'danger');
|
||||
}
|
||||
}
|
||||
|
||||
async function suspendUser(userId) {
|
||||
if (!confirm('Are you sure you want to suspend this user?')) return;
|
||||
|
||||
try {
|
||||
const reason = prompt('Enter suspension reason:');
|
||||
if (!reason) return;
|
||||
|
||||
await axios.post(`/api/auth/admin/users/${userId}/suspend`, { reason });
|
||||
showAlert('User suspended successfully', 'success');
|
||||
loadUsers(currentPage);
|
||||
loadStats();
|
||||
} catch (error) {
|
||||
console.error('Failed to suspend user:', error);
|
||||
showAlert('Failed to suspend user', 'danger');
|
||||
}
|
||||
}
|
||||
|
||||
async function unsuspendUser(userId) {
|
||||
if (!confirm('Are you sure you want to unsuspend this user?')) return;
|
||||
|
||||
try {
|
||||
await axios.post(`/api/auth/admin/users/${userId}/unsuspend`);
|
||||
showAlert('User unsuspended successfully', 'success');
|
||||
loadUsers(currentPage);
|
||||
loadStats();
|
||||
} catch (error) {
|
||||
console.error('Failed to unsuspend user:', error);
|
||||
showAlert('Failed to unsuspend user', 'danger');
|
||||
}
|
||||
}
|
||||
|
||||
async function deleteUser(userId) {
|
||||
if (!confirm('Are you sure you want to delete this user? This action cannot be undone.')) return;
|
||||
|
||||
try {
|
||||
await axios.delete(`/api/auth/admin/users/${userId}`);
|
||||
showAlert('User deleted successfully', 'success');
|
||||
loadUsers(currentPage);
|
||||
loadStats();
|
||||
} catch (error) {
|
||||
console.error('Failed to delete user:', error);
|
||||
showAlert('Failed to delete user', 'danger');
|
||||
}
|
||||
}
|
||||
|
||||
function copyToClipboard(text) {
|
||||
navigator.clipboard.writeText(text).then(() => {
|
||||
showAlert('Copied to clipboard', 'success');
|
||||
});
|
||||
}
|
||||
|
||||
function showAlert(message, type) {
|
||||
const alertDiv = document.createElement('div');
|
||||
alertDiv.className = `alert alert-${type} alert-dismissible fade show position-fixed top-0 start-50 translate-middle-x mt-3`;
|
||||
alertDiv.style.zIndex = '9999';
|
||||
alertDiv.innerHTML = `
|
||||
${message}
|
||||
<button type="button" class="btn-close" data-bs-dismiss="alert"></button>
|
||||
`;
|
||||
document.body.appendChild(alertDiv);
|
||||
|
||||
setTimeout(() => {
|
||||
alertDiv.remove();
|
||||
}, 5000);
|
||||
}
|
||||
|
||||
function logout() {
|
||||
localStorage.removeItem('auth_token');
|
||||
window.location.href = '/login';
|
||||
}
|
||||
</script>
|
||||
</body>
|
||||
</html>
|
@ -2,21 +2,107 @@
|
||||
<html lang="en">
|
||||
<head>
|
||||
<meta charset="UTF-8">
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=no">
|
||||
<title>Voice Language Translator</title>
|
||||
<link href="https://cdn.jsdelivr.net/npm/bootstrap@5.3.0-alpha1/dist/css/bootstrap.min.css" rel="stylesheet">
|
||||
<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.0.0/css/all.min.css">
|
||||
<link rel="icon" href="/favicon.ico" sizes="any">
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=no, viewport-fit=cover">
|
||||
<title>Talk2Me - Real-time Voice Translation</title>
|
||||
|
||||
<!-- Icons for various platforms -->
|
||||
<link rel="icon" href="/static/icons/favicon.ico" sizes="any">
|
||||
<link rel="apple-touch-icon" href="/static/icons/apple-icon-180x180.png">
|
||||
<link rel="apple-touch-icon" sizes="120x120" href="/static/icons/apple-icon-120x120.png">
|
||||
<link rel="apple-touch-icon" sizes="152x152" href="/static/icons/apple-icon-152x152.png">
|
||||
<link rel="apple-touch-icon" sizes="180x180" href="/static/icons/apple-icon-180x180.png">
|
||||
<link rel="apple-touch-icon" sizes="167x167" href="/static/icons/apple-icon-167x167.png">
|
||||
<link rel="apple-touch-icon" sizes="180x180" href="/static/icons/apple-icon-180x180.png">
|
||||
<style>
|
||||
body {
|
||||
padding-top: 20px;
|
||||
padding-bottom: 20px;
|
||||
padding-top: 10px;
|
||||
padding-bottom: 10px;
|
||||
background-color: #f8f9fa;
|
||||
}
|
||||
|
||||
/* Mobile-first approach */
|
||||
@media (max-width: 768px) {
|
||||
.container {
|
||||
padding-left: 10px;
|
||||
padding-right: 10px;
|
||||
}
|
||||
|
||||
h1 {
|
||||
font-size: 1.5rem;
|
||||
margin-bottom: 0.5rem !important;
|
||||
}
|
||||
|
||||
.card {
|
||||
margin-bottom: 10px !important;
|
||||
}
|
||||
|
||||
.card-body {
|
||||
padding: 0.75rem !important;
|
||||
}
|
||||
|
||||
.card-header {
|
||||
padding: 0.5rem 0.75rem !important;
|
||||
}
|
||||
|
||||
.card-header h5 {
|
||||
font-size: 1rem;
|
||||
margin-bottom: 0;
|
||||
}
|
||||
|
||||
.text-display {
|
||||
min-height: 60px !important;
|
||||
max-height: 100px;
|
||||
overflow-y: auto;
|
||||
padding: 10px !important;
|
||||
margin-bottom: 10px !important;
|
||||
font-size: 0.9rem;
|
||||
}
|
||||
|
||||
.language-select {
|
||||
padding: 5px 10px !important;
|
||||
font-size: 0.9rem;
|
||||
margin-bottom: 10px !important;
|
||||
}
|
||||
|
||||
.btn-action {
|
||||
padding: 5px 10px !important;
|
||||
font-size: 0.875rem;
|
||||
margin: 2px !important;
|
||||
}
|
||||
|
||||
.record-btn {
|
||||
width: 60px !important;
|
||||
height: 60px !important;
|
||||
font-size: 24px !important;
|
||||
margin: 10px auto !important;
|
||||
}
|
||||
|
||||
.status-indicator {
|
||||
font-size: 0.8rem !important;
|
||||
margin-top: 5px !important;
|
||||
}
|
||||
|
||||
/* Hide speaker toolbar on mobile by default */
|
||||
#speakerToolbar {
|
||||
position: fixed;
|
||||
bottom: 70px;
|
||||
left: 0;
|
||||
right: 0;
|
||||
z-index: 100;
|
||||
border-radius: 0 !important;
|
||||
}
|
||||
|
||||
#conversationView {
|
||||
position: fixed;
|
||||
bottom: 0;
|
||||
left: 0;
|
||||
right: 0;
|
||||
top: 50%;
|
||||
z-index: 99;
|
||||
border-radius: 15px 15px 0 0 !important;
|
||||
margin: 0 !important;
|
||||
}
|
||||
}
|
||||
|
||||
.record-btn {
|
||||
width: 80px;
|
||||
height: 80px;
|
||||
@ -74,6 +160,7 @@
|
||||
background-color: #f8f9fa;
|
||||
border-radius: 10px;
|
||||
margin-bottom: 15px;
|
||||
position: relative;
|
||||
}
|
||||
.btn-action {
|
||||
border-radius: 10px;
|
||||
@ -90,19 +177,39 @@
|
||||
font-style: italic;
|
||||
color: #6c757d;
|
||||
}
|
||||
|
||||
/* Ensure record button area is always visible on mobile */
|
||||
@media (max-width: 768px) {
|
||||
.record-section {
|
||||
position: sticky;
|
||||
bottom: 0;
|
||||
background: white;
|
||||
padding: 10px 0;
|
||||
box-shadow: 0 -2px 10px rgba(0,0,0,0.1);
|
||||
z-index: 50;
|
||||
margin-left: -10px;
|
||||
margin-right: -10px;
|
||||
padding-left: 10px;
|
||||
padding-right: 10px;
|
||||
}
|
||||
}
|
||||
</style>
|
||||
|
||||
<!-- PWA Meta Tags -->
|
||||
<meta name="description" content="Translate spoken language between multiple languages with speech input and output">
|
||||
<meta name="description" content="Real-time voice translation app - translate spoken language instantly">
|
||||
<meta name="theme-color" content="#007bff">
|
||||
<meta name="mobile-web-app-capable" content="yes">
|
||||
<meta name="apple-mobile-web-app-capable" content="yes">
|
||||
<meta name="apple-mobile-web-app-status-bar-style" content="black-translucent">
|
||||
<meta name="apple-mobile-web-app-title" content="Translator">
|
||||
<meta name="apple-mobile-web-app-title" content="Talk2Me">
|
||||
<meta name="application-name" content="Talk2Me">
|
||||
<meta name="msapplication-TileColor" content="#007bff">
|
||||
<meta name="msapplication-TileImage" content="/static/icons/icon-192x192.png">
|
||||
|
||||
<!-- PWA Icons and Manifest -->
|
||||
<link rel="manifest" href="/static/manifest.json">
|
||||
<link rel="icon" type="image/png" href="/static/icons/icon-192x192.png">
|
||||
<link rel="apple-touch-icon" href="/static/icons/apple-icon-180x180.png">
|
||||
<link rel="icon" type="image/png" sizes="192x192" href="/static/icons/icon-192x192.png">
|
||||
<link rel="icon" type="image/png" sizes="512x512" href="/static/icons/icon-512x512.png">
|
||||
|
||||
<!-- Apple Splash Screens -->
|
||||
<link rel="apple-touch-startup-image" href="/static/splash/apple-splash-2048-2732.png" media="(device-width: 1024px) and (device-height: 1366px) and (-webkit-device-pixel-ratio: 2) and (orientation: portrait)">
|
||||
@ -121,8 +228,27 @@
|
||||
</head>
|
||||
<body>
|
||||
<div class="container">
|
||||
<h1 class="text-center mb-4">Voice Language Translator</h1>
|
||||
<h1 class="text-center mb-4">Talk2Me</h1>
|
||||
<!--<p class="text-center text-muted">Powered by Gemma 3, Whisper & Edge TTS</p>-->
|
||||
|
||||
<!-- Multi-speaker toolbar -->
|
||||
<div id="speakerToolbar" class="card mb-3" style="display: none;">
|
||||
<div class="card-body p-2">
|
||||
<div class="d-flex align-items-center justify-content-between flex-wrap">
|
||||
<div class="d-flex align-items-center gap-2 mb-2 mb-md-0">
|
||||
<button id="addSpeakerBtn" class="btn btn-sm btn-outline-primary">
|
||||
<i class="fas fa-user-plus"></i> Add Speaker
|
||||
</button>
|
||||
<button id="toggleMultiSpeaker" class="btn btn-sm btn-secondary">
|
||||
<i class="fas fa-users"></i> Multi-Speaker: <span id="multiSpeakerStatus">OFF</span>
|
||||
</button>
|
||||
</div>
|
||||
<div id="speakerList" class="d-flex gap-2 flex-wrap">
|
||||
<!-- Speaker buttons will be added here dynamically -->
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="row">
|
||||
<div class="col-md-6 mb-3">
|
||||
@ -132,6 +258,7 @@
|
||||
</div>
|
||||
<div class="card-body">
|
||||
<select id="sourceLanguage" class="form-select language-select mb-3">
|
||||
<option value="auto">Auto-detect</option>
|
||||
{% for language in languages %}
|
||||
<option value="{{ language }}">{{ language }}</option>
|
||||
{% endfor %}
|
||||
@ -178,17 +305,25 @@
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="text-center">
|
||||
<div class="text-center record-section">
|
||||
<button id="recordBtn" class="btn btn-primary record-btn">
|
||||
<i class="fas fa-microphone"></i>
|
||||
</button>
|
||||
<p class="status-indicator" id="statusIndicator">Click to start recording</p>
|
||||
</div>
|
||||
|
||||
<div class="text-center mt-3">
|
||||
<button id="translateBtn" class="btn btn-success" disabled>
|
||||
<i class="fas fa-language"></i> Translate
|
||||
</button>
|
||||
<!-- Queue Status Indicator -->
|
||||
<div id="queueStatus" class="text-center mt-2" style="display: none;">
|
||||
<small class="text-muted">
|
||||
<i class="fas fa-list"></i> Queue: <span id="queueLength">0</span> |
|
||||
<i class="fas fa-sync"></i> Active: <span id="activeRequests">0</span>
|
||||
</small>
|
||||
</div>
|
||||
|
||||
<div class="mt-2">
|
||||
<button id="translateBtn" class="btn btn-outline-secondary btn-sm" disabled title="Translation happens automatically after transcription">
|
||||
<i class="fas fa-redo"></i> Re-translate
|
||||
</button>
|
||||
<small class="text-muted d-block mt-1">Translation happens automatically after transcription</small>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="mt-3">
|
||||
@ -196,285 +331,183 @@
|
||||
<div id="progressBar" class="progress-bar progress-bar-striped progress-bar-animated" role="progressbar" style="width: 0%"></div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Multi-speaker conversation view -->
|
||||
<div id="conversationView" class="card mt-4" style="display: none;">
|
||||
<div class="card-header bg-info text-white d-flex justify-content-between align-items-center">
|
||||
<h5 class="mb-0">Conversation</h5>
|
||||
<div>
|
||||
<button id="exportConversation" class="btn btn-sm btn-light">
|
||||
<i class="fas fa-download"></i> Export
|
||||
</button>
|
||||
<button id="clearConversation" class="btn btn-sm btn-light">
|
||||
<i class="fas fa-trash"></i> Clear
|
||||
</button>
|
||||
</div>
|
||||
</div>
|
||||
<div class="card-body" style="max-height: 400px; overflow-y: auto;">
|
||||
<div id="conversationContent">
|
||||
<!-- Conversation entries will be added here -->
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<audio id="audioPlayer" style="display: none;"></audio>
|
||||
|
||||
<!-- TTS Server Configuration Alert - Moved to Admin Dashboard -->
|
||||
<!-- Commenting out user-facing TTS server configuration
|
||||
<div id="ttsServerAlert" class="alert alert-warning d-none" role="alert">
|
||||
<strong>TTS Server Status:</strong> <span id="ttsServerMessage">Checking...</span>
|
||||
<div class="mt-2">
|
||||
<input type="text" id="ttsServerUrl" class="form-control mb-2" placeholder="TTS Server URL">
|
||||
<input type="password" id="ttsApiKey" class="form-control mb-2" placeholder="API Key">
|
||||
<button id="updateTtsServer" class="btn btn-sm btn-primary">Update Configuration</button>
|
||||
</div>
|
||||
</div>
|
||||
-->
|
||||
|
||||
<!-- Loading Overlay -->
|
||||
<div id="loadingOverlay" class="loading-overlay">
|
||||
<div class="loading-content">
|
||||
<div class="spinner-custom"></div>
|
||||
<p id="loadingText" class="mt-3">Processing...</p>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Notification Settings -->
|
||||
<div class="position-fixed bottom-0 end-0 p-3" style="z-index: 5">
|
||||
<div id="notificationPrompt" class="toast" role="alert" aria-live="assertive" aria-atomic="true">
|
||||
<div class="toast-header">
|
||||
<i class="fas fa-bell text-primary me-2"></i>
|
||||
<strong class="me-auto">Enable Notifications</strong>
|
||||
<button type="button" class="btn-close" data-bs-dismiss="toast" aria-label="Close"></button>
|
||||
</div>
|
||||
<div class="toast-body">
|
||||
Get notified when translations are complete!
|
||||
<div class="mt-2">
|
||||
<button type="button" class="btn btn-sm btn-primary" id="enableNotifications">Enable</button>
|
||||
<button type="button" class="btn btn-sm btn-secondary" data-bs-dismiss="toast">Not now</button>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Success Toast -->
|
||||
<div id="successToast" class="toast align-items-center text-white bg-success border-0" role="alert" aria-live="assertive" aria-atomic="true">
|
||||
<div class="d-flex">
|
||||
<div class="toast-body">
|
||||
<i class="fas fa-check-circle me-2"></i>
|
||||
<span id="successMessage">Settings saved successfully!</span>
|
||||
</div>
|
||||
<button type="button" class="btn-close btn-close-white me-2 m-auto" data-bs-dismiss="toast" aria-label="Close"></button>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Settings Modal -->
|
||||
<div class="modal fade" id="settingsModal" tabindex="-1" aria-labelledby="settingsModalLabel" aria-hidden="true">
|
||||
<div class="modal-dialog">
|
||||
<div class="modal-content">
|
||||
<div class="modal-header">
|
||||
<h5 class="modal-title" id="settingsModalLabel">Settings</h5>
|
||||
<button type="button" class="btn-close" data-bs-dismiss="modal" aria-label="Close"></button>
|
||||
</div>
|
||||
<div class="modal-body">
|
||||
<h6>Notifications</h6>
|
||||
<div class="form-check form-switch">
|
||||
<input class="form-check-input" type="checkbox" id="notificationToggle">
|
||||
<label class="form-check-label" for="notificationToggle">
|
||||
Enable push notifications
|
||||
</label>
|
||||
</div>
|
||||
<p class="text-muted small mt-2">Get notified when transcriptions and translations complete</p>
|
||||
|
||||
<hr>
|
||||
|
||||
<h6>Notification Types</h6>
|
||||
<div class="form-check">
|
||||
<input class="form-check-input" type="checkbox" id="notifyTranscription" checked>
|
||||
<label class="form-check-label" for="notifyTranscription">
|
||||
Transcription complete
|
||||
</label>
|
||||
</div>
|
||||
<div class="form-check">
|
||||
<input class="form-check-input" type="checkbox" id="notifyTranslation" checked>
|
||||
<label class="form-check-label" for="notifyTranslation">
|
||||
Translation complete
|
||||
</label>
|
||||
</div>
|
||||
<div class="form-check">
|
||||
<input class="form-check-input" type="checkbox" id="notifyErrors">
|
||||
<label class="form-check-label" for="notifyErrors">
|
||||
Error notifications
|
||||
</label>
|
||||
</div>
|
||||
|
||||
<hr>
|
||||
|
||||
<h6 class="mb-3">Translation Settings</h6>
|
||||
<div class="form-check form-switch mb-3">
|
||||
<input class="form-check-input" type="checkbox" id="streamingTranslation" checked>
|
||||
<label class="form-check-label" for="streamingTranslation">
|
||||
Enable streaming translation
|
||||
<small class="text-muted d-block">Shows translation as it's generated for faster feedback</small>
|
||||
</label>
|
||||
</div>
|
||||
<div class="form-check form-switch mb-3">
|
||||
<input class="form-check-input" type="checkbox" id="multiSpeakerMode">
|
||||
<label class="form-check-label" for="multiSpeakerMode">
|
||||
Enable multi-speaker mode
|
||||
<small class="text-muted d-block">Track multiple speakers in conversations</small>
|
||||
</label>
|
||||
</div>
|
||||
|
||||
<hr>
|
||||
|
||||
<h6>Offline Cache</h6>
|
||||
<div class="mb-3">
|
||||
<div class="d-flex justify-content-between align-items-center mb-2">
|
||||
<span>Cached translations:</span>
|
||||
<span id="cacheCount" class="badge bg-primary">0</span>
|
||||
</div>
|
||||
<div class="d-flex justify-content-between align-items-center mb-2">
|
||||
<span>Cache size:</span>
|
||||
<span id="cacheSize" class="badge bg-secondary">0 KB</span>
|
||||
</div>
|
||||
<div class="form-check form-switch mb-2">
|
||||
<input class="form-check-input" type="checkbox" id="offlineMode" checked>
|
||||
<label class="form-check-label" for="offlineMode">
|
||||
Enable offline caching
|
||||
</label>
|
||||
</div>
|
||||
<button type="button" class="btn btn-sm btn-outline-danger" id="clearCache">
|
||||
<i class="fas fa-trash"></i> Clear Cache
|
||||
</button>
|
||||
</div>
|
||||
</div>
|
||||
<div class="modal-footer">
|
||||
<div id="settingsSaveStatus" class="text-success me-auto" style="display: none;">
|
||||
<i class="fas fa-check-circle"></i> Saved!
|
||||
</div>
|
||||
<button type="button" class="btn btn-secondary" data-bs-dismiss="modal">Close</button>
|
||||
<button type="button" class="btn btn-primary" id="saveSettings">Save settings</button>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<!-- Settings Button -->
|
||||
<button type="button" class="btn btn-outline-secondary position-fixed top-0 end-0 m-3" data-bs-toggle="modal" data-bs-target="#settingsModal">
|
||||
<i class="fas fa-cog"></i>
|
||||
</button>
|
||||
|
||||
<!-- Simple Success Notification -->
|
||||
<div id="successNotification" class="success-notification">
|
||||
<i class="fas fa-check-circle"></i>
|
||||
<span id="successText">Settings saved successfully!</span>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<script src="https://cdn.jsdelivr.net/npm/bootstrap@5.3.0-alpha1/dist/js/bootstrap.bundle.min.js"></script>
|
||||
<script>
|
||||
document.addEventListener('DOMContentLoaded', function() {
|
||||
// DOM elements
|
||||
const recordBtn = document.getElementById('recordBtn');
|
||||
const translateBtn = document.getElementById('translateBtn');
|
||||
const sourceText = document.getElementById('sourceText');
|
||||
const translatedText = document.getElementById('translatedText');
|
||||
const sourceLanguage = document.getElementById('sourceLanguage');
|
||||
const targetLanguage = document.getElementById('targetLanguage');
|
||||
const playSource = document.getElementById('playSource');
|
||||
const playTranslation = document.getElementById('playTranslation');
|
||||
const clearSource = document.getElementById('clearSource');
|
||||
const clearTranslation = document.getElementById('clearTranslation');
|
||||
const statusIndicator = document.getElementById('statusIndicator');
|
||||
const progressContainer = document.getElementById('progressContainer');
|
||||
const progressBar = document.getElementById('progressBar');
|
||||
const audioPlayer = document.getElementById('audioPlayer');
|
||||
|
||||
// Set initial values
|
||||
let isRecording = false;
|
||||
let mediaRecorder = null;
|
||||
let audioChunks = [];
|
||||
let currentSourceText = '';
|
||||
let currentTranslationText = '';
|
||||
|
||||
// Make sure target language is different from source
|
||||
if (targetLanguage.options[0].value === sourceLanguage.value) {
|
||||
targetLanguage.selectedIndex = 1;
|
||||
}
|
||||
|
||||
// Event listeners for language selection
|
||||
sourceLanguage.addEventListener('change', function() {
|
||||
if (targetLanguage.value === sourceLanguage.value) {
|
||||
for (let i = 0; i < targetLanguage.options.length; i++) {
|
||||
if (targetLanguage.options[i].value !== sourceLanguage.value) {
|
||||
targetLanguage.selectedIndex = i;
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
});
|
||||
|
||||
targetLanguage.addEventListener('change', function() {
|
||||
if (targetLanguage.value === sourceLanguage.value) {
|
||||
for (let i = 0; i < sourceLanguage.options.length; i++) {
|
||||
if (sourceLanguage.options[i].value !== targetLanguage.value) {
|
||||
sourceLanguage.selectedIndex = i;
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
});
|
||||
|
||||
// Record button click event
|
||||
recordBtn.addEventListener('click', function() {
|
||||
if (isRecording) {
|
||||
stopRecording();
|
||||
} else {
|
||||
startRecording();
|
||||
}
|
||||
});
|
||||
|
||||
// Function to start recording
|
||||
function startRecording() {
|
||||
navigator.mediaDevices.getUserMedia({ audio: true })
|
||||
.then(stream => {
|
||||
mediaRecorder = new MediaRecorder(stream);
|
||||
audioChunks = [];
|
||||
|
||||
mediaRecorder.addEventListener('dataavailable', event => {
|
||||
audioChunks.push(event.data);
|
||||
});
|
||||
|
||||
mediaRecorder.addEventListener('stop', () => {
|
||||
const audioBlob = new Blob(audioChunks, { type: 'audio/wav' });
|
||||
transcribeAudio(audioBlob);
|
||||
});
|
||||
|
||||
mediaRecorder.start();
|
||||
isRecording = true;
|
||||
recordBtn.classList.add('recording');
|
||||
recordBtn.classList.replace('btn-primary', 'btn-danger');
|
||||
recordBtn.innerHTML = '<i class="fas fa-stop"></i>';
|
||||
statusIndicator.textContent = 'Recording... Click to stop';
|
||||
})
|
||||
.catch(error => {
|
||||
console.error('Error accessing microphone:', error);
|
||||
alert('Error accessing microphone. Please make sure you have given permission for microphone access.');
|
||||
});
|
||||
}
|
||||
|
||||
// Function to stop recording
|
||||
function stopRecording() {
|
||||
mediaRecorder.stop();
|
||||
isRecording = false;
|
||||
recordBtn.classList.remove('recording');
|
||||
recordBtn.classList.replace('btn-danger', 'btn-primary');
|
||||
recordBtn.innerHTML = '<i class="fas fa-microphone"></i>';
|
||||
statusIndicator.textContent = 'Processing audio...';
|
||||
|
||||
// Stop all audio tracks
|
||||
mediaRecorder.stream.getTracks().forEach(track => track.stop());
|
||||
}
|
||||
|
||||
// Function to transcribe audio
|
||||
function transcribeAudio(audioBlob) {
|
||||
const formData = new FormData();
|
||||
formData.append('audio', audioBlob);
|
||||
formData.append('source_lang', sourceLanguage.value);
|
||||
|
||||
showProgress();
|
||||
|
||||
fetch('/transcribe', {
|
||||
method: 'POST',
|
||||
body: formData
|
||||
})
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
hideProgress();
|
||||
|
||||
if (data.success) {
|
||||
currentSourceText = data.text;
|
||||
sourceText.innerHTML = `<p>${data.text}</p>`;
|
||||
playSource.disabled = false;
|
||||
translateBtn.disabled = false;
|
||||
statusIndicator.textContent = 'Transcription complete';
|
||||
} else {
|
||||
sourceText.innerHTML = `<p class="text-danger">Error: ${data.error}</p>`;
|
||||
statusIndicator.textContent = 'Transcription failed';
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
hideProgress();
|
||||
console.error('Transcription error:', error);
|
||||
sourceText.innerHTML = `<p class="text-danger">Failed to transcribe audio. Please try again.</p>`;
|
||||
statusIndicator.textContent = 'Transcription failed';
|
||||
});
|
||||
}
|
||||
|
||||
// Translate button click event
|
||||
translateBtn.addEventListener('click', function() {
|
||||
if (!currentSourceText) {
|
||||
return;
|
||||
}
|
||||
|
||||
statusIndicator.textContent = 'Translating...';
|
||||
showProgress();
|
||||
|
||||
fetch('/translate', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
text: currentSourceText,
|
||||
source_lang: sourceLanguage.value,
|
||||
target_lang: targetLanguage.value
|
||||
})
|
||||
})
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
hideProgress();
|
||||
|
||||
if (data.success) {
|
||||
currentTranslationText = data.translation;
|
||||
translatedText.innerHTML = `<p>${data.translation}</p>`;
|
||||
playTranslation.disabled = false;
|
||||
statusIndicator.textContent = 'Translation complete';
|
||||
} else {
|
||||
translatedText.innerHTML = `<p class="text-danger">Error: ${data.error}</p>`;
|
||||
statusIndicator.textContent = 'Translation failed';
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
hideProgress();
|
||||
console.error('Translation error:', error);
|
||||
translatedText.innerHTML = `<p class="text-danger">Failed to translate. Please try again.</p>`;
|
||||
statusIndicator.textContent = 'Translation failed';
|
||||
});
|
||||
});
|
||||
|
||||
// Play source text
|
||||
playSource.addEventListener('click', function() {
|
||||
if (!currentSourceText) return;
|
||||
|
||||
playAudio(currentSourceText, sourceLanguage.value);
|
||||
statusIndicator.textContent = 'Playing source audio...';
|
||||
});
|
||||
|
||||
// Play translation
|
||||
playTranslation.addEventListener('click', function() {
|
||||
if (!currentTranslationText) return;
|
||||
|
||||
playAudio(currentTranslationText, targetLanguage.value);
|
||||
statusIndicator.textContent = 'Playing translation audio...';
|
||||
});
|
||||
|
||||
// Function to play audio via TTS
|
||||
function playAudio(text, language) {
|
||||
showProgress();
|
||||
|
||||
fetch('/speak', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
text: text,
|
||||
language: language
|
||||
})
|
||||
})
|
||||
.then(response => response.json())
|
||||
.then(data => {
|
||||
hideProgress();
|
||||
|
||||
if (data.success) {
|
||||
audioPlayer.src = data.audio_url;
|
||||
audioPlayer.onended = function() {
|
||||
statusIndicator.textContent = 'Ready';
|
||||
};
|
||||
audioPlayer.play();
|
||||
} else {
|
||||
statusIndicator.textContent = 'TTS failed';
|
||||
alert('Failed to play audio: ' + data.error);
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
hideProgress();
|
||||
console.error('TTS error:', error);
|
||||
statusIndicator.textContent = 'TTS failed';
|
||||
});
|
||||
}
|
||||
|
||||
// Clear buttons
|
||||
clearSource.addEventListener('click', function() {
|
||||
sourceText.innerHTML = '<p class="text-muted">Your transcribed text will appear here...</p>';
|
||||
currentSourceText = '';
|
||||
playSource.disabled = true;
|
||||
translateBtn.disabled = true;
|
||||
});
|
||||
|
||||
clearTranslation.addEventListener('click', function() {
|
||||
translatedText.innerHTML = '<p class="text-muted">Translation will appear here...</p>';
|
||||
currentTranslationText = '';
|
||||
playTranslation.disabled = true;
|
||||
});
|
||||
|
||||
// Progress indicator functions
|
||||
function showProgress() {
|
||||
progressContainer.classList.remove('d-none');
|
||||
let progress = 0;
|
||||
const interval = setInterval(() => {
|
||||
progress += 5;
|
||||
if (progress > 90) {
|
||||
clearInterval(interval);
|
||||
}
|
||||
progressBar.style.width = `${progress}%`;
|
||||
}, 100);
|
||||
progressBar.dataset.interval = interval;
|
||||
}
|
||||
|
||||
function hideProgress() {
|
||||
const interval = progressBar.dataset.interval;
|
||||
if (interval) {
|
||||
clearInterval(Number(interval));
|
||||
}
|
||||
progressBar.style.width = '100%';
|
||||
setTimeout(() => {
|
||||
progressContainer.classList.add('d-none');
|
||||
progressBar.style.width = '0%';
|
||||
}, 500);
|
||||
}
|
||||
});
|
||||
</script>
|
||||
<script src="/static/js/app.js"></script>
|
||||
<script src="/static/js/dist/app.bundle.js"></script>
|
||||
</body>
|
||||
</html>
|
||||
|
287
templates/login.html
Normal file
287
templates/login.html
Normal file
@ -0,0 +1,287 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="en">
|
||||
<head>
|
||||
<meta charset="UTF-8">
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||||
<title>Login - Talk2Me</title>
|
||||
<link href="https://cdn.jsdelivr.net/npm/bootstrap@5.1.3/dist/css/bootstrap.min.css" rel="stylesheet">
|
||||
<link rel="stylesheet" href="https://cdn.jsdelivr.net/npm/bootstrap-icons@1.7.2/font/bootstrap-icons.css">
|
||||
<style>
|
||||
body {
|
||||
background: linear-gradient(135deg, #667eea 0%, #764ba2 100%);
|
||||
height: 100vh;
|
||||
display: flex;
|
||||
align-items: center;
|
||||
justify-content: center;
|
||||
}
|
||||
.login-container {
|
||||
background: white;
|
||||
padding: 2rem;
|
||||
border-radius: 10px;
|
||||
box-shadow: 0 10px 30px rgba(0, 0, 0, 0.1);
|
||||
width: 100%;
|
||||
max-width: 400px;
|
||||
}
|
||||
.login-header {
|
||||
text-align: center;
|
||||
margin-bottom: 2rem;
|
||||
}
|
||||
.login-header h1 {
|
||||
color: #333;
|
||||
font-size: 2rem;
|
||||
margin-bottom: 0.5rem;
|
||||
}
|
||||
.login-header p {
|
||||
color: #666;
|
||||
margin: 0;
|
||||
}
|
||||
.form-control:focus {
|
||||
border-color: #667eea;
|
||||
box-shadow: 0 0 0 0.2rem rgba(102, 126, 234, 0.25);
|
||||
}
|
||||
.btn-primary {
|
||||
background-color: #667eea;
|
||||
border-color: #667eea;
|
||||
width: 100%;
|
||||
padding: 0.75rem;
|
||||
font-weight: 500;
|
||||
margin-top: 1rem;
|
||||
}
|
||||
.btn-primary:hover {
|
||||
background-color: #5a67d8;
|
||||
border-color: #5a67d8;
|
||||
}
|
||||
.alert {
|
||||
font-size: 0.875rem;
|
||||
}
|
||||
.divider {
|
||||
text-align: center;
|
||||
margin: 1.5rem 0;
|
||||
position: relative;
|
||||
}
|
||||
.divider::before {
|
||||
content: '';
|
||||
position: absolute;
|
||||
top: 50%;
|
||||
left: 0;
|
||||
right: 0;
|
||||
height: 1px;
|
||||
background: #e0e0e0;
|
||||
}
|
||||
.divider span {
|
||||
background: white;
|
||||
padding: 0 1rem;
|
||||
position: relative;
|
||||
color: #666;
|
||||
font-size: 0.875rem;
|
||||
}
|
||||
.api-key-section {
|
||||
margin-top: 1.5rem;
|
||||
padding-top: 1.5rem;
|
||||
border-top: 1px solid #e0e0e0;
|
||||
}
|
||||
.api-key-display {
|
||||
background: #f8f9fa;
|
||||
padding: 0.75rem;
|
||||
border-radius: 5px;
|
||||
font-family: monospace;
|
||||
font-size: 0.875rem;
|
||||
word-break: break-all;
|
||||
margin-top: 0.5rem;
|
||||
}
|
||||
.back-link {
|
||||
text-align: center;
|
||||
margin-top: 1rem;
|
||||
}
|
||||
.loading {
|
||||
display: none;
|
||||
}
|
||||
.loading.show {
|
||||
display: inline-block;
|
||||
}
|
||||
</style>
|
||||
</head>
|
||||
<body>
|
||||
<div class="login-container">
|
||||
<div class="login-header">
|
||||
<h1><i class="bi bi-translate"></i> Talk2Me</h1>
|
||||
<p>Voice Translation Made Simple</p>
|
||||
</div>
|
||||
|
||||
<div id="alertContainer">
|
||||
{% if error %}
|
||||
<div class="alert alert-danger alert-dismissible fade show" role="alert">
|
||||
{{ error }}
|
||||
<button type="button" class="btn-close" data-bs-dismiss="alert"></button>
|
||||
</div>
|
||||
{% endif %}
|
||||
</div>
|
||||
|
||||
<form id="loginForm" method="POST" action="{{ url_for('login', next=request.args.get('next')) }}">
|
||||
<div class="mb-3">
|
||||
<label for="username" class="form-label">Email or Username</label>
|
||||
<div class="input-group">
|
||||
<span class="input-group-text"><i class="bi bi-person"></i></span>
|
||||
<input type="text" class="form-control" id="username" name="username" required autofocus>
|
||||
</div>
|
||||
</div>
|
||||
<div class="mb-3">
|
||||
<label for="password" class="form-label">Password</label>
|
||||
<div class="input-group">
|
||||
<span class="input-group-text"><i class="bi bi-lock"></i></span>
|
||||
<input type="password" class="form-control" id="password" name="password" required>
|
||||
</div>
|
||||
</div>
|
||||
<button type="submit" class="btn btn-primary">
|
||||
<span class="loading spinner-border spinner-border-sm me-2" role="status"></span>
|
||||
<span class="btn-text">Sign In</span>
|
||||
</button>
|
||||
</form>
|
||||
|
||||
<div class="divider">
|
||||
<span>OR</span>
|
||||
</div>
|
||||
|
||||
<div class="text-center">
|
||||
<p class="mb-2">Use the app without signing in</p>
|
||||
<a href="/" class="btn btn-outline-secondary w-100">
|
||||
<i class="bi bi-arrow-right-circle"></i> Continue as Guest
|
||||
</a>
|
||||
</div>
|
||||
|
||||
<div class="api-key-section" id="apiKeySection" style="display: none;">
|
||||
<h6 class="mb-2">Your API Key</h6>
|
||||
<p class="text-muted small">Use this key to authenticate API requests:</p>
|
||||
<div class="api-key-display" id="apiKeyDisplay">
|
||||
<span id="apiKey"></span>
|
||||
<button class="btn btn-sm btn-outline-secondary float-end" onclick="copyApiKey()">
|
||||
<i class="bi bi-clipboard"></i> Copy
|
||||
</button>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<div class="back-link">
|
||||
<a href="/" class="text-muted small">
|
||||
<i class="bi bi-arrow-left"></i> Back to App
|
||||
</a>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<script src="https://cdn.jsdelivr.net/npm/bootstrap@5.1.3/dist/js/bootstrap.bundle.min.js"></script>
|
||||
<script src="https://cdn.jsdelivr.net/npm/axios/dist/axios.min.js"></script>
|
||||
<script>
|
||||
// Check if already logged in
|
||||
const authToken = localStorage.getItem('auth_token');
|
||||
if (authToken) {
|
||||
// Verify token is still valid
|
||||
axios.defaults.headers.common['Authorization'] = `Bearer ${authToken}`;
|
||||
axios.get('/api/auth/profile').then(response => {
|
||||
// Token is valid, redirect to app
|
||||
window.location.href = '/';
|
||||
}).catch(() => {
|
||||
// Token invalid, clear it
|
||||
localStorage.removeItem('auth_token');
|
||||
delete axios.defaults.headers.common['Authorization'];
|
||||
});
|
||||
}
|
||||
|
||||
document.getElementById('loginForm').addEventListener('submit', async (e) => {
|
||||
e.preventDefault();
|
||||
|
||||
const username = document.getElementById('username').value;
|
||||
const password = document.getElementById('password').value;
|
||||
const loadingSpinner = document.querySelector('.loading');
|
||||
const btnText = document.querySelector('.btn-text');
|
||||
const submitBtn = e.target.querySelector('button[type="submit"]');
|
||||
|
||||
// Show loading state
|
||||
loadingSpinner.classList.add('show');
|
||||
btnText.textContent = 'Signing in...';
|
||||
submitBtn.disabled = true;
|
||||
|
||||
try {
|
||||
const response = await axios.post('/api/auth/login', {
|
||||
username: username,
|
||||
password: password
|
||||
});
|
||||
|
||||
if (response.data.success) {
|
||||
// Store token
|
||||
const token = response.data.tokens.access_token;
|
||||
localStorage.setItem('auth_token', token);
|
||||
localStorage.setItem('refresh_token', response.data.tokens.refresh_token);
|
||||
localStorage.setItem('user_id', response.data.user.id);
|
||||
localStorage.setItem('username', response.data.user.username);
|
||||
localStorage.setItem('user_role', response.data.user.role);
|
||||
|
||||
// Show API key
|
||||
document.getElementById('apiKey').textContent = response.data.user.api_key;
|
||||
document.getElementById('apiKeySection').style.display = 'block';
|
||||
|
||||
showAlert('Login successful! Redirecting...', 'success');
|
||||
|
||||
// Redirect based on role or next parameter
|
||||
setTimeout(() => {
|
||||
const urlParams = new URLSearchParams(window.location.search);
|
||||
const nextUrl = urlParams.get('next');
|
||||
|
||||
if (nextUrl) {
|
||||
window.location.href = nextUrl;
|
||||
} else if (response.data.user.role === 'admin') {
|
||||
window.location.href = '/admin';
|
||||
} else {
|
||||
window.location.href = '/';
|
||||
}
|
||||
}, 1500);
|
||||
}
|
||||
} catch (error) {
|
||||
console.error('Login error:', error);
|
||||
const errorMessage = error.response?.data?.error || 'Login failed. Please try again.';
|
||||
showAlert(errorMessage, 'danger');
|
||||
|
||||
// Reset button state
|
||||
loadingSpinner.classList.remove('show');
|
||||
btnText.textContent = 'Sign In';
|
||||
submitBtn.disabled = false;
|
||||
|
||||
// Clear password field on error
|
||||
document.getElementById('password').value = '';
|
||||
document.getElementById('password').focus();
|
||||
}
|
||||
});
|
||||
|
||||
function showAlert(message, type) {
|
||||
const alertContainer = document.getElementById('alertContainer');
|
||||
const alert = document.createElement('div');
|
||||
alert.className = `alert alert-${type} alert-dismissible fade show`;
|
||||
alert.innerHTML = `
|
||||
${message}
|
||||
<button type="button" class="btn-close" data-bs-dismiss="alert"></button>
|
||||
`;
|
||||
alertContainer.innerHTML = '';
|
||||
alertContainer.appendChild(alert);
|
||||
|
||||
// Auto-dismiss after 5 seconds
|
||||
setTimeout(() => {
|
||||
alert.remove();
|
||||
}, 5000);
|
||||
}
|
||||
|
||||
function copyApiKey() {
|
||||
const apiKey = document.getElementById('apiKey').textContent;
|
||||
navigator.clipboard.writeText(apiKey).then(() => {
|
||||
showAlert('API key copied to clipboard!', 'success');
|
||||
}).catch(() => {
|
||||
showAlert('Failed to copy API key', 'danger');
|
||||
});
|
||||
}
|
||||
|
||||
// Handle Enter key in form fields
|
||||
document.getElementById('username').addEventListener('keypress', (e) => {
|
||||
if (e.key === 'Enter') {
|
||||
document.getElementById('password').focus();
|
||||
}
|
||||
});
|
||||
</script>
|
||||
</body>
|
||||
</html>
|
41
tsconfig.json
Normal file
41
tsconfig.json
Normal file
@ -0,0 +1,41 @@
|
||||
{
|
||||
"compilerOptions": {
|
||||
"target": "ES2020",
|
||||
"module": "ES2020",
|
||||
"lib": ["ES2020", "DOM", "DOM.Iterable"],
|
||||
"outDir": "./static/js/dist",
|
||||
"rootDir": "./static/js/src",
|
||||
"strict": true,
|
||||
"esModuleInterop": true,
|
||||
"skipLibCheck": true,
|
||||
"forceConsistentCasingInFileNames": true,
|
||||
"moduleResolution": "node",
|
||||
"resolveJsonModule": true,
|
||||
"declaration": true,
|
||||
"declarationMap": true,
|
||||
"sourceMap": true,
|
||||
"removeComments": false,
|
||||
"noEmitOnError": true,
|
||||
"noImplicitAny": true,
|
||||
"noImplicitThis": true,
|
||||
"noUnusedLocals": true,
|
||||
"noUnusedParameters": true,
|
||||
"noImplicitReturns": true,
|
||||
"noFallthroughCasesInSwitch": true,
|
||||
"strictNullChecks": true,
|
||||
"strictFunctionTypes": true,
|
||||
"strictBindCallApply": true,
|
||||
"strictPropertyInitialization": true,
|
||||
"allowJs": false,
|
||||
"types": [
|
||||
"node"
|
||||
]
|
||||
},
|
||||
"include": [
|
||||
"static/js/src/**/*"
|
||||
],
|
||||
"exclude": [
|
||||
"node_modules",
|
||||
"static/js/dist"
|
||||
]
|
||||
}
|
@ -1,78 +0,0 @@
|
||||
#!/usr/bin/env python
|
||||
"""
|
||||
TTS Debug Script - Tests connection to the OpenAI TTS server
|
||||
"""
|
||||
|
||||
import os
|
||||
import sys
|
||||
import json
|
||||
import requests
|
||||
from argparse import ArgumentParser
|
||||
|
||||
def test_tts_connection(server_url, api_key, text="Hello, this is a test message"):
|
||||
"""Test connection to the TTS server"""
|
||||
|
||||
headers = {
|
||||
"Content-Type": "application/json",
|
||||
"Authorization": f"Bearer {api_key}"
|
||||
}
|
||||
|
||||
payload = {
|
||||
"input": text,
|
||||
"voice": "echo",
|
||||
"response_format": "mp3",
|
||||
"speed": 1.0
|
||||
}
|
||||
|
||||
print(f"Sending request to: {server_url}")
|
||||
print(f"Headers: {headers}")
|
||||
print(f"Payload: {json.dumps(payload, indent=2)}")
|
||||
|
||||
try:
|
||||
response = requests.post(
|
||||
server_url,
|
||||
headers=headers,
|
||||
json=payload,
|
||||
timeout=15
|
||||
)
|
||||
|
||||
print(f"Response status code: {response.status_code}")
|
||||
|
||||
if response.status_code == 200:
|
||||
print("Success! Received audio data")
|
||||
# Save to file
|
||||
output_file = "tts_test_output.mp3"
|
||||
with open(output_file, "wb") as f:
|
||||
f.write(response.content)
|
||||
print(f"Saved audio to {output_file}")
|
||||
return True
|
||||
else:
|
||||
print("Error in response")
|
||||
try:
|
||||
error_data = response.json()
|
||||
print(f"Error details: {json.dumps(error_data, indent=2)}")
|
||||
except:
|
||||
print(f"Raw response: {response.text[:500]}")
|
||||
return False
|
||||
|
||||
except Exception as e:
|
||||
print(f"Error during request: {str(e)}")
|
||||
return False
|
||||
|
||||
def main():
|
||||
parser = ArgumentParser(description="Test connection to OpenAI TTS server")
|
||||
parser.add_argument("--url", default="http://localhost:5050/v1/audio/speech", help="TTS server URL")
|
||||
parser.add_argument("--key", default=os.environ.get("TTS_API_KEY", ""), help="API key")
|
||||
parser.add_argument("--text", default="Hello, this is a test message", help="Text to synthesize")
|
||||
|
||||
args = parser.parse_args()
|
||||
|
||||
if not args.key:
|
||||
print("Error: API key is required. Use --key argument or set TTS_API_KEY environment variable.")
|
||||
return 1
|
||||
|
||||
success = test_tts_connection(args.url, args.key, args.text)
|
||||
return 0 if success else 1
|
||||
|
||||
if __name__ == "__main__":
|
||||
sys.exit(main())
|
Binary file not shown.
352
user_rate_limiter.py
Normal file
352
user_rate_limiter.py
Normal file
@ -0,0 +1,352 @@
|
||||
"""User-specific rate limiting that integrates with authentication"""
|
||||
import time
|
||||
import logging
|
||||
from functools import wraps
|
||||
from flask import request, jsonify, g
|
||||
from collections import defaultdict, deque
|
||||
from threading import Lock
|
||||
from datetime import datetime, timedelta
|
||||
|
||||
from auth import get_current_user
|
||||
from auth_models import User
|
||||
from database import db
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class UserRateLimiter:
|
||||
"""Enhanced rate limiter with user-specific limits"""
|
||||
|
||||
def __init__(self, default_limiter):
|
||||
self.default_limiter = default_limiter
|
||||
self.user_buckets = defaultdict(lambda: {
|
||||
'tokens': 0,
|
||||
'last_update': time.time(),
|
||||
'requests': deque(maxlen=1000)
|
||||
})
|
||||
self.lock = Lock()
|
||||
|
||||
def get_user_limits(self, user: User, endpoint: str):
|
||||
"""Get rate limits for a specific user"""
|
||||
# Start with endpoint-specific or default limits
|
||||
base_limits = self.default_limiter.get_limits(endpoint)
|
||||
|
||||
if not user:
|
||||
return base_limits
|
||||
|
||||
# Override with user-specific limits
|
||||
user_limits = {
|
||||
'requests_per_minute': user.rate_limit_per_minute,
|
||||
'requests_per_hour': user.rate_limit_per_hour,
|
||||
'requests_per_day': user.rate_limit_per_day,
|
||||
'burst_size': base_limits.get('burst_size', 10),
|
||||
'token_refresh_rate': user.rate_limit_per_minute / 60.0 # Convert to per-second
|
||||
}
|
||||
|
||||
# Admin users get higher limits
|
||||
if user.is_admin:
|
||||
user_limits['requests_per_minute'] *= 10
|
||||
user_limits['requests_per_hour'] *= 10
|
||||
user_limits['requests_per_day'] *= 10
|
||||
user_limits['burst_size'] *= 5
|
||||
|
||||
return user_limits
|
||||
|
||||
def check_user_rate_limit(self, user: User, endpoint: str, request_size: int = 0):
|
||||
"""Check rate limit for authenticated user"""
|
||||
if not user:
|
||||
# Fall back to IP-based limiting
|
||||
client_id = self.default_limiter.get_client_id(request)
|
||||
return self.default_limiter.check_rate_limit(client_id, endpoint, request_size)
|
||||
|
||||
with self.lock:
|
||||
user_id = str(user.id)
|
||||
limits = self.get_user_limits(user, endpoint)
|
||||
|
||||
# Get or create bucket for user
|
||||
bucket = self.user_buckets[user_id]
|
||||
now = time.time()
|
||||
|
||||
# Update tokens based on time passed
|
||||
time_passed = now - bucket['last_update']
|
||||
new_tokens = time_passed * limits['token_refresh_rate']
|
||||
bucket['tokens'] = min(
|
||||
limits['burst_size'],
|
||||
bucket['tokens'] + new_tokens
|
||||
)
|
||||
bucket['last_update'] = now
|
||||
|
||||
# Clean old requests from sliding windows
|
||||
minute_ago = now - 60
|
||||
hour_ago = now - 3600
|
||||
day_ago = now - 86400
|
||||
|
||||
bucket['requests'] = deque(
|
||||
(t for t in bucket['requests'] if t > day_ago),
|
||||
maxlen=1000
|
||||
)
|
||||
|
||||
# Count requests in windows
|
||||
requests_last_minute = sum(1 for t in bucket['requests'] if t > minute_ago)
|
||||
requests_last_hour = sum(1 for t in bucket['requests'] if t > hour_ago)
|
||||
requests_last_day = len(bucket['requests'])
|
||||
|
||||
# Check limits
|
||||
if requests_last_minute >= limits['requests_per_minute']:
|
||||
return False, "Rate limit exceeded (per minute)", {
|
||||
'retry_after': 60,
|
||||
'limit': limits['requests_per_minute'],
|
||||
'remaining': 0,
|
||||
'reset': int(minute_ago + 60),
|
||||
'scope': 'user'
|
||||
}
|
||||
|
||||
if requests_last_hour >= limits['requests_per_hour']:
|
||||
return False, "Rate limit exceeded (per hour)", {
|
||||
'retry_after': 3600,
|
||||
'limit': limits['requests_per_hour'],
|
||||
'remaining': 0,
|
||||
'reset': int(hour_ago + 3600),
|
||||
'scope': 'user'
|
||||
}
|
||||
|
||||
if requests_last_day >= limits['requests_per_day']:
|
||||
return False, "Rate limit exceeded (per day)", {
|
||||
'retry_after': 86400,
|
||||
'limit': limits['requests_per_day'],
|
||||
'remaining': 0,
|
||||
'reset': int(day_ago + 86400),
|
||||
'scope': 'user'
|
||||
}
|
||||
|
||||
# Check token bucket
|
||||
if bucket['tokens'] < 1:
|
||||
retry_after = int(1 / limits['token_refresh_rate'])
|
||||
return False, "Rate limit exceeded (burst)", {
|
||||
'retry_after': retry_after,
|
||||
'limit': limits['burst_size'],
|
||||
'remaining': 0,
|
||||
'reset': int(now + retry_after),
|
||||
'scope': 'user'
|
||||
}
|
||||
|
||||
# Request allowed
|
||||
bucket['tokens'] -= 1
|
||||
bucket['requests'].append(now)
|
||||
|
||||
# Calculate remaining
|
||||
remaining_minute = limits['requests_per_minute'] - requests_last_minute - 1
|
||||
remaining_hour = limits['requests_per_hour'] - requests_last_hour - 1
|
||||
remaining_day = limits['requests_per_day'] - requests_last_day - 1
|
||||
|
||||
return True, None, {
|
||||
'limit_minute': limits['requests_per_minute'],
|
||||
'limit_hour': limits['requests_per_hour'],
|
||||
'limit_day': limits['requests_per_day'],
|
||||
'remaining_minute': remaining_minute,
|
||||
'remaining_hour': remaining_hour,
|
||||
'remaining_day': remaining_day,
|
||||
'reset': int(minute_ago + 60),
|
||||
'scope': 'user'
|
||||
}
|
||||
|
||||
def get_user_usage_stats(self, user: User):
|
||||
"""Get usage statistics for a user"""
|
||||
if not user:
|
||||
return None
|
||||
|
||||
with self.lock:
|
||||
user_id = str(user.id)
|
||||
if user_id not in self.user_buckets:
|
||||
return {
|
||||
'requests_last_minute': 0,
|
||||
'requests_last_hour': 0,
|
||||
'requests_last_day': 0,
|
||||
'tokens_available': 0
|
||||
}
|
||||
|
||||
bucket = self.user_buckets[user_id]
|
||||
now = time.time()
|
||||
minute_ago = now - 60
|
||||
hour_ago = now - 3600
|
||||
day_ago = now - 86400
|
||||
|
||||
requests_last_minute = sum(1 for t in bucket['requests'] if t > minute_ago)
|
||||
requests_last_hour = sum(1 for t in bucket['requests'] if t > hour_ago)
|
||||
requests_last_day = sum(1 for t in bucket['requests'] if t > day_ago)
|
||||
|
||||
return {
|
||||
'requests_last_minute': requests_last_minute,
|
||||
'requests_last_hour': requests_last_hour,
|
||||
'requests_last_day': requests_last_day,
|
||||
'tokens_available': bucket['tokens'],
|
||||
'last_request': bucket['last_update']
|
||||
}
|
||||
|
||||
def reset_user_limits(self, user: User):
|
||||
"""Reset rate limits for a user (admin action)"""
|
||||
if not user:
|
||||
return False
|
||||
|
||||
with self.lock:
|
||||
user_id = str(user.id)
|
||||
if user_id in self.user_buckets:
|
||||
del self.user_buckets[user_id]
|
||||
return True
|
||||
return False
|
||||
|
||||
|
||||
# Global user rate limiter instance
|
||||
from rate_limiter import rate_limiter as default_rate_limiter
|
||||
user_rate_limiter = UserRateLimiter(default_rate_limiter)
|
||||
|
||||
|
||||
def user_aware_rate_limit(endpoint=None,
|
||||
requests_per_minute=None,
|
||||
requests_per_hour=None,
|
||||
requests_per_day=None,
|
||||
burst_size=None,
|
||||
check_size=False,
|
||||
require_auth=False):
|
||||
"""
|
||||
Enhanced rate limiting decorator that considers user authentication
|
||||
|
||||
Usage:
|
||||
@app.route('/api/endpoint')
|
||||
@user_aware_rate_limit(requests_per_minute=10)
|
||||
def endpoint():
|
||||
return jsonify({'status': 'ok'})
|
||||
"""
|
||||
def decorator(f):
|
||||
@wraps(f)
|
||||
def decorated_function(*args, **kwargs):
|
||||
# Get current user (if authenticated)
|
||||
user = get_current_user()
|
||||
|
||||
# If auth is required but no user, return 401
|
||||
if require_auth and not user:
|
||||
return jsonify({
|
||||
'success': False,
|
||||
'error': 'Authentication required',
|
||||
'code': 'auth_required'
|
||||
}), 401
|
||||
|
||||
# Get endpoint
|
||||
endpoint_path = endpoint or request.endpoint
|
||||
|
||||
# Check request size if needed
|
||||
request_size = 0
|
||||
if check_size:
|
||||
request_size = request.content_length or 0
|
||||
|
||||
# Check rate limit
|
||||
if user:
|
||||
# User-specific rate limiting
|
||||
allowed, message, headers = user_rate_limiter.check_user_rate_limit(
|
||||
user, endpoint_path, request_size
|
||||
)
|
||||
else:
|
||||
# Fall back to IP-based rate limiting
|
||||
client_id = default_rate_limiter.get_client_id(request)
|
||||
allowed, message, headers = default_rate_limiter.check_rate_limit(
|
||||
client_id, endpoint_path, request_size
|
||||
)
|
||||
|
||||
if not allowed:
|
||||
# Log excessive requests
|
||||
identifier = f"user:{user.username}" if user else f"ip:{request.remote_addr}"
|
||||
logger.warning(f"Rate limit exceeded for {identifier} on {endpoint_path}: {message}")
|
||||
|
||||
# Update user stats if authenticated
|
||||
if user:
|
||||
user.last_active_at = datetime.utcnow()
|
||||
db.session.commit()
|
||||
|
||||
response = jsonify({
|
||||
'success': False,
|
||||
'error': message,
|
||||
'retry_after': headers.get('retry_after') if headers else 60
|
||||
})
|
||||
response.status_code = 429
|
||||
|
||||
# Add rate limit headers
|
||||
if headers:
|
||||
if headers.get('scope') == 'user':
|
||||
response.headers['X-RateLimit-Limit'] = str(headers.get('limit_minute', 60))
|
||||
response.headers['X-RateLimit-Remaining'] = str(headers.get('remaining_minute', 0))
|
||||
response.headers['X-RateLimit-Limit-Hour'] = str(headers.get('limit_hour', 1000))
|
||||
response.headers['X-RateLimit-Remaining-Hour'] = str(headers.get('remaining_hour', 0))
|
||||
response.headers['X-RateLimit-Limit-Day'] = str(headers.get('limit_day', 10000))
|
||||
response.headers['X-RateLimit-Remaining-Day'] = str(headers.get('remaining_day', 0))
|
||||
else:
|
||||
response.headers['X-RateLimit-Limit'] = str(headers['limit'])
|
||||
response.headers['X-RateLimit-Remaining'] = str(headers['remaining'])
|
||||
response.headers['X-RateLimit-Reset'] = str(headers['reset'])
|
||||
response.headers['Retry-After'] = str(headers['retry_after'])
|
||||
|
||||
return response
|
||||
|
||||
# Track concurrent requests
|
||||
default_rate_limiter.increment_concurrent()
|
||||
|
||||
try:
|
||||
# Store user in g if authenticated
|
||||
if user:
|
||||
g.current_user = user
|
||||
|
||||
# Add rate limit info to response
|
||||
g.rate_limit_headers = headers
|
||||
response = f(*args, **kwargs)
|
||||
|
||||
# Add headers to successful response
|
||||
if headers and hasattr(response, 'headers'):
|
||||
if headers.get('scope') == 'user':
|
||||
response.headers['X-RateLimit-Limit'] = str(headers.get('limit_minute', 60))
|
||||
response.headers['X-RateLimit-Remaining'] = str(headers.get('remaining_minute', 0))
|
||||
response.headers['X-RateLimit-Limit-Hour'] = str(headers.get('limit_hour', 1000))
|
||||
response.headers['X-RateLimit-Remaining-Hour'] = str(headers.get('remaining_hour', 0))
|
||||
response.headers['X-RateLimit-Limit-Day'] = str(headers.get('limit_day', 10000))
|
||||
response.headers['X-RateLimit-Remaining-Day'] = str(headers.get('remaining_day', 0))
|
||||
else:
|
||||
response.headers['X-RateLimit-Limit'] = str(headers.get('limit', 60))
|
||||
response.headers['X-RateLimit-Remaining'] = str(headers.get('remaining', 0))
|
||||
response.headers['X-RateLimit-Reset'] = str(headers['reset'])
|
||||
|
||||
return response
|
||||
finally:
|
||||
default_rate_limiter.decrement_concurrent()
|
||||
|
||||
return decorated_function
|
||||
return decorator
|
||||
|
||||
|
||||
def get_user_rate_limit_status(user: User = None):
|
||||
"""Get current rate limit status for a user or IP"""
|
||||
if not user:
|
||||
user = get_current_user()
|
||||
|
||||
if user:
|
||||
stats = user_rate_limiter.get_user_usage_stats(user)
|
||||
limits = user_rate_limiter.get_user_limits(user, request.endpoint or '/')
|
||||
|
||||
return {
|
||||
'type': 'user',
|
||||
'identifier': user.username,
|
||||
'limits': {
|
||||
'per_minute': limits['requests_per_minute'],
|
||||
'per_hour': limits['requests_per_hour'],
|
||||
'per_day': limits['requests_per_day']
|
||||
},
|
||||
'usage': stats
|
||||
}
|
||||
else:
|
||||
# IP-based stats
|
||||
client_id = default_rate_limiter.get_client_id(request)
|
||||
stats = default_rate_limiter.get_client_stats(client_id)
|
||||
|
||||
return {
|
||||
'type': 'ip',
|
||||
'identifier': request.remote_addr,
|
||||
'limits': default_rate_limiter.default_limits,
|
||||
'usage': stats
|
||||
}
|
121
validate-pwa.html
Normal file
121
validate-pwa.html
Normal file
@ -0,0 +1,121 @@
|
||||
<!DOCTYPE html>
|
||||
<html>
|
||||
<head>
|
||||
<title>PWA Validation - Talk2Me</title>
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||||
<style>
|
||||
body { font-family: Arial, sans-serif; padding: 20px; }
|
||||
.status { padding: 10px; margin: 10px 0; border-radius: 5px; }
|
||||
.success { background: #d4edda; color: #155724; }
|
||||
.error { background: #f8d7da; color: #721c24; }
|
||||
.info { background: #d1ecf1; color: #0c5460; }
|
||||
img { max-width: 100px; height: auto; margin: 10px; }
|
||||
</style>
|
||||
</head>
|
||||
<body>
|
||||
<h1>Talk2Me PWA Validation</h1>
|
||||
|
||||
<h2>Manifest Check</h2>
|
||||
<div id="manifest-status"></div>
|
||||
|
||||
<h2>Icon Check</h2>
|
||||
<div id="icon-status"></div>
|
||||
|
||||
<h2>Service Worker Check</h2>
|
||||
<div id="sw-status"></div>
|
||||
|
||||
<h2>Installation Test</h2>
|
||||
<button id="install-btn" style="display:none; padding: 10px 20px; font-size: 16px;">Install Talk2Me</button>
|
||||
<div id="install-status"></div>
|
||||
|
||||
<script>
|
||||
// Check manifest
|
||||
fetch('/static/manifest.json')
|
||||
.then(res => res.json())
|
||||
.then(manifest => {
|
||||
const status = document.getElementById('manifest-status');
|
||||
status.innerHTML = `
|
||||
<div class="status success">✓ Manifest loaded successfully</div>
|
||||
<div class="status info">Name: ${manifest.name}</div>
|
||||
<div class="status info">Short Name: ${manifest.short_name}</div>
|
||||
<div class="status info">Icons: ${manifest.icons.length} defined</div>
|
||||
`;
|
||||
|
||||
// Check icons
|
||||
const iconStatus = document.getElementById('icon-status');
|
||||
manifest.icons.forEach(icon => {
|
||||
const img = new Image();
|
||||
img.onload = () => {
|
||||
const div = document.createElement('div');
|
||||
div.className = 'status success';
|
||||
div.innerHTML = `✓ ${icon.src} (${icon.sizes}) - ${icon.purpose}`;
|
||||
iconStatus.appendChild(div);
|
||||
iconStatus.appendChild(img);
|
||||
};
|
||||
img.onerror = () => {
|
||||
const div = document.createElement('div');
|
||||
div.className = 'status error';
|
||||
div.innerHTML = `✗ ${icon.src} (${icon.sizes}) - Failed to load`;
|
||||
iconStatus.appendChild(div);
|
||||
};
|
||||
img.src = icon.src;
|
||||
img.style.maxWidth = '50px';
|
||||
});
|
||||
})
|
||||
.catch(err => {
|
||||
document.getElementById('manifest-status').innerHTML =
|
||||
`<div class="status error">✗ Failed to load manifest: ${err.message}</div>`;
|
||||
});
|
||||
|
||||
// Check service worker
|
||||
if ('serviceWorker' in navigator) {
|
||||
navigator.serviceWorker.getRegistration()
|
||||
.then(reg => {
|
||||
const status = document.getElementById('sw-status');
|
||||
if (reg) {
|
||||
status.innerHTML = `
|
||||
<div class="status success">✓ Service Worker is registered</div>
|
||||
<div class="status info">Scope: ${reg.scope}</div>
|
||||
<div class="status info">State: ${reg.active ? 'Active' : 'Not Active'}</div>
|
||||
`;
|
||||
} else {
|
||||
status.innerHTML = '<div class="status error">✗ Service Worker not registered</div>';
|
||||
}
|
||||
})
|
||||
.catch(err => {
|
||||
document.getElementById('sw-status').innerHTML =
|
||||
`<div class="status error">✗ Service Worker error: ${err.message}</div>`;
|
||||
});
|
||||
} else {
|
||||
document.getElementById('sw-status').innerHTML =
|
||||
'<div class="status error">✗ Service Workers not supported</div>';
|
||||
}
|
||||
|
||||
// Check install prompt
|
||||
let deferredPrompt;
|
||||
window.addEventListener('beforeinstallprompt', (e) => {
|
||||
e.preventDefault();
|
||||
deferredPrompt = e;
|
||||
document.getElementById('install-btn').style.display = 'block';
|
||||
document.getElementById('install-status').innerHTML =
|
||||
'<div class="status success">✓ App is installable</div>';
|
||||
});
|
||||
|
||||
document.getElementById('install-btn').addEventListener('click', async () => {
|
||||
if (deferredPrompt) {
|
||||
deferredPrompt.prompt();
|
||||
const { outcome } = await deferredPrompt.userChoice;
|
||||
document.getElementById('install-status').innerHTML +=
|
||||
`<div class="status info">User ${outcome} the install</div>`;
|
||||
deferredPrompt = null;
|
||||
}
|
||||
});
|
||||
|
||||
// Check if already installed
|
||||
if (window.matchMedia('(display-mode: standalone)').matches) {
|
||||
document.getElementById('install-status').innerHTML =
|
||||
'<div class="status success">✓ App is already installed</div>';
|
||||
}
|
||||
</script>
|
||||
</body>
|
||||
</html>
|
256
validators.py
Normal file
256
validators.py
Normal file
@ -0,0 +1,256 @@
|
||||
"""
|
||||
Input validation and sanitization for the Talk2Me application
|
||||
"""
|
||||
import re
|
||||
import html
|
||||
from typing import Optional, Dict, Any, Tuple
|
||||
import os
|
||||
|
||||
class Validators:
|
||||
# Maximum sizes
|
||||
MAX_TEXT_LENGTH = 10000
|
||||
MAX_AUDIO_SIZE = 25 * 1024 * 1024 # 25MB
|
||||
MAX_URL_LENGTH = 2048
|
||||
MAX_API_KEY_LENGTH = 128
|
||||
|
||||
# Allowed audio formats
|
||||
ALLOWED_AUDIO_EXTENSIONS = {'.webm', '.ogg', '.wav', '.mp3', '.mp4', '.m4a'}
|
||||
ALLOWED_AUDIO_MIMETYPES = {
|
||||
'audio/webm', 'audio/ogg', 'audio/wav', 'audio/mp3',
|
||||
'audio/mpeg', 'audio/mp4', 'audio/x-m4a', 'audio/x-wav'
|
||||
}
|
||||
|
||||
@staticmethod
|
||||
def sanitize_text(text: str, max_length: int = None) -> str:
|
||||
"""Sanitize text input by removing dangerous characters"""
|
||||
if not isinstance(text, str):
|
||||
return ""
|
||||
|
||||
if max_length is None:
|
||||
max_length = Validators.MAX_TEXT_LENGTH
|
||||
|
||||
# Trim and limit length
|
||||
text = text.strip()[:max_length]
|
||||
|
||||
# Remove null bytes
|
||||
text = text.replace('\x00', '')
|
||||
|
||||
# Remove control characters except newlines and tabs
|
||||
text = re.sub(r'[\x00-\x08\x0B\x0C\x0E-\x1F\x7F]', '', text)
|
||||
|
||||
return text
|
||||
|
||||
@staticmethod
|
||||
def sanitize_html(text: str) -> str:
|
||||
"""Escape HTML to prevent XSS"""
|
||||
if not isinstance(text, str):
|
||||
return ""
|
||||
return html.escape(text)
|
||||
|
||||
@staticmethod
|
||||
def validate_language_code(code: str, allowed_languages: set) -> Optional[str]:
|
||||
"""Validate language code against allowed list"""
|
||||
if not code or not isinstance(code, str):
|
||||
return None
|
||||
|
||||
code = code.strip().lower()
|
||||
|
||||
# Check if it's in the allowed list or is 'auto'
|
||||
if code in allowed_languages or code == 'auto':
|
||||
return code
|
||||
|
||||
return None
|
||||
|
||||
@staticmethod
|
||||
def validate_audio_file(file_storage) -> Tuple[bool, Optional[str]]:
|
||||
"""Validate uploaded audio file"""
|
||||
if not file_storage:
|
||||
return False, "No file provided"
|
||||
|
||||
# Check file size
|
||||
file_storage.seek(0, os.SEEK_END)
|
||||
size = file_storage.tell()
|
||||
file_storage.seek(0)
|
||||
|
||||
if size > Validators.MAX_AUDIO_SIZE:
|
||||
return False, f"File size exceeds {Validators.MAX_AUDIO_SIZE // (1024*1024)}MB limit"
|
||||
|
||||
# Check file extension
|
||||
if file_storage.filename:
|
||||
ext = os.path.splitext(file_storage.filename.lower())[1]
|
||||
if ext not in Validators.ALLOWED_AUDIO_EXTENSIONS:
|
||||
return False, "Invalid audio file type"
|
||||
|
||||
# Check MIME type if available
|
||||
if hasattr(file_storage, 'content_type') and file_storage.content_type:
|
||||
if file_storage.content_type not in Validators.ALLOWED_AUDIO_MIMETYPES:
|
||||
# Allow generic application/octet-stream as browsers sometimes use this
|
||||
if file_storage.content_type != 'application/octet-stream':
|
||||
return False, "Invalid audio MIME type"
|
||||
|
||||
return True, None
|
||||
|
||||
@staticmethod
|
||||
def validate_email(email: str) -> bool:
|
||||
"""Validate email address format"""
|
||||
if not email or not isinstance(email, str):
|
||||
return False
|
||||
|
||||
# Basic email pattern
|
||||
email_pattern = re.compile(
|
||||
r'^[a-zA-Z0-9._%+-]+@[a-zA-Z0-9.-]+\.[a-zA-Z]{2,}$'
|
||||
)
|
||||
|
||||
return bool(email_pattern.match(email))
|
||||
|
||||
@staticmethod
|
||||
def validate_url(url: str) -> Optional[str]:
|
||||
"""Validate and sanitize URL"""
|
||||
if not url or not isinstance(url, str):
|
||||
return None
|
||||
|
||||
url = url.strip()
|
||||
|
||||
# Check length
|
||||
if len(url) > Validators.MAX_URL_LENGTH:
|
||||
return None
|
||||
|
||||
# Basic URL pattern check
|
||||
url_pattern = re.compile(
|
||||
r'^https?://' # http:// or https://
|
||||
r'(?:(?:[A-Z0-9](?:[A-Z0-9-]{0,61}[A-Z0-9])?\.)+[A-Z]{2,6}\.?|' # domain...
|
||||
r'localhost|' # localhost...
|
||||
r'\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}|' # ...or ipv4
|
||||
r'\[?[A-F0-9]*:[A-F0-9:]+\]?)' # ...or ipv6
|
||||
r'(?::\d+)?' # optional port
|
||||
r'(?:/?|[/?]\S+)$', re.IGNORECASE)
|
||||
|
||||
if not url_pattern.match(url):
|
||||
return None
|
||||
|
||||
# Prevent some common injection attempts
|
||||
dangerous_patterns = [
|
||||
'javascript:', 'data:', 'vbscript:', 'file:', 'about:', 'chrome:'
|
||||
]
|
||||
if any(pattern in url.lower() for pattern in dangerous_patterns):
|
||||
return None
|
||||
|
||||
return url
|
||||
|
||||
@staticmethod
|
||||
def validate_api_key(key: str) -> Optional[str]:
|
||||
"""Validate API key format"""
|
||||
if not key or not isinstance(key, str):
|
||||
return None
|
||||
|
||||
key = key.strip()
|
||||
|
||||
# Check length
|
||||
if len(key) < 20 or len(key) > Validators.MAX_API_KEY_LENGTH:
|
||||
return None
|
||||
|
||||
# Only allow alphanumeric, dash, and underscore
|
||||
if not re.match(r'^[a-zA-Z0-9\-_]+$', key):
|
||||
return None
|
||||
|
||||
return key
|
||||
|
||||
@staticmethod
|
||||
def sanitize_filename(filename: str) -> str:
|
||||
"""Sanitize filename to prevent directory traversal"""
|
||||
if not filename or not isinstance(filename, str):
|
||||
return "file"
|
||||
|
||||
# Remove any path components
|
||||
filename = os.path.basename(filename)
|
||||
|
||||
# Remove dangerous characters
|
||||
filename = re.sub(r'[^a-zA-Z0-9.\-_]', '_', filename)
|
||||
|
||||
# Limit length
|
||||
if len(filename) > 255:
|
||||
name, ext = os.path.splitext(filename)
|
||||
max_name_length = 255 - len(ext)
|
||||
filename = name[:max_name_length] + ext
|
||||
|
||||
# Don't allow hidden files
|
||||
if filename.startswith('.'):
|
||||
filename = '_' + filename[1:]
|
||||
|
||||
return filename or "file"
|
||||
|
||||
@staticmethod
|
||||
def validate_json_size(data: Dict[str, Any], max_size_kb: int = 1024) -> bool:
|
||||
"""Check if JSON data size is within limits"""
|
||||
try:
|
||||
import json
|
||||
json_str = json.dumps(data)
|
||||
size_kb = len(json_str.encode('utf-8')) / 1024
|
||||
return size_kb <= max_size_kb
|
||||
except:
|
||||
return False
|
||||
|
||||
@staticmethod
|
||||
def validate_settings(settings: Dict[str, Any]) -> Tuple[bool, Dict[str, Any], list]:
|
||||
"""Validate settings object"""
|
||||
errors = []
|
||||
sanitized = {}
|
||||
|
||||
# Boolean settings
|
||||
bool_settings = [
|
||||
'notificationsEnabled', 'notifyTranscription',
|
||||
'notifyTranslation', 'notifyErrors', 'offlineMode'
|
||||
]
|
||||
|
||||
for setting in bool_settings:
|
||||
if setting in settings:
|
||||
sanitized[setting] = bool(settings[setting])
|
||||
|
||||
# URL validation
|
||||
if 'ttsServerUrl' in settings and settings['ttsServerUrl']:
|
||||
url = Validators.validate_url(settings['ttsServerUrl'])
|
||||
if not url:
|
||||
errors.append('Invalid TTS server URL')
|
||||
else:
|
||||
sanitized['ttsServerUrl'] = url
|
||||
|
||||
# API key validation
|
||||
if 'ttsApiKey' in settings and settings['ttsApiKey']:
|
||||
key = Validators.validate_api_key(settings['ttsApiKey'])
|
||||
if not key:
|
||||
errors.append('Invalid API key format')
|
||||
else:
|
||||
sanitized['ttsApiKey'] = key
|
||||
|
||||
return len(errors) == 0, sanitized, errors
|
||||
|
||||
@staticmethod
|
||||
def rate_limit_check(identifier: str, action: str, max_requests: int = 10,
|
||||
window_seconds: int = 60, storage: Dict = None) -> bool:
|
||||
"""
|
||||
Simple rate limiting check
|
||||
Returns True if request is allowed, False if rate limited
|
||||
"""
|
||||
import time
|
||||
|
||||
if storage is None:
|
||||
return True # Can't track without storage
|
||||
|
||||
key = f"{identifier}:{action}"
|
||||
current_time = time.time()
|
||||
window_start = current_time - window_seconds
|
||||
|
||||
# Get or create request list
|
||||
if key not in storage:
|
||||
storage[key] = []
|
||||
|
||||
# Remove old requests outside the window
|
||||
storage[key] = [t for t in storage[key] if t > window_start]
|
||||
|
||||
# Check if limit exceeded
|
||||
if len(storage[key]) >= max_requests:
|
||||
return False
|
||||
|
||||
# Add current request
|
||||
storage[key].append(current_time)
|
||||
return True
|
23
webpack.config.js
Normal file
23
webpack.config.js
Normal file
@ -0,0 +1,23 @@
|
||||
const path = require('path');
|
||||
|
||||
module.exports = {
|
||||
entry: './static/js/src/app.ts',
|
||||
module: {
|
||||
rules: [
|
||||
{
|
||||
test: /\.tsx?$/,
|
||||
use: 'ts-loader',
|
||||
exclude: /node_modules/,
|
||||
},
|
||||
],
|
||||
},
|
||||
resolve: {
|
||||
extensions: ['.tsx', '.ts', '.js'],
|
||||
},
|
||||
output: {
|
||||
filename: 'app.bundle.js',
|
||||
path: path.resolve(__dirname, 'static/js/dist'),
|
||||
},
|
||||
mode: 'production',
|
||||
devtool: 'source-map',
|
||||
};
|
39
whisper_config.py
Normal file
39
whisper_config.py
Normal file
@ -0,0 +1,39 @@
|
||||
"""
|
||||
Whisper Model Configuration and Optimization Settings
|
||||
"""
|
||||
|
||||
# Model selection based on available resources
|
||||
# Available models: tiny, base, small, medium, large
|
||||
MODEL_SIZE = "base" # ~140MB, good balance of speed and accuracy
|
||||
|
||||
# GPU Optimization Settings
|
||||
GPU_OPTIMIZATIONS = {
|
||||
"enable_tf32": True, # TensorFloat-32 for Ampere GPUs
|
||||
"enable_cudnn_benchmark": True, # Auto-tune convolution algorithms
|
||||
"use_fp16": True, # Half precision for faster inference
|
||||
"pre_allocate_memory": True, # Reduce memory fragmentation
|
||||
"warm_up_gpu": True # Cache CUDA kernels on startup
|
||||
}
|
||||
|
||||
# Transcription Settings for Speed
|
||||
TRANSCRIBE_OPTIONS = {
|
||||
"task": "transcribe",
|
||||
"temperature": 0, # Disable sampling
|
||||
"best_of": 1, # No beam search
|
||||
"beam_size": 1, # Single beam
|
||||
"condition_on_previous_text": False, # Faster inference
|
||||
"compression_ratio_threshold": 2.4,
|
||||
"logprob_threshold": -1.0,
|
||||
"no_speech_threshold": 0.6,
|
||||
"word_timestamps": False # Disable if not needed
|
||||
}
|
||||
|
||||
# Memory Management
|
||||
MEMORY_SETTINGS = {
|
||||
"clear_cache_after_transcribe": True,
|
||||
"force_garbage_collection": True,
|
||||
"max_concurrent_transcriptions": 1 # Prevent memory overflow
|
||||
}
|
||||
|
||||
# Performance Monitoring
|
||||
ENABLE_PERFORMANCE_LOGGING = True
|
34
wsgi.py
Normal file
34
wsgi.py
Normal file
@ -0,0 +1,34 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
WSGI entry point for production deployment
|
||||
"""
|
||||
import os
|
||||
import sys
|
||||
from pathlib import Path
|
||||
|
||||
# Add the project directory to the Python path
|
||||
project_root = Path(__file__).parent.absolute()
|
||||
sys.path.insert(0, str(project_root))
|
||||
|
||||
# Set production environment
|
||||
os.environ['FLASK_ENV'] = 'production'
|
||||
|
||||
# Import and configure the Flask app
|
||||
from app import app
|
||||
|
||||
# Production configuration overrides
|
||||
app.config.update(
|
||||
DEBUG=False,
|
||||
TESTING=False,
|
||||
# Ensure proper secret key is set in production
|
||||
SECRET_KEY=os.environ.get('SECRET_KEY', app.config.get('SECRET_KEY'))
|
||||
)
|
||||
|
||||
# Create the WSGI application
|
||||
application = app
|
||||
|
||||
if __name__ == '__main__':
|
||||
# This is only for development/testing
|
||||
# In production, use: gunicorn wsgi:application
|
||||
print("Warning: Running WSGI directly. Use a proper WSGI server in production!")
|
||||
application.run(host='0.0.0.0', port=5005)
|
Loading…
Reference in New Issue
Block a user