221
qwen3-tts-backend/main.py
Normal file
221
qwen3-tts-backend/main.py
Normal file
@@ -0,0 +1,221 @@
|
||||
import logging
|
||||
import sys
|
||||
from contextlib import asynccontextmanager
|
||||
from pathlib import Path
|
||||
|
||||
import torch
|
||||
from fastapi import FastAPI, Request
|
||||
from fastapi.middleware.cors import CORSMiddleware
|
||||
from slowapi import Limiter, _rate_limit_exceeded_handler
|
||||
from slowapi.util import get_remote_address
|
||||
from slowapi.errors import RateLimitExceeded
|
||||
from sqlalchemy import text
|
||||
|
||||
from core.config import settings
|
||||
from core.database import init_db
|
||||
from core.model_manager import ModelManager
|
||||
from core.cleanup import run_scheduled_cleanup
|
||||
from api import auth, jobs, tts, cache, metrics, users
|
||||
from apscheduler.schedulers.asyncio import AsyncIOScheduler
|
||||
|
||||
logging.basicConfig(
|
||||
level=getattr(logging, settings.LOG_LEVEL.upper()),
|
||||
format='%(asctime)s - %(name)s - %(levelname)s - %(message)s',
|
||||
handlers=[
|
||||
logging.StreamHandler(sys.stdout),
|
||||
logging.FileHandler(settings.LOG_FILE)
|
||||
]
|
||||
)
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
def get_user_identifier(request: Request) -> str:
|
||||
from jose import jwt
|
||||
from core.config import settings
|
||||
|
||||
auth_header = request.headers.get("Authorization", "")
|
||||
if auth_header.startswith("Bearer "):
|
||||
token = auth_header.split(" ")[1]
|
||||
try:
|
||||
payload = jwt.decode(token, settings.SECRET_KEY, algorithms=[settings.ALGORITHM])
|
||||
user_id = payload.get("sub")
|
||||
if user_id:
|
||||
return f"user:{user_id}"
|
||||
except Exception:
|
||||
pass
|
||||
|
||||
return get_remote_address(request)
|
||||
|
||||
limiter = Limiter(key_func=get_user_identifier)
|
||||
|
||||
@asynccontextmanager
|
||||
async def lifespan(app: FastAPI):
|
||||
logger.info("Starting Qwen3-TTS Backend Service...")
|
||||
logger.info(f"Model base path: {settings.MODEL_BASE_PATH}")
|
||||
logger.info(f"Cache directory: {settings.CACHE_DIR}")
|
||||
logger.info(f"Output directory: {settings.OUTPUT_DIR}")
|
||||
logger.info(f"Device: {settings.MODEL_DEVICE}")
|
||||
|
||||
try:
|
||||
settings.validate()
|
||||
logger.info("Configuration validated successfully")
|
||||
except Exception as e:
|
||||
logger.error(f"Configuration validation failed: {e}")
|
||||
raise
|
||||
|
||||
try:
|
||||
init_db()
|
||||
logger.info("Database initialized successfully")
|
||||
except Exception as e:
|
||||
logger.error(f"Database initialization failed: {e}")
|
||||
raise
|
||||
|
||||
try:
|
||||
model_manager = await ModelManager.get_instance()
|
||||
await model_manager.load_model("custom-voice")
|
||||
logger.info("Preloaded custom-voice model")
|
||||
except Exception as e:
|
||||
logger.warning(f"Model preload failed: {e}")
|
||||
|
||||
scheduler = AsyncIOScheduler()
|
||||
scheduler.add_job(
|
||||
run_scheduled_cleanup,
|
||||
'interval',
|
||||
hours=6,
|
||||
args=[str(settings.DATABASE_URL)],
|
||||
id='cleanup_task'
|
||||
)
|
||||
scheduler.start()
|
||||
logger.info("Background cleanup scheduler started (runs every 6 hours)")
|
||||
|
||||
yield
|
||||
|
||||
logger.info("Shutting down Qwen3-TTS Backend Service...")
|
||||
|
||||
scheduler.shutdown()
|
||||
logger.info("Scheduler shutdown completed")
|
||||
|
||||
try:
|
||||
model_manager = await ModelManager.get_instance()
|
||||
await model_manager.unload_model()
|
||||
logger.info("Model cleanup completed")
|
||||
except Exception as e:
|
||||
logger.error(f"Model cleanup failed: {e}")
|
||||
|
||||
app = FastAPI(
|
||||
title="Qwen3-TTS-WebUI Backend API",
|
||||
description="Backend service for Qwen3-TTS-WebUI text-to-speech system",
|
||||
version="0.1.0",
|
||||
lifespan=lifespan
|
||||
)
|
||||
|
||||
app.state.limiter = limiter
|
||||
app.add_exception_handler(RateLimitExceeded, _rate_limit_exceeded_handler)
|
||||
|
||||
app.add_middleware(
|
||||
CORSMiddleware,
|
||||
allow_origins=["*"],
|
||||
allow_credentials=True,
|
||||
allow_methods=["*"],
|
||||
allow_headers=["*"],
|
||||
)
|
||||
|
||||
app.include_router(auth.router)
|
||||
app.include_router(jobs.router)
|
||||
app.include_router(tts.router)
|
||||
app.include_router(cache.router)
|
||||
app.include_router(metrics.router)
|
||||
app.include_router(users.router)
|
||||
|
||||
@app.get("/health")
|
||||
async def health_check():
|
||||
from core.batch_processor import BatchProcessor
|
||||
from core.database import SessionLocal
|
||||
|
||||
gpu_available = torch.cuda.is_available()
|
||||
|
||||
gpu_memory_used_mb = 0
|
||||
gpu_memory_total_mb = 0
|
||||
if gpu_available:
|
||||
gpu_memory_used_mb = torch.cuda.memory_allocated(0) / 1024**2
|
||||
gpu_memory_total_mb = torch.cuda.get_device_properties(0).total_memory / 1024**2
|
||||
|
||||
model_manager = await ModelManager.get_instance()
|
||||
current_model, _ = await model_manager.get_current_model()
|
||||
|
||||
batch_processor = await BatchProcessor.get_instance()
|
||||
queue_length = await batch_processor.get_queue_length()
|
||||
|
||||
database_connected = True
|
||||
try:
|
||||
db = SessionLocal()
|
||||
db.execute(text("SELECT 1"))
|
||||
db.close()
|
||||
except Exception as e:
|
||||
logger.error(f"Database health check failed: {e}")
|
||||
database_connected = False
|
||||
|
||||
cache_dir_writable = True
|
||||
try:
|
||||
test_file = Path(settings.CACHE_DIR) / ".health_check"
|
||||
test_file.write_text("test")
|
||||
test_file.unlink()
|
||||
except Exception as e:
|
||||
logger.error(f"Cache directory health check failed: {e}")
|
||||
cache_dir_writable = False
|
||||
|
||||
output_dir_writable = True
|
||||
try:
|
||||
test_file = Path(settings.OUTPUT_DIR) / ".health_check"
|
||||
test_file.write_text("test")
|
||||
test_file.unlink()
|
||||
except Exception as e:
|
||||
logger.error(f"Output directory health check failed: {e}")
|
||||
output_dir_writable = False
|
||||
|
||||
critical_issues = []
|
||||
if not database_connected:
|
||||
critical_issues.append("database_disconnected")
|
||||
if not cache_dir_writable:
|
||||
critical_issues.append("cache_dir_not_writable")
|
||||
if not output_dir_writable:
|
||||
critical_issues.append("output_dir_not_writable")
|
||||
|
||||
minor_issues = []
|
||||
if not gpu_available:
|
||||
minor_issues.append("gpu_not_available")
|
||||
if queue_length > 50:
|
||||
minor_issues.append("queue_congested")
|
||||
|
||||
if critical_issues:
|
||||
status = "unhealthy"
|
||||
elif minor_issues:
|
||||
status = "degraded"
|
||||
else:
|
||||
status = "healthy"
|
||||
|
||||
return {
|
||||
"status": status,
|
||||
"gpu_available": gpu_available,
|
||||
"gpu_memory_used_mb": gpu_memory_used_mb,
|
||||
"gpu_memory_total_mb": gpu_memory_total_mb,
|
||||
"queue_length": queue_length,
|
||||
"active_model": current_model,
|
||||
"database_connected": database_connected,
|
||||
"cache_dir_writable": cache_dir_writable,
|
||||
"output_dir_writable": output_dir_writable,
|
||||
"issues": {
|
||||
"critical": critical_issues,
|
||||
"minor": minor_issues
|
||||
}
|
||||
}
|
||||
|
||||
if __name__ == "__main__":
|
||||
import uvicorn
|
||||
uvicorn.run(
|
||||
"main:app",
|
||||
host=settings.HOST,
|
||||
port=settings.PORT,
|
||||
workers=settings.WORKERS,
|
||||
log_level=settings.LOG_LEVEL.lower()
|
||||
)
|
||||
Reference in New Issue
Block a user