273 lines
8.6 KiB
Python
273 lines
8.6 KiB
Python
import logging
|
|
import sys
|
|
from contextlib import asynccontextmanager
|
|
from pathlib import Path
|
|
|
|
import torch
|
|
from fastapi import FastAPI, Request, Depends, HTTPException
|
|
from fastapi.middleware.cors import CORSMiddleware
|
|
from slowapi import Limiter, _rate_limit_exceeded_handler
|
|
from slowapi.util import get_remote_address
|
|
from slowapi.errors import RateLimitExceeded
|
|
from sqlalchemy import text
|
|
|
|
from core.config import settings
|
|
from core.database import init_db
|
|
from core.model_manager import ModelManager
|
|
from core.cleanup import run_scheduled_cleanup
|
|
from api import auth, jobs, tts, users, voice_designs, audiobook, admin
|
|
from api.auth import get_current_user
|
|
from schemas.user import User
|
|
from apscheduler.schedulers.asyncio import AsyncIOScheduler
|
|
|
|
logging.basicConfig(
|
|
level=getattr(logging, settings.LOG_LEVEL.upper()),
|
|
format='%(asctime)s - %(name)s - %(levelname)s - %(message)s',
|
|
handlers=[
|
|
logging.StreamHandler(sys.stdout)
|
|
]
|
|
)
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
def get_user_identifier(request: Request) -> str:
|
|
from jose import jwt
|
|
from core.config import settings
|
|
|
|
auth_header = request.headers.get("Authorization", "")
|
|
if auth_header.startswith("Bearer "):
|
|
token = auth_header.split(" ")[1]
|
|
try:
|
|
payload = jwt.decode(token, settings.SECRET_KEY, algorithms=[settings.ALGORITHM])
|
|
user_id = payload.get("sub")
|
|
if user_id:
|
|
return f"user:{user_id}"
|
|
except Exception:
|
|
pass
|
|
|
|
return get_remote_address(request)
|
|
|
|
limiter = Limiter(key_func=get_user_identifier)
|
|
|
|
@asynccontextmanager
|
|
async def lifespan(app: FastAPI):
|
|
logger.info("Starting Qwen3-TTS Backend Service...")
|
|
logger.info(f"Model base path: {settings.MODEL_BASE_PATH}")
|
|
logger.info(f"Cache directory: {settings.CACHE_DIR}")
|
|
logger.info(f"Output directory: {settings.OUTPUT_DIR}")
|
|
logger.info(f"Device: {settings.MODEL_DEVICE}")
|
|
|
|
try:
|
|
settings.validate()
|
|
logger.info("Configuration validated successfully")
|
|
except Exception as e:
|
|
logger.error(f"Configuration validation failed: {e}")
|
|
raise
|
|
|
|
try:
|
|
init_db()
|
|
logger.info("Database initialized successfully")
|
|
|
|
# Reset stale processing statuses from interrupted sessions
|
|
from core.database import SessionLocal
|
|
from db.models import AudiobookChapter, AudiobookSegment
|
|
startup_db = SessionLocal()
|
|
try:
|
|
stale_chapters = startup_db.query(AudiobookChapter).filter(AudiobookChapter.status == "parsing").all()
|
|
for ch in stale_chapters:
|
|
ch.status = "pending"
|
|
stale_segments = startup_db.query(AudiobookSegment).filter(AudiobookSegment.status == "generating").all()
|
|
for seg in stale_segments:
|
|
seg.status = "pending"
|
|
if stale_chapters or stale_segments:
|
|
startup_db.commit()
|
|
logger.info(f"Reset {len(stale_chapters)} stale parsing chapters, {len(stale_segments)} stale generating segments")
|
|
finally:
|
|
startup_db.close()
|
|
except Exception as e:
|
|
logger.error(f"Database initialization failed: {e}")
|
|
raise
|
|
|
|
try:
|
|
from core.init_admin import init_superuser
|
|
init_superuser()
|
|
except Exception as e:
|
|
logger.error(f"Superuser initialization failed: {e}")
|
|
raise
|
|
|
|
try:
|
|
from core.model_manager import IndexTTS2ModelManager
|
|
indextts2_manager = await IndexTTS2ModelManager.get_instance()
|
|
await indextts2_manager.get_model()
|
|
logger.info("Preloaded IndexTTS2 model")
|
|
except Exception as e:
|
|
logger.warning(f"IndexTTS2 model preload failed: {e}")
|
|
|
|
scheduler = AsyncIOScheduler()
|
|
scheduler.add_job(
|
|
run_scheduled_cleanup,
|
|
'interval',
|
|
hours=6,
|
|
args=[str(settings.DATABASE_URL)],
|
|
id='cleanup_task'
|
|
)
|
|
scheduler.start()
|
|
logger.info("Background cleanup scheduler started (runs every 6 hours)")
|
|
|
|
yield
|
|
|
|
logger.info("Shutting down Qwen3-TTS Backend Service...")
|
|
|
|
scheduler.shutdown()
|
|
logger.info("Scheduler shutdown completed")
|
|
|
|
try:
|
|
model_manager = await ModelManager.get_instance()
|
|
await model_manager.unload_model()
|
|
logger.info("Model cleanup completed")
|
|
except Exception as e:
|
|
logger.error(f"Model cleanup failed: {e}")
|
|
|
|
app = FastAPI(
|
|
title="Qwen3-TTS-WebUI Backend API",
|
|
description="Backend service for Qwen3-TTS-WebUI text-to-speech system",
|
|
version="0.1.0",
|
|
lifespan=lifespan
|
|
)
|
|
|
|
app.state.limiter = limiter
|
|
app.add_exception_handler(RateLimitExceeded, _rate_limit_exceeded_handler)
|
|
|
|
app.add_middleware(
|
|
CORSMiddleware,
|
|
allow_origins=["http://localhost:5173", "http://127.0.0.1:5173"],
|
|
allow_credentials=True,
|
|
allow_methods=["*"],
|
|
allow_headers=["*"],
|
|
)
|
|
|
|
app.include_router(auth.router)
|
|
app.include_router(jobs.router)
|
|
app.include_router(tts.router)
|
|
app.include_router(users.router)
|
|
app.include_router(voice_designs.router)
|
|
app.include_router(audiobook.router)
|
|
app.include_router(admin.router)
|
|
|
|
@app.get("/health")
|
|
async def health_check():
|
|
return {"status": "ok"}
|
|
|
|
|
|
@app.get("/health/details")
|
|
async def health_check_details(current_user: User = Depends(get_current_user)):
|
|
if not current_user.is_superuser:
|
|
raise HTTPException(status_code=403, detail="Superuser access required")
|
|
|
|
from core.batch_processor import BatchProcessor
|
|
from core.database import SessionLocal
|
|
|
|
gpu_available = torch.cuda.is_available()
|
|
|
|
gpu_memory_used_mb = 0
|
|
gpu_memory_total_mb = 0
|
|
if gpu_available:
|
|
gpu_memory_used_mb = torch.cuda.memory_allocated(0) / 1024**2
|
|
gpu_memory_total_mb = torch.cuda.get_device_properties(0).total_memory / 1024**2
|
|
|
|
model_manager = await ModelManager.get_instance()
|
|
current_model, _ = await model_manager.get_current_model()
|
|
|
|
batch_processor = await BatchProcessor.get_instance()
|
|
queue_length = await batch_processor.get_queue_length()
|
|
|
|
database_connected = True
|
|
try:
|
|
db = SessionLocal()
|
|
db.execute(text("SELECT 1"))
|
|
db.close()
|
|
except Exception as e:
|
|
logger.error(f"Database health check failed: {e}")
|
|
database_connected = False
|
|
|
|
cache_dir_writable = True
|
|
try:
|
|
test_file = Path(settings.CACHE_DIR) / ".health_check"
|
|
test_file.write_text("test")
|
|
test_file.unlink()
|
|
except Exception as e:
|
|
logger.error(f"Cache directory health check failed: {e}")
|
|
cache_dir_writable = False
|
|
|
|
output_dir_writable = True
|
|
try:
|
|
test_file = Path(settings.OUTPUT_DIR) / ".health_check"
|
|
test_file.write_text("test")
|
|
test_file.unlink()
|
|
except Exception as e:
|
|
logger.error(f"Output directory health check failed: {e}")
|
|
output_dir_writable = False
|
|
|
|
critical_issues = []
|
|
if not database_connected:
|
|
critical_issues.append("database_disconnected")
|
|
if not cache_dir_writable:
|
|
critical_issues.append("cache_dir_not_writable")
|
|
if not output_dir_writable:
|
|
critical_issues.append("output_dir_not_writable")
|
|
|
|
minor_issues = []
|
|
if not gpu_available:
|
|
minor_issues.append("gpu_not_available")
|
|
if queue_length > 50:
|
|
minor_issues.append("queue_congested")
|
|
|
|
backends_status = {}
|
|
|
|
try:
|
|
from core.tts_service import TTSServiceFactory
|
|
|
|
try:
|
|
local_backend = await TTSServiceFactory.get_backend("local")
|
|
local_health = await local_backend.health_check()
|
|
backends_status["local"] = local_health
|
|
except Exception as e:
|
|
backends_status["local"] = {"available": False, "error": str(e)}
|
|
except Exception as e:
|
|
logger.error(f"Backend health check failed: {e}")
|
|
backends_status = {"error": str(e)}
|
|
|
|
if critical_issues:
|
|
status = "unhealthy"
|
|
elif minor_issues:
|
|
status = "degraded"
|
|
else:
|
|
status = "healthy"
|
|
|
|
return {
|
|
"status": status,
|
|
"gpu_available": gpu_available,
|
|
"gpu_memory_used_mb": gpu_memory_used_mb,
|
|
"gpu_memory_total_mb": gpu_memory_total_mb,
|
|
"queue_length": queue_length,
|
|
"active_model": current_model,
|
|
"database_connected": database_connected,
|
|
"cache_dir_writable": cache_dir_writable,
|
|
"output_dir_writable": output_dir_writable,
|
|
"backends": backends_status,
|
|
"issues": {
|
|
"critical": critical_issues,
|
|
"minor": minor_issues
|
|
}
|
|
}
|
|
|
|
if __name__ == "__main__":
|
|
import uvicorn
|
|
uvicorn.run(
|
|
"main:app",
|
|
host=settings.HOST,
|
|
port=settings.PORT,
|
|
workers=settings.WORKERS,
|
|
log_level=settings.LOG_LEVEL.lower()
|
|
)
|