refactor: rename canto-backend → backend, canto-frontend → frontend
Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
This commit is contained in:
272
backend/main.py
Normal file
272
backend/main.py
Normal file
@@ -0,0 +1,272 @@
|
||||
import logging
|
||||
import sys
|
||||
from contextlib import asynccontextmanager
|
||||
from pathlib import Path
|
||||
|
||||
import torch
|
||||
from fastapi import FastAPI, Request, Depends, HTTPException
|
||||
from fastapi.middleware.cors import CORSMiddleware
|
||||
from slowapi import Limiter, _rate_limit_exceeded_handler
|
||||
from slowapi.util import get_remote_address
|
||||
from slowapi.errors import RateLimitExceeded
|
||||
from sqlalchemy import text
|
||||
|
||||
from core.config import settings
|
||||
from core.database import init_db
|
||||
from core.model_manager import ModelManager
|
||||
from core.cleanup import run_scheduled_cleanup
|
||||
from api import auth, jobs, tts, users, voice_designs, audiobook, admin
|
||||
from api.auth import get_current_user
|
||||
from schemas.user import User
|
||||
from apscheduler.schedulers.asyncio import AsyncIOScheduler
|
||||
|
||||
logging.basicConfig(
|
||||
level=getattr(logging, settings.LOG_LEVEL.upper()),
|
||||
format='%(asctime)s - %(name)s - %(levelname)s - %(message)s',
|
||||
handlers=[
|
||||
logging.StreamHandler(sys.stdout)
|
||||
]
|
||||
)
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
def get_user_identifier(request: Request) -> str:
|
||||
from jose import jwt
|
||||
from core.config import settings
|
||||
|
||||
auth_header = request.headers.get("Authorization", "")
|
||||
if auth_header.startswith("Bearer "):
|
||||
token = auth_header.split(" ")[1]
|
||||
try:
|
||||
payload = jwt.decode(token, settings.SECRET_KEY, algorithms=[settings.ALGORITHM])
|
||||
user_id = payload.get("sub")
|
||||
if user_id:
|
||||
return f"user:{user_id}"
|
||||
except Exception:
|
||||
pass
|
||||
|
||||
return get_remote_address(request)
|
||||
|
||||
limiter = Limiter(key_func=get_user_identifier)
|
||||
|
||||
@asynccontextmanager
|
||||
async def lifespan(app: FastAPI):
|
||||
logger.info("Starting Canto Backend Service...")
|
||||
logger.info(f"Model base path: {settings.MODEL_BASE_PATH}")
|
||||
logger.info(f"Cache directory: {settings.CACHE_DIR}")
|
||||
logger.info(f"Output directory: {settings.OUTPUT_DIR}")
|
||||
logger.info(f"Device: {settings.MODEL_DEVICE}")
|
||||
|
||||
try:
|
||||
settings.validate()
|
||||
logger.info("Configuration validated successfully")
|
||||
except Exception as e:
|
||||
logger.error(f"Configuration validation failed: {e}")
|
||||
raise
|
||||
|
||||
try:
|
||||
init_db()
|
||||
logger.info("Database initialized successfully")
|
||||
|
||||
# Reset stale processing statuses from interrupted sessions
|
||||
from core.database import SessionLocal
|
||||
from db.models import AudiobookChapter, AudiobookSegment
|
||||
startup_db = SessionLocal()
|
||||
try:
|
||||
stale_chapters = startup_db.query(AudiobookChapter).filter(AudiobookChapter.status == "parsing").all()
|
||||
for ch in stale_chapters:
|
||||
ch.status = "pending"
|
||||
stale_segments = startup_db.query(AudiobookSegment).filter(AudiobookSegment.status == "generating").all()
|
||||
for seg in stale_segments:
|
||||
seg.status = "pending"
|
||||
if stale_chapters or stale_segments:
|
||||
startup_db.commit()
|
||||
logger.info(f"Reset {len(stale_chapters)} stale parsing chapters, {len(stale_segments)} stale generating segments")
|
||||
finally:
|
||||
startup_db.close()
|
||||
except Exception as e:
|
||||
logger.error(f"Database initialization failed: {e}")
|
||||
raise
|
||||
|
||||
try:
|
||||
from core.init_admin import init_superuser
|
||||
init_superuser()
|
||||
except Exception as e:
|
||||
logger.error(f"Superuser initialization failed: {e}")
|
||||
raise
|
||||
|
||||
try:
|
||||
from core.model_manager import IndexTTS2ModelManager
|
||||
indextts2_manager = await IndexTTS2ModelManager.get_instance()
|
||||
await indextts2_manager.get_model()
|
||||
logger.info("Preloaded IndexTTS2 model")
|
||||
except Exception as e:
|
||||
logger.warning(f"IndexTTS2 model preload failed: {e}")
|
||||
|
||||
scheduler = AsyncIOScheduler()
|
||||
scheduler.add_job(
|
||||
run_scheduled_cleanup,
|
||||
'interval',
|
||||
hours=6,
|
||||
args=[str(settings.DATABASE_URL)],
|
||||
id='cleanup_task'
|
||||
)
|
||||
scheduler.start()
|
||||
logger.info("Background cleanup scheduler started (runs every 6 hours)")
|
||||
|
||||
yield
|
||||
|
||||
logger.info("Shutting down Canto Backend Service...")
|
||||
|
||||
scheduler.shutdown()
|
||||
logger.info("Scheduler shutdown completed")
|
||||
|
||||
try:
|
||||
model_manager = await ModelManager.get_instance()
|
||||
await model_manager.unload_model()
|
||||
logger.info("Model cleanup completed")
|
||||
except Exception as e:
|
||||
logger.error(f"Model cleanup failed: {e}")
|
||||
|
||||
app = FastAPI(
|
||||
title="Canto Backend API",
|
||||
description="Backend service for Canto text-to-speech system",
|
||||
version="0.1.0",
|
||||
lifespan=lifespan
|
||||
)
|
||||
|
||||
app.state.limiter = limiter
|
||||
app.add_exception_handler(RateLimitExceeded, _rate_limit_exceeded_handler)
|
||||
|
||||
app.add_middleware(
|
||||
CORSMiddleware,
|
||||
allow_origins=["http://localhost:5173", "http://127.0.0.1:5173"],
|
||||
allow_credentials=True,
|
||||
allow_methods=["*"],
|
||||
allow_headers=["*"],
|
||||
)
|
||||
|
||||
app.include_router(auth.router)
|
||||
app.include_router(jobs.router)
|
||||
app.include_router(tts.router)
|
||||
app.include_router(users.router)
|
||||
app.include_router(voice_designs.router)
|
||||
app.include_router(audiobook.router)
|
||||
app.include_router(admin.router)
|
||||
|
||||
@app.get("/health")
|
||||
async def health_check():
|
||||
return {"status": "ok"}
|
||||
|
||||
|
||||
@app.get("/health/details")
|
||||
async def health_check_details(current_user: User = Depends(get_current_user)):
|
||||
if not current_user.is_superuser:
|
||||
raise HTTPException(status_code=403, detail="Superuser access required")
|
||||
|
||||
from core.batch_processor import BatchProcessor
|
||||
from core.database import SessionLocal
|
||||
|
||||
gpu_available = torch.cuda.is_available()
|
||||
|
||||
gpu_memory_used_mb = 0
|
||||
gpu_memory_total_mb = 0
|
||||
if gpu_available:
|
||||
gpu_memory_used_mb = torch.cuda.memory_allocated(0) / 1024**2
|
||||
gpu_memory_total_mb = torch.cuda.get_device_properties(0).total_memory / 1024**2
|
||||
|
||||
model_manager = await ModelManager.get_instance()
|
||||
current_model, _ = await model_manager.get_current_model()
|
||||
|
||||
batch_processor = await BatchProcessor.get_instance()
|
||||
queue_length = await batch_processor.get_queue_length()
|
||||
|
||||
database_connected = True
|
||||
try:
|
||||
db = SessionLocal()
|
||||
db.execute(text("SELECT 1"))
|
||||
db.close()
|
||||
except Exception as e:
|
||||
logger.error(f"Database health check failed: {e}")
|
||||
database_connected = False
|
||||
|
||||
cache_dir_writable = True
|
||||
try:
|
||||
test_file = Path(settings.CACHE_DIR) / ".health_check"
|
||||
test_file.write_text("test")
|
||||
test_file.unlink()
|
||||
except Exception as e:
|
||||
logger.error(f"Cache directory health check failed: {e}")
|
||||
cache_dir_writable = False
|
||||
|
||||
output_dir_writable = True
|
||||
try:
|
||||
test_file = Path(settings.OUTPUT_DIR) / ".health_check"
|
||||
test_file.write_text("test")
|
||||
test_file.unlink()
|
||||
except Exception as e:
|
||||
logger.error(f"Output directory health check failed: {e}")
|
||||
output_dir_writable = False
|
||||
|
||||
critical_issues = []
|
||||
if not database_connected:
|
||||
critical_issues.append("database_disconnected")
|
||||
if not cache_dir_writable:
|
||||
critical_issues.append("cache_dir_not_writable")
|
||||
if not output_dir_writable:
|
||||
critical_issues.append("output_dir_not_writable")
|
||||
|
||||
minor_issues = []
|
||||
if not gpu_available:
|
||||
minor_issues.append("gpu_not_available")
|
||||
if queue_length > 50:
|
||||
minor_issues.append("queue_congested")
|
||||
|
||||
backends_status = {}
|
||||
|
||||
try:
|
||||
from core.tts_service import TTSServiceFactory
|
||||
|
||||
try:
|
||||
local_backend = await TTSServiceFactory.get_backend("local")
|
||||
local_health = await local_backend.health_check()
|
||||
backends_status["local"] = local_health
|
||||
except Exception as e:
|
||||
backends_status["local"] = {"available": False, "error": str(e)}
|
||||
except Exception as e:
|
||||
logger.error(f"Backend health check failed: {e}")
|
||||
backends_status = {"error": str(e)}
|
||||
|
||||
if critical_issues:
|
||||
status = "unhealthy"
|
||||
elif minor_issues:
|
||||
status = "degraded"
|
||||
else:
|
||||
status = "healthy"
|
||||
|
||||
return {
|
||||
"status": status,
|
||||
"gpu_available": gpu_available,
|
||||
"gpu_memory_used_mb": gpu_memory_used_mb,
|
||||
"gpu_memory_total_mb": gpu_memory_total_mb,
|
||||
"queue_length": queue_length,
|
||||
"active_model": current_model,
|
||||
"database_connected": database_connected,
|
||||
"cache_dir_writable": cache_dir_writable,
|
||||
"output_dir_writable": output_dir_writable,
|
||||
"backends": backends_status,
|
||||
"issues": {
|
||||
"critical": critical_issues,
|
||||
"minor": minor_issues
|
||||
}
|
||||
}
|
||||
|
||||
if __name__ == "__main__":
|
||||
import uvicorn
|
||||
uvicorn.run(
|
||||
"main:app",
|
||||
host=settings.HOST,
|
||||
port=settings.PORT,
|
||||
workers=settings.WORKERS,
|
||||
log_level=settings.LOG_LEVEL.lower()
|
||||
)
|
||||
Reference in New Issue
Block a user