From 4f0d9f5ed6e6b0772878290122cf84458083f8b8 Mon Sep 17 00:00:00 2001 From: bdim404 Date: Wed, 11 Mar 2026 19:05:03 +0800 Subject: [PATCH 1/2] fix: Adjust chunk size in parse_one_chapter to 1500 and add enable_thinking parameter to LLMService methods --- qwen3-tts-backend/core/audiobook_service.py | 2 +- qwen3-tts-backend/core/llm_service.py | 3 +++ 2 files changed, 4 insertions(+), 1 deletion(-) diff --git a/qwen3-tts-backend/core/audiobook_service.py b/qwen3-tts-backend/core/audiobook_service.py index 25640e4..746b404 100644 --- a/qwen3-tts-backend/core/audiobook_service.py +++ b/qwen3-tts-backend/core/audiobook_service.py @@ -361,7 +361,7 @@ async def parse_one_chapter(project_id: int, chapter_id: int, user: User, db) -> crud.delete_audiobook_segments_for_chapter(db, project_id, chapter.chapter_index) - chunks = _chunk_chapter(chapter.source_text, max_chars=4000) + chunks = _chunk_chapter(chapter.source_text, max_chars=1500) ps.append_line(key, f"共 {len(chunks)} 块\n") seg_counter = 0 diff --git a/qwen3-tts-backend/core/llm_service.py b/qwen3-tts-backend/core/llm_service.py index 78287a1..968271e 100644 --- a/qwen3-tts-backend/core/llm_service.py +++ b/qwen3-tts-backend/core/llm_service.py @@ -29,6 +29,7 @@ class LLMService: "temperature": 0.3, "max_tokens": max_tokens, "stream": True, + "enable_thinking": False, } full_text = "" timeout = httpx.Timeout(connect=10.0, read=90.0, write=10.0, pool=5.0) @@ -87,6 +88,8 @@ class LLMService: {"role": "user", "content": user_message}, ], "temperature": 0.3, + "max_tokens": 8192, + "enable_thinking": False, } timeout = httpx.Timeout(connect=10.0, read=90.0, write=10.0, pool=5.0) From 29bd45e0e0bf8096cbf0345bba0c2e58e6245518 Mon Sep 17 00:00:00 2001 From: bdim404 Date: Wed, 11 Mar 2026 19:09:17 +0800 Subject: [PATCH 2/2] fix: Remove enable_thinking parameter from stream_chat methods --- qwen3-tts-backend/core/llm_service.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/qwen3-tts-backend/core/llm_service.py b/qwen3-tts-backend/core/llm_service.py index 968271e..9908104 100644 --- a/qwen3-tts-backend/core/llm_service.py +++ b/qwen3-tts-backend/core/llm_service.py @@ -29,7 +29,6 @@ class LLMService: "temperature": 0.3, "max_tokens": max_tokens, "stream": True, - "enable_thinking": False, } full_text = "" timeout = httpx.Timeout(connect=10.0, read=90.0, write=10.0, pool=5.0) @@ -89,7 +88,6 @@ class LLMService: ], "temperature": 0.3, "max_tokens": 8192, - "enable_thinking": False, } timeout = httpx.Timeout(connect=10.0, read=90.0, write=10.0, pool=5.0)