fix: Adjust chunk size in parse_one_chapter to 1500 and add enable_thinking parameter to LLMService methods
This commit is contained in:
@@ -29,6 +29,7 @@ class LLMService:
|
||||
"temperature": 0.3,
|
||||
"max_tokens": max_tokens,
|
||||
"stream": True,
|
||||
"enable_thinking": False,
|
||||
}
|
||||
full_text = ""
|
||||
timeout = httpx.Timeout(connect=10.0, read=90.0, write=10.0, pool=5.0)
|
||||
@@ -87,6 +88,8 @@ class LLMService:
|
||||
{"role": "user", "content": user_message},
|
||||
],
|
||||
"temperature": 0.3,
|
||||
"max_tokens": 8192,
|
||||
"enable_thinking": False,
|
||||
}
|
||||
|
||||
timeout = httpx.Timeout(connect=10.0, read=90.0, write=10.0, pool=5.0)
|
||||
|
||||
Reference in New Issue
Block a user