1414
1515from bisheng .common .errcode .server import NoLlmModelConfigError , LlmModelConfigDeletedError , LlmProviderDeletedError , \
1616 LlmModelTypeError , LlmModelOfflineError , InitLlmError
17- from bisheng .core .ai import CustomChatOllamaWithReasoning , ChatOpenAI , ChatOpenAICompatible , \
17+ from bisheng .core .ai import ChatOllama , ChatOpenAI , ChatOpenAICompatible , \
1818 AzureChatOpenAI , ChatTongyi , ChatZhipuAI , MiniMaxChat , ChatAnthropic , ChatDeepSeek , \
1919 MoonshotChat
2020from bisheng .llm .domain .const import LLMModelType , LLMServerType
@@ -35,7 +35,6 @@ def _get_user_kwargs(model_config: dict) -> dict:
3535def _get_ollama_params (params : dict , server_config : dict , model_config : dict ) -> dict :
3636 params ['base_url' ] = server_config .get ('base_url' , '' ).rstrip ('/' )
3737 # some bugs
38- params ['extract_reasoning' ] = False
3938 params ['stream' ] = params .pop ('streaming' , True )
4039 if params .get ('max_tokens' ):
4140 params ['num_ctx' ] = params .pop ('max_tokens' , None )
@@ -155,7 +154,7 @@ def _get_spark_params(params: dict, server_config: dict, model_config: dict) ->
155154
156155_llm_node_type : Dict = {
157156 # Open source inference framework
158- LLMServerType .OLLAMA .value : {'client' : CustomChatOllamaWithReasoning , 'params_handler' : _get_ollama_params },
157+ LLMServerType .OLLAMA .value : {'client' : ChatOllama , 'params_handler' : _get_ollama_params },
159158 LLMServerType .XINFERENCE .value : {'client' : ChatOpenAI , 'params_handler' : _get_xinference_params },
160159 LLMServerType .LLAMACPP .value : {'client' : ChatOpenAI , 'params_handler' : _get_openai_params },
161160 LLMServerType .VLLM .value : {'client' : ChatOpenAICompatible , 'params_handler' : _get_openai_params },
0 commit comments