|
15 | 15 | from bisheng.common.errcode.server import NoLlmModelConfigError, LlmModelConfigDeletedError, LlmProviderDeletedError, \ |
16 | 16 | LlmModelTypeError, LlmModelOfflineError, InitLlmError |
17 | 17 | from bisheng.core.ai import ChatOllama, ChatOpenAI, ChatOpenAICompatible, \ |
18 | | - AzureChatOpenAI, ChatZhipuAI, MiniMaxChat, ChatAnthropic, MoonshotChat |
| 18 | + AzureChatOpenAI, ChatZhipuAI, ChatAnthropic, MoonshotChat |
19 | 19 | from bisheng.core.ai.llm.custom_chat_deepseek import CustomChatDeepSeek |
20 | 20 | from bisheng.core.ai.llm.custom_chat_tongyi import CustomChatTongYi |
21 | 21 | from bisheng.llm.domain.const import LLMModelType, LLMServerType |
@@ -126,19 +126,6 @@ def _get_qwen_params(params: dict, server_config: dict, model_config: dict) -> d |
126 | 126 | return user_kwargs |
127 | 127 |
|
128 | 128 |
|
129 | | -def _get_minimax_params(params: dict, server_config: dict, model_config: dict) -> dict: |
130 | | - params['minimax_api_key'] = server_config.get('openai_api_key') |
131 | | - params['base_url'] = server_config.get('openai_api_base').rstrip('/') |
132 | | - if 'max_tokens' not in params: |
133 | | - params['max_tokens'] = 2048 |
134 | | - if '/chat/completions' not in params['base_url']: |
135 | | - params['base_url'] = f"{params['base_url']}/chat/completions" |
136 | | - |
137 | | - user_kwargs = _get_user_kwargs(model_config) |
138 | | - user_kwargs.update(params) |
139 | | - return user_kwargs |
140 | | - |
141 | | - |
142 | 129 | def _get_anthropic_params(params: dict, server_config: dict, model_config: dict) -> dict: |
143 | 130 | params.update(server_config) |
144 | 131 |
|
@@ -182,7 +169,7 @@ def _get_spark_params(params: dict, server_config: dict, model_config: dict) -> |
182 | 169 | LLMServerType.QWEN.value: {'client': CustomChatTongYi, 'params_handler': _get_qwen_params}, |
183 | 170 | LLMServerType.QIAN_FAN.value: {'client': ChatOpenAICompatible, 'params_handler': _get_openai_params}, |
184 | 171 | LLMServerType.ZHIPU.value: {'client': ChatZhipuAI, 'params_handler': _get_zhipu_params}, |
185 | | - LLMServerType.MINIMAX.value: {'client': MiniMaxChat, 'params_handler': _get_minimax_params}, |
| 172 | + LLMServerType.MINIMAX.value: {'client': ChatOpenAICompatible, 'params_handler': _get_openai_params}, |
186 | 173 | LLMServerType.ANTHROPIC.value: {'client': ChatAnthropic, 'params_handler': _get_anthropic_params}, |
187 | 174 | LLMServerType.DEEPSEEK.value: {'client': CustomChatDeepSeek, 'params_handler': _get_deepseek_params}, |
188 | 175 | LLMServerType.SPARK.value: {'client': ChatOpenAICompatible, 'params_handler': _get_spark_params}, |
|
0 commit comments