fix:【ai 大模型】硅基流动 thinking 返回
This commit is contained in:
@@ -122,12 +122,12 @@ public class AiAutoConfiguration {
|
|||||||
if (StrUtil.isEmpty(properties.getModel())) {
|
if (StrUtil.isEmpty(properties.getModel())) {
|
||||||
properties.setModel(SiliconFlowApiConstants.MODEL_DEFAULT);
|
properties.setModel(SiliconFlowApiConstants.MODEL_DEFAULT);
|
||||||
}
|
}
|
||||||
OpenAiChatModel openAiChatModel = OpenAiChatModel.builder()
|
DeepSeekChatModel openAiChatModel = DeepSeekChatModel.builder()
|
||||||
.openAiApi(OpenAiApi.builder()
|
.deepSeekApi(DeepSeekApi.builder()
|
||||||
.baseUrl(SiliconFlowApiConstants.DEFAULT_BASE_URL)
|
.baseUrl(SiliconFlowApiConstants.DEFAULT_BASE_URL)
|
||||||
.apiKey(properties.getApiKey())
|
.apiKey(properties.getApiKey())
|
||||||
.build())
|
.build())
|
||||||
.defaultOptions(OpenAiChatOptions.builder()
|
.defaultOptions(DeepSeekChatOptions.builder()
|
||||||
.model(properties.getModel())
|
.model(properties.getModel())
|
||||||
.temperature(properties.getTemperature())
|
.temperature(properties.getTemperature())
|
||||||
.maxTokens(properties.getMaxTokens())
|
.maxTokens(properties.getMaxTokens())
|
||||||
|
|||||||
@@ -23,7 +23,7 @@ public class SiliconFlowChatModel implements ChatModel {
|
|||||||
/**
|
/**
|
||||||
* 兼容 OpenAI 接口,进行复用
|
* 兼容 OpenAI 接口,进行复用
|
||||||
*/
|
*/
|
||||||
private final OpenAiChatModel openAiChatModel;
|
private final ChatModel openAiChatModel;
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ChatResponse call(Prompt prompt) {
|
public ChatResponse call(Prompt prompt) {
|
||||||
|
|||||||
@@ -50,9 +50,10 @@ public class AiUtils {
|
|||||||
return QianFanChatOptions.builder().model(model).temperature(temperature).maxTokens(maxTokens).build();
|
return QianFanChatOptions.builder().model(model).temperature(temperature).maxTokens(maxTokens).build();
|
||||||
case DEEP_SEEK:
|
case DEEP_SEEK:
|
||||||
case DOU_BAO: // 复用 DeepSeek 客户端
|
case DOU_BAO: // 复用 DeepSeek 客户端
|
||||||
|
case HUN_YUAN: // 复用 DeepSeek 客户端
|
||||||
|
case SILICON_FLOW: // 复用 OpenAI 客户端
|
||||||
return DeepSeekChatOptions.builder().model(model).temperature(temperature).maxTokens(maxTokens)
|
return DeepSeekChatOptions.builder().model(model).temperature(temperature).maxTokens(maxTokens)
|
||||||
.toolNames(toolNames).toolContext(toolContext).build();
|
.toolNames(toolNames).toolContext(toolContext).build();
|
||||||
case HUN_YUAN: // 复用 DeepSeek 客户端
|
|
||||||
case ZHI_PU:
|
case ZHI_PU:
|
||||||
return ZhiPuAiChatOptions.builder().model(model).temperature(temperature).maxTokens(maxTokens)
|
return ZhiPuAiChatOptions.builder().model(model).temperature(temperature).maxTokens(maxTokens)
|
||||||
.toolNames(toolNames).toolContext(toolContext).build();
|
.toolNames(toolNames).toolContext(toolContext).build();
|
||||||
@@ -65,7 +66,6 @@ public class AiUtils {
|
|||||||
case OPENAI:
|
case OPENAI:
|
||||||
case GEMINI: // 复用 OpenAI 客户端
|
case GEMINI: // 复用 OpenAI 客户端
|
||||||
case XING_HUO: // 复用 OpenAI 客户端
|
case XING_HUO: // 复用 OpenAI 客户端
|
||||||
case SILICON_FLOW: // 复用 OpenAI 客户端
|
|
||||||
case BAI_CHUAN: // 复用 OpenAI 客户端
|
case BAI_CHUAN: // 复用 OpenAI 客户端
|
||||||
return OpenAiChatOptions.builder().model(model).temperature(temperature).maxTokens(maxTokens)
|
return OpenAiChatOptions.builder().model(model).temperature(temperature).maxTokens(maxTokens)
|
||||||
.toolNames(toolNames).toolContext(toolContext).build();
|
.toolNames(toolNames).toolContext(toolContext).build();
|
||||||
|
|||||||
@@ -9,9 +9,9 @@ import org.springframework.ai.chat.messages.SystemMessage;
|
|||||||
import org.springframework.ai.chat.messages.UserMessage;
|
import org.springframework.ai.chat.messages.UserMessage;
|
||||||
import org.springframework.ai.chat.model.ChatResponse;
|
import org.springframework.ai.chat.model.ChatResponse;
|
||||||
import org.springframework.ai.chat.prompt.Prompt;
|
import org.springframework.ai.chat.prompt.Prompt;
|
||||||
import org.springframework.ai.openai.OpenAiChatModel;
|
import org.springframework.ai.deepseek.DeepSeekChatModel;
|
||||||
import org.springframework.ai.openai.OpenAiChatOptions;
|
import org.springframework.ai.deepseek.DeepSeekChatOptions;
|
||||||
import org.springframework.ai.openai.api.OpenAiApi;
|
import org.springframework.ai.deepseek.api.DeepSeekApi;
|
||||||
import reactor.core.publisher.Flux;
|
import reactor.core.publisher.Flux;
|
||||||
|
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
@@ -24,12 +24,12 @@ import java.util.List;
|
|||||||
*/
|
*/
|
||||||
public class SiliconFlowChatModelTests {
|
public class SiliconFlowChatModelTests {
|
||||||
|
|
||||||
private final OpenAiChatModel openAiChatModel = OpenAiChatModel.builder()
|
private final DeepSeekChatModel openAiChatModel = DeepSeekChatModel.builder()
|
||||||
.openAiApi(OpenAiApi.builder()
|
.deepSeekApi(DeepSeekApi.builder()
|
||||||
.baseUrl(SiliconFlowApiConstants.DEFAULT_BASE_URL)
|
.baseUrl(SiliconFlowApiConstants.DEFAULT_BASE_URL)
|
||||||
.apiKey("sk-epsakfenqnyzoxhmbucsxlhkdqlcbnimslqoivkshalvdozz") // apiKey
|
.apiKey("sk-epsakfenqnyzoxhmbucsxlhkdqlcbnimslqoivkshalvdozz") // apiKey
|
||||||
.build())
|
.build())
|
||||||
.defaultOptions(OpenAiChatOptions.builder()
|
.defaultOptions(DeepSeekChatOptions.builder()
|
||||||
.model(SiliconFlowApiConstants.MODEL_DEFAULT) // 模型
|
.model(SiliconFlowApiConstants.MODEL_DEFAULT) // 模型
|
||||||
// .model("deepseek-ai/DeepSeek-R1") // 模型(deepseek-ai/DeepSeek-R1)可用赠费
|
// .model("deepseek-ai/DeepSeek-R1") // 模型(deepseek-ai/DeepSeek-R1)可用赠费
|
||||||
// .model("Pro/deepseek-ai/DeepSeek-R1") // 模型(Pro/deepseek-ai/DeepSeek-R1)需要付费
|
// .model("Pro/deepseek-ai/DeepSeek-R1") // 模型(Pro/deepseek-ai/DeepSeek-R1)需要付费
|
||||||
@@ -67,4 +67,23 @@ public class SiliconFlowChatModelTests {
|
|||||||
flux.doOnNext(System.out::println).then().block();
|
flux.doOnNext(System.out::println).then().block();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
@Disabled
|
||||||
|
public void testStream_thinking() {
|
||||||
|
// 准备参数
|
||||||
|
List<Message> messages = new ArrayList<>();
|
||||||
|
messages.add(new UserMessage("详细分析下,如何设计一个电商系统?"));
|
||||||
|
DeepSeekChatOptions options = DeepSeekChatOptions.builder()
|
||||||
|
.model("deepseek-ai/DeepSeek-R1")
|
||||||
|
.build();
|
||||||
|
|
||||||
|
// 调用
|
||||||
|
Flux<ChatResponse> flux = chatModel.stream(new Prompt(messages, options));
|
||||||
|
// 打印结果
|
||||||
|
flux.doOnNext(response -> {
|
||||||
|
// System.out.println(response);
|
||||||
|
System.out.println(response.getResult().getOutput());
|
||||||
|
}).then().block();
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user