AI rule node: support top N for all models
This commit is contained in:
parent
3f2a6440a7
commit
3f58ff01c3
@ -56,6 +56,7 @@ class Langchain4jChatModelConfigurerImpl implements Langchain4jChatModelConfigur
|
|||||||
.apiKey(chatModel.providerConfig().apiKey())
|
.apiKey(chatModel.providerConfig().apiKey())
|
||||||
.modelName(modelConfig.modelId())
|
.modelName(modelConfig.modelId())
|
||||||
.temperature(modelConfig.temperature())
|
.temperature(modelConfig.temperature())
|
||||||
|
.topP(modelConfig.topP())
|
||||||
.timeout(toDuration(modelConfig.timeoutSeconds()))
|
.timeout(toDuration(modelConfig.timeoutSeconds()))
|
||||||
.maxRetries(modelConfig.maxRetries())
|
.maxRetries(modelConfig.maxRetries())
|
||||||
.build();
|
.build();
|
||||||
@ -68,6 +69,7 @@ class Langchain4jChatModelConfigurerImpl implements Langchain4jChatModelConfigur
|
|||||||
.apiKey(chatModel.providerConfig().apiKey())
|
.apiKey(chatModel.providerConfig().apiKey())
|
||||||
.deploymentName(modelConfig.modelId())
|
.deploymentName(modelConfig.modelId())
|
||||||
.temperature(modelConfig.temperature())
|
.temperature(modelConfig.temperature())
|
||||||
|
.topP(modelConfig.topP())
|
||||||
.timeout(toDuration(modelConfig.timeoutSeconds()))
|
.timeout(toDuration(modelConfig.timeoutSeconds()))
|
||||||
.maxRetries(modelConfig.maxRetries())
|
.maxRetries(modelConfig.maxRetries())
|
||||||
.build();
|
.build();
|
||||||
@ -80,6 +82,7 @@ class Langchain4jChatModelConfigurerImpl implements Langchain4jChatModelConfigur
|
|||||||
.apiKey(chatModel.providerConfig().apiKey())
|
.apiKey(chatModel.providerConfig().apiKey())
|
||||||
.modelName(modelConfig.modelId())
|
.modelName(modelConfig.modelId())
|
||||||
.temperature(modelConfig.temperature())
|
.temperature(modelConfig.temperature())
|
||||||
|
.topP(modelConfig.topP())
|
||||||
.timeout(toDuration(modelConfig.timeoutSeconds()))
|
.timeout(toDuration(modelConfig.timeoutSeconds()))
|
||||||
.maxRetries(modelConfig.maxRetries())
|
.maxRetries(modelConfig.maxRetries())
|
||||||
.build();
|
.build();
|
||||||
@ -112,6 +115,9 @@ class Langchain4jChatModelConfigurerImpl implements Langchain4jChatModelConfigur
|
|||||||
if (modelConfig.temperature() != null) {
|
if (modelConfig.temperature() != null) {
|
||||||
generationConfigBuilder.setTemperature(modelConfig.temperature().floatValue());
|
generationConfigBuilder.setTemperature(modelConfig.temperature().floatValue());
|
||||||
}
|
}
|
||||||
|
if (modelConfig.topP() != null) {
|
||||||
|
generationConfigBuilder.setTopP(modelConfig.topP().floatValue());
|
||||||
|
}
|
||||||
var generationConfig = generationConfigBuilder.build();
|
var generationConfig = generationConfigBuilder.build();
|
||||||
|
|
||||||
// construct generative model instance
|
// construct generative model instance
|
||||||
@ -128,6 +134,7 @@ class Langchain4jChatModelConfigurerImpl implements Langchain4jChatModelConfigur
|
|||||||
.apiKey(chatModel.providerConfig().apiKey())
|
.apiKey(chatModel.providerConfig().apiKey())
|
||||||
.modelName(modelConfig.modelId())
|
.modelName(modelConfig.modelId())
|
||||||
.temperature(modelConfig.temperature())
|
.temperature(modelConfig.temperature())
|
||||||
|
.topP(modelConfig.topP())
|
||||||
.timeout(toDuration(modelConfig.timeoutSeconds()))
|
.timeout(toDuration(modelConfig.timeoutSeconds()))
|
||||||
.maxRetries(modelConfig.maxRetries())
|
.maxRetries(modelConfig.maxRetries())
|
||||||
.build();
|
.build();
|
||||||
@ -140,6 +147,7 @@ class Langchain4jChatModelConfigurerImpl implements Langchain4jChatModelConfigur
|
|||||||
.apiKey(chatModel.providerConfig().apiKey())
|
.apiKey(chatModel.providerConfig().apiKey())
|
||||||
.modelName(modelConfig.modelId())
|
.modelName(modelConfig.modelId())
|
||||||
.temperature(modelConfig.temperature())
|
.temperature(modelConfig.temperature())
|
||||||
|
.topP(modelConfig.topP())
|
||||||
.timeout(toDuration(modelConfig.timeoutSeconds()))
|
.timeout(toDuration(modelConfig.timeoutSeconds()))
|
||||||
.maxRetries(modelConfig.maxRetries())
|
.maxRetries(modelConfig.maxRetries())
|
||||||
.build();
|
.build();
|
||||||
@ -161,6 +169,7 @@ class Langchain4jChatModelConfigurerImpl implements Langchain4jChatModelConfigur
|
|||||||
|
|
||||||
var defaultChatRequestParams = ChatRequestParameters.builder()
|
var defaultChatRequestParams = ChatRequestParameters.builder()
|
||||||
.temperature(modelConfig.temperature())
|
.temperature(modelConfig.temperature())
|
||||||
|
.topP(modelConfig.topP())
|
||||||
.build();
|
.build();
|
||||||
|
|
||||||
return BedrockChatModel.builder()
|
return BedrockChatModel.builder()
|
||||||
@ -179,6 +188,7 @@ class Langchain4jChatModelConfigurerImpl implements Langchain4jChatModelConfigur
|
|||||||
.gitHubToken(chatModel.providerConfig().personalAccessToken())
|
.gitHubToken(chatModel.providerConfig().personalAccessToken())
|
||||||
.modelName(modelConfig.modelId())
|
.modelName(modelConfig.modelId())
|
||||||
.temperature(modelConfig.temperature())
|
.temperature(modelConfig.temperature())
|
||||||
|
.topP(modelConfig.topP())
|
||||||
.timeout(toDuration(modelConfig.timeoutSeconds()))
|
.timeout(toDuration(modelConfig.timeoutSeconds()))
|
||||||
.maxRetries(modelConfig.maxRetries())
|
.maxRetries(modelConfig.maxRetries())
|
||||||
.build();
|
.build();
|
||||||
|
|||||||
@ -23,14 +23,10 @@ public sealed interface AiChatModelConfig<C extends AiChatModelConfig<C>> extend
|
|||||||
GoogleVertexAiGeminiChatModel.Config, MistralAiChatModel.Config, AnthropicChatModel.Config,
|
GoogleVertexAiGeminiChatModel.Config, MistralAiChatModel.Config, AnthropicChatModel.Config,
|
||||||
AmazonBedrockChatModel.Config, GitHubModelsChatModel.Config {
|
AmazonBedrockChatModel.Config, GitHubModelsChatModel.Config {
|
||||||
|
|
||||||
Double temperature();
|
|
||||||
|
|
||||||
Integer timeoutSeconds();
|
Integer timeoutSeconds();
|
||||||
|
|
||||||
Integer maxRetries();
|
Integer maxRetries();
|
||||||
|
|
||||||
C withTemperature(Double temperature);
|
|
||||||
|
|
||||||
C withTimeoutSeconds(Integer timeoutSeconds);
|
C withTimeoutSeconds(Integer timeoutSeconds);
|
||||||
|
|
||||||
C withMaxRetries(Integer maxRetries);
|
C withMaxRetries(Integer maxRetries);
|
||||||
|
|||||||
@ -16,47 +16,28 @@
|
|||||||
package org.thingsboard.server.common.data.ai.model.chat;
|
package org.thingsboard.server.common.data.ai.model.chat;
|
||||||
|
|
||||||
import dev.langchain4j.model.chat.ChatModel;
|
import dev.langchain4j.model.chat.ChatModel;
|
||||||
|
import lombok.With;
|
||||||
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
||||||
import org.thingsboard.server.common.data.ai.provider.AmazonBedrockProviderConfig;
|
import org.thingsboard.server.common.data.ai.provider.AmazonBedrockProviderConfig;
|
||||||
|
|
||||||
public record AmazonBedrockChatModel(
|
public record AmazonBedrockChatModel(
|
||||||
AiModelType modelType,
|
AiModelType modelType,
|
||||||
AmazonBedrockProviderConfig providerConfig,
|
AmazonBedrockProviderConfig providerConfig,
|
||||||
Config modelConfig
|
@With Config modelConfig
|
||||||
) implements AiChatModel<AmazonBedrockChatModel.Config> {
|
) implements AiChatModel<AmazonBedrockChatModel.Config> {
|
||||||
|
|
||||||
|
@With
|
||||||
public record Config(
|
public record Config(
|
||||||
String modelId,
|
String modelId,
|
||||||
Double temperature,
|
Double temperature,
|
||||||
|
Double topP,
|
||||||
Integer timeoutSeconds,
|
Integer timeoutSeconds,
|
||||||
Integer maxRetries
|
Integer maxRetries
|
||||||
) implements AiChatModelConfig<AmazonBedrockChatModel.Config> {
|
) implements AiChatModelConfig<AmazonBedrockChatModel.Config> {}
|
||||||
|
|
||||||
@Override
|
|
||||||
public AmazonBedrockChatModel.Config withTemperature(Double temperature) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public AmazonBedrockChatModel.Config withTimeoutSeconds(Integer timeoutSeconds) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public AmazonBedrockChatModel.Config withMaxRetries(Integer maxRetries) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
||||||
return configurer.configureChatModel(this);
|
return configurer.configureChatModel(this);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
|
||||||
public AmazonBedrockChatModel withModelConfig(AmazonBedrockChatModel.Config config) {
|
|
||||||
return new AmazonBedrockChatModel(modelType, providerConfig, config);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|||||||
@ -16,47 +16,28 @@
|
|||||||
package org.thingsboard.server.common.data.ai.model.chat;
|
package org.thingsboard.server.common.data.ai.model.chat;
|
||||||
|
|
||||||
import dev.langchain4j.model.chat.ChatModel;
|
import dev.langchain4j.model.chat.ChatModel;
|
||||||
|
import lombok.With;
|
||||||
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
||||||
import org.thingsboard.server.common.data.ai.provider.AnthropicProviderConfig;
|
import org.thingsboard.server.common.data.ai.provider.AnthropicProviderConfig;
|
||||||
|
|
||||||
public record AnthropicChatModel(
|
public record AnthropicChatModel(
|
||||||
AiModelType modelType,
|
AiModelType modelType,
|
||||||
AnthropicProviderConfig providerConfig,
|
AnthropicProviderConfig providerConfig,
|
||||||
Config modelConfig
|
@With Config modelConfig
|
||||||
) implements AiChatModel<AnthropicChatModel.Config> {
|
) implements AiChatModel<AnthropicChatModel.Config> {
|
||||||
|
|
||||||
|
@With
|
||||||
public record Config(
|
public record Config(
|
||||||
String modelId,
|
String modelId,
|
||||||
Double temperature,
|
Double temperature,
|
||||||
|
Double topP,
|
||||||
Integer timeoutSeconds,
|
Integer timeoutSeconds,
|
||||||
Integer maxRetries
|
Integer maxRetries
|
||||||
) implements AiChatModelConfig<AnthropicChatModel.Config> {
|
) implements AiChatModelConfig<AnthropicChatModel.Config> {}
|
||||||
|
|
||||||
@Override
|
|
||||||
public AnthropicChatModel.Config withTemperature(Double temperature) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public AnthropicChatModel.Config withTimeoutSeconds(Integer timeoutSeconds) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public AnthropicChatModel.Config withMaxRetries(Integer maxRetries) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
||||||
return configurer.configureChatModel(this);
|
return configurer.configureChatModel(this);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
|
||||||
public AnthropicChatModel withModelConfig(AnthropicChatModel.Config config) {
|
|
||||||
return new AnthropicChatModel(modelType, providerConfig, config);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|||||||
@ -16,47 +16,28 @@
|
|||||||
package org.thingsboard.server.common.data.ai.model.chat;
|
package org.thingsboard.server.common.data.ai.model.chat;
|
||||||
|
|
||||||
import dev.langchain4j.model.chat.ChatModel;
|
import dev.langchain4j.model.chat.ChatModel;
|
||||||
|
import lombok.With;
|
||||||
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
||||||
import org.thingsboard.server.common.data.ai.provider.AzureOpenAiProviderConfig;
|
import org.thingsboard.server.common.data.ai.provider.AzureOpenAiProviderConfig;
|
||||||
|
|
||||||
public record AzureOpenAiChatModel(
|
public record AzureOpenAiChatModel(
|
||||||
AiModelType modelType,
|
AiModelType modelType,
|
||||||
AzureOpenAiProviderConfig providerConfig,
|
AzureOpenAiProviderConfig providerConfig,
|
||||||
Config modelConfig
|
@With Config modelConfig
|
||||||
) implements AiChatModel<AzureOpenAiChatModel.Config> {
|
) implements AiChatModel<AzureOpenAiChatModel.Config> {
|
||||||
|
|
||||||
|
@With
|
||||||
public record Config(
|
public record Config(
|
||||||
String modelId,
|
String modelId,
|
||||||
Double temperature,
|
Double temperature,
|
||||||
|
Double topP,
|
||||||
Integer timeoutSeconds,
|
Integer timeoutSeconds,
|
||||||
Integer maxRetries
|
Integer maxRetries
|
||||||
) implements AiChatModelConfig<AzureOpenAiChatModel.Config> {
|
) implements AiChatModelConfig<AzureOpenAiChatModel.Config> {}
|
||||||
|
|
||||||
@Override
|
|
||||||
public AzureOpenAiChatModel.Config withTemperature(Double temperature) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public AzureOpenAiChatModel.Config withTimeoutSeconds(Integer timeoutSeconds) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public AzureOpenAiChatModel.Config withMaxRetries(Integer maxRetries) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
||||||
return configurer.configureChatModel(this);
|
return configurer.configureChatModel(this);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
|
||||||
public AzureOpenAiChatModel withModelConfig(AzureOpenAiChatModel.Config config) {
|
|
||||||
return new AzureOpenAiChatModel(modelType, providerConfig, config);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|||||||
@ -16,47 +16,28 @@
|
|||||||
package org.thingsboard.server.common.data.ai.model.chat;
|
package org.thingsboard.server.common.data.ai.model.chat;
|
||||||
|
|
||||||
import dev.langchain4j.model.chat.ChatModel;
|
import dev.langchain4j.model.chat.ChatModel;
|
||||||
|
import lombok.With;
|
||||||
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
||||||
import org.thingsboard.server.common.data.ai.provider.GithubModelsProviderConfig;
|
import org.thingsboard.server.common.data.ai.provider.GithubModelsProviderConfig;
|
||||||
|
|
||||||
public record GitHubModelsChatModel(
|
public record GitHubModelsChatModel(
|
||||||
AiModelType modelType,
|
AiModelType modelType,
|
||||||
GithubModelsProviderConfig providerConfig,
|
GithubModelsProviderConfig providerConfig,
|
||||||
Config modelConfig
|
@With Config modelConfig
|
||||||
) implements AiChatModel<GitHubModelsChatModel.Config> {
|
) implements AiChatModel<GitHubModelsChatModel.Config> {
|
||||||
|
|
||||||
|
@With
|
||||||
public record Config(
|
public record Config(
|
||||||
String modelId,
|
String modelId,
|
||||||
Double temperature,
|
Double temperature,
|
||||||
|
Double topP,
|
||||||
Integer timeoutSeconds,
|
Integer timeoutSeconds,
|
||||||
Integer maxRetries
|
Integer maxRetries
|
||||||
) implements AiChatModelConfig<GitHubModelsChatModel.Config> {
|
) implements AiChatModelConfig<GitHubModelsChatModel.Config> {}
|
||||||
|
|
||||||
@Override
|
|
||||||
public GitHubModelsChatModel.Config withTemperature(Double temperature) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public GitHubModelsChatModel.Config withTimeoutSeconds(Integer timeoutSeconds) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public GitHubModelsChatModel.Config withMaxRetries(Integer maxRetries) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
||||||
return configurer.configureChatModel(this);
|
return configurer.configureChatModel(this);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
|
||||||
public GitHubModelsChatModel withModelConfig(GitHubModelsChatModel.Config config) {
|
|
||||||
return new GitHubModelsChatModel(modelType, providerConfig, config);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|||||||
@ -16,47 +16,28 @@
|
|||||||
package org.thingsboard.server.common.data.ai.model.chat;
|
package org.thingsboard.server.common.data.ai.model.chat;
|
||||||
|
|
||||||
import dev.langchain4j.model.chat.ChatModel;
|
import dev.langchain4j.model.chat.ChatModel;
|
||||||
|
import lombok.With;
|
||||||
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
||||||
import org.thingsboard.server.common.data.ai.provider.GoogleAiGeminiProviderConfig;
|
import org.thingsboard.server.common.data.ai.provider.GoogleAiGeminiProviderConfig;
|
||||||
|
|
||||||
public record GoogleAiGeminiChatModel(
|
public record GoogleAiGeminiChatModel(
|
||||||
AiModelType modelType,
|
AiModelType modelType,
|
||||||
GoogleAiGeminiProviderConfig providerConfig,
|
GoogleAiGeminiProviderConfig providerConfig,
|
||||||
Config modelConfig
|
@With Config modelConfig
|
||||||
) implements AiChatModel<GoogleAiGeminiChatModel.Config> {
|
) implements AiChatModel<GoogleAiGeminiChatModel.Config> {
|
||||||
|
|
||||||
|
@With
|
||||||
public record Config(
|
public record Config(
|
||||||
String modelId,
|
String modelId,
|
||||||
Double temperature,
|
Double temperature,
|
||||||
|
Double topP,
|
||||||
Integer timeoutSeconds,
|
Integer timeoutSeconds,
|
||||||
Integer maxRetries
|
Integer maxRetries
|
||||||
) implements AiChatModelConfig<GoogleAiGeminiChatModel.Config> {
|
) implements AiChatModelConfig<GoogleAiGeminiChatModel.Config> {}
|
||||||
|
|
||||||
@Override
|
|
||||||
public Config withTemperature(Double temperature) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Config withTimeoutSeconds(Integer timeoutSeconds) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Config withMaxRetries(Integer maxRetries) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
||||||
return configurer.configureChatModel(this);
|
return configurer.configureChatModel(this);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
|
||||||
public GoogleAiGeminiChatModel withModelConfig(GoogleAiGeminiChatModel.Config config) {
|
|
||||||
return new GoogleAiGeminiChatModel(modelType, providerConfig, config);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|||||||
@ -16,47 +16,28 @@
|
|||||||
package org.thingsboard.server.common.data.ai.model.chat;
|
package org.thingsboard.server.common.data.ai.model.chat;
|
||||||
|
|
||||||
import dev.langchain4j.model.chat.ChatModel;
|
import dev.langchain4j.model.chat.ChatModel;
|
||||||
|
import lombok.With;
|
||||||
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
||||||
import org.thingsboard.server.common.data.ai.provider.GoogleVertexAiGeminiProviderConfig;
|
import org.thingsboard.server.common.data.ai.provider.GoogleVertexAiGeminiProviderConfig;
|
||||||
|
|
||||||
public record GoogleVertexAiGeminiChatModel(
|
public record GoogleVertexAiGeminiChatModel(
|
||||||
AiModelType modelType,
|
AiModelType modelType,
|
||||||
GoogleVertexAiGeminiProviderConfig providerConfig,
|
GoogleVertexAiGeminiProviderConfig providerConfig,
|
||||||
Config modelConfig
|
@With Config modelConfig
|
||||||
) implements AiChatModel<GoogleVertexAiGeminiChatModel.Config> {
|
) implements AiChatModel<GoogleVertexAiGeminiChatModel.Config> {
|
||||||
|
|
||||||
|
@With
|
||||||
public record Config(
|
public record Config(
|
||||||
String modelId,
|
String modelId,
|
||||||
Double temperature,
|
Double temperature,
|
||||||
|
Double topP,
|
||||||
Integer timeoutSeconds, // TODO: not supported by Vertex AI
|
Integer timeoutSeconds, // TODO: not supported by Vertex AI
|
||||||
Integer maxRetries
|
Integer maxRetries
|
||||||
) implements AiChatModelConfig<GoogleVertexAiGeminiChatModel.Config> {
|
) implements AiChatModelConfig<GoogleVertexAiGeminiChatModel.Config> {}
|
||||||
|
|
||||||
@Override
|
|
||||||
public Config withTemperature(Double temperature) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Config withTimeoutSeconds(Integer timeoutSeconds) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Config withMaxRetries(Integer maxRetries) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
||||||
return configurer.configureChatModel(this);
|
return configurer.configureChatModel(this);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
|
||||||
public GoogleVertexAiGeminiChatModel withModelConfig(GoogleVertexAiGeminiChatModel.Config config) {
|
|
||||||
return new GoogleVertexAiGeminiChatModel(modelType, providerConfig, config);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|||||||
@ -16,47 +16,28 @@
|
|||||||
package org.thingsboard.server.common.data.ai.model.chat;
|
package org.thingsboard.server.common.data.ai.model.chat;
|
||||||
|
|
||||||
import dev.langchain4j.model.chat.ChatModel;
|
import dev.langchain4j.model.chat.ChatModel;
|
||||||
|
import lombok.With;
|
||||||
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
||||||
import org.thingsboard.server.common.data.ai.provider.MistralAiProviderConfig;
|
import org.thingsboard.server.common.data.ai.provider.MistralAiProviderConfig;
|
||||||
|
|
||||||
public record MistralAiChatModel(
|
public record MistralAiChatModel(
|
||||||
AiModelType modelType,
|
AiModelType modelType,
|
||||||
MistralAiProviderConfig providerConfig,
|
MistralAiProviderConfig providerConfig,
|
||||||
Config modelConfig
|
@With Config modelConfig
|
||||||
) implements AiChatModel<MistralAiChatModel.Config> {
|
) implements AiChatModel<MistralAiChatModel.Config> {
|
||||||
|
|
||||||
|
@With
|
||||||
public record Config(
|
public record Config(
|
||||||
String modelId,
|
String modelId,
|
||||||
Double temperature,
|
Double temperature,
|
||||||
|
Double topP,
|
||||||
Integer timeoutSeconds,
|
Integer timeoutSeconds,
|
||||||
Integer maxRetries
|
Integer maxRetries
|
||||||
) implements AiChatModelConfig<MistralAiChatModel.Config> {
|
) implements AiChatModelConfig<MistralAiChatModel.Config> {}
|
||||||
|
|
||||||
@Override
|
|
||||||
public Config withTemperature(Double temperature) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Config withTimeoutSeconds(Integer timeoutSeconds) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public Config withMaxRetries(Integer maxRetries) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
||||||
return configurer.configureChatModel(this);
|
return configurer.configureChatModel(this);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
|
||||||
public MistralAiChatModel withModelConfig(Config config) {
|
|
||||||
return new MistralAiChatModel(modelType, providerConfig, config);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|||||||
@ -16,47 +16,28 @@
|
|||||||
package org.thingsboard.server.common.data.ai.model.chat;
|
package org.thingsboard.server.common.data.ai.model.chat;
|
||||||
|
|
||||||
import dev.langchain4j.model.chat.ChatModel;
|
import dev.langchain4j.model.chat.ChatModel;
|
||||||
|
import lombok.With;
|
||||||
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
import org.thingsboard.server.common.data.ai.model.AiModelType;
|
||||||
import org.thingsboard.server.common.data.ai.provider.OpenAiProviderConfig;
|
import org.thingsboard.server.common.data.ai.provider.OpenAiProviderConfig;
|
||||||
|
|
||||||
public record OpenAiChatModel(
|
public record OpenAiChatModel(
|
||||||
AiModelType modelType,
|
AiModelType modelType,
|
||||||
OpenAiProviderConfig providerConfig,
|
OpenAiProviderConfig providerConfig,
|
||||||
Config modelConfig
|
@With Config modelConfig
|
||||||
) implements AiChatModel<OpenAiChatModel.Config> {
|
) implements AiChatModel<OpenAiChatModel.Config> {
|
||||||
|
|
||||||
|
@With
|
||||||
public record Config(
|
public record Config(
|
||||||
String modelId,
|
String modelId,
|
||||||
Double temperature,
|
Double temperature,
|
||||||
|
Double topP,
|
||||||
Integer timeoutSeconds,
|
Integer timeoutSeconds,
|
||||||
Integer maxRetries
|
Integer maxRetries
|
||||||
) implements AiChatModelConfig<OpenAiChatModel.Config> {
|
) implements AiChatModelConfig<OpenAiChatModel.Config> {}
|
||||||
|
|
||||||
@Override
|
|
||||||
public OpenAiChatModel.Config withTemperature(Double temperature) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public OpenAiChatModel.Config withTimeoutSeconds(Integer timeoutSeconds) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public OpenAiChatModel.Config withMaxRetries(Integer maxRetries) {
|
|
||||||
return new Config(modelId, temperature, timeoutSeconds, maxRetries);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
public ChatModel configure(Langchain4jChatModelConfigurer configurer) {
|
||||||
return configurer.configureChatModel(this);
|
return configurer.configureChatModel(this);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
|
||||||
public OpenAiChatModel withModelConfig(OpenAiChatModel.Config config) {
|
|
||||||
return new OpenAiChatModel(modelType, providerConfig, config);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user