From 2047a05f84c2fb0b521db3990f5da873c723319a Mon Sep 17 00:00:00 2001 From: Daniel Strobusch <1847260+dastrobu@users.noreply.github.com> Date: Tue, 27 Aug 2024 11:23:20 +0200 Subject: [PATCH] fix docs on topP Signed-off-by: Daniel Strobusch <1847260+dastrobu@users.noreply.github.com> --- .../azure/openai/runtime/config/ChatModelConfig.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/model-providers/openai/azure-openai/runtime/src/main/java/io/quarkiverse/langchain4j/azure/openai/runtime/config/ChatModelConfig.java b/model-providers/openai/azure-openai/runtime/src/main/java/io/quarkiverse/langchain4j/azure/openai/runtime/config/ChatModelConfig.java index 140ae3875..a40dc122f 100644 --- a/model-providers/openai/azure-openai/runtime/src/main/java/io/quarkiverse/langchain4j/azure/openai/runtime/config/ChatModelConfig.java +++ b/model-providers/openai/azure-openai/runtime/src/main/java/io/quarkiverse/langchain4j/azure/openai/runtime/config/ChatModelConfig.java @@ -50,7 +50,7 @@ public interface ChatModelConfig { * An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens * with topP probability mass. * 0.1 means only the tokens comprising the top 10% probability mass are considered. - * It is recommended to alter this or topP, but not both. + * It is recommended to alter this or temperature, but not both. */ @WithDefault("1.0") Double topP();