From be1ddb142f4bc206b9020621a736d66af84041f6 Mon Sep 17 00:00:00 2001 From: JohannesGaessler Date: Wed, 30 Aug 2023 20:10:53 +0200 Subject: [PATCH] CUDA: mul_mat_q=true llama_context_params default --- llama.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama.cpp b/llama.cpp index fcd6f276a0655..95ee6ffe41c3a 100644 --- a/llama.cpp +++ b/llama.cpp @@ -5287,7 +5287,7 @@ struct llama_context_params llama_context_default_params() { /*.progress_callback =*/ nullptr, /*.progress_callback_user_data =*/ nullptr, /*.low_vram =*/ false, - /*.mul_mat_q =*/ false, + /*.mul_mat_q =*/ true, /*.f16_kv =*/ true, /*.logits_all =*/ false, /*.vocab_only =*/ false,