Replies: 1 comment
-
Hey there, @mathabes! I'm here to assist you with any bugs, questions, or contributions. Let's tackle this issue together! To address the performance issues with the groq llama 3-70b-8192 model, you can try the following optimizations and configurations:
These configurations will help ensure that Langflow can efficiently handle large models by scaling horizontally and vertically, using a more robust database system, and optimizing model-specific settings [1][2]. |
Beta Was this translation helpful? Give feedback.
-
Hi, I'm using the groq llama 3-70b-8192 template component to make a conversational agent. My requests are taking longer and longer, reaching 30 seconds. This only happens when I use the 70b template, because with 8b the requests take a maximum of 4 seconds.
I'm hosting langflow on railway with a Pro plan. Is there any way to solve this problem?
I am sending a high amount of instructions in the conversation, could this be affecting the 70b model?
Beta Was this translation helpful? Give feedback.
All reactions