Replies: 2 comments
-
I just noticed that when I use "chat" mode in textgen-ui I get the same poor response as when I prompt the model with llama-cpp-python. I guess now I need to find out why does textgen 'default' mode work better than 'chat' mode and native llama-cpp-python |
Beta Was this translation helpful? Give feedback.
0 replies
-
never mind... thank you for participating in my rubber ducky moment. The difference was due to the prompting. I changed the default prompt structure from my formerly trusty 'alpaca' to just asking the question, and the model gained 50 IQ points. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Well this sucks because my code relies on llapa-cpp-python directly, not through an API and I have an entire roadmap for how I am going to build an agent system using the llama-cpp-python as the provider.
But textgen uses llama-cpp-python for some of its functionality, it is the requirement, so I am confused and kind of exhausted. And I am hoping there is an easy answer that would allow me to get the same performance from llama-cpp-python as I get with what is mostly its GUI.
Beta Was this translation helpful? Give feedback.
All reactions