Replies: 1 comment 2 replies
-
For that prompt specifically you wouldn't need
|
Beta Was this translation helpful? Give feedback.
-
For that prompt specifically you wouldn't need
|
Beta Was this translation helpful? Give feedback.
-
Hello, I am using a llama-2-chat model with exllamav2. My typical prompt is something like
[INST] <<SYS>> You are a helpful assistant..<</SYS>> Instruction ...[/INST]
. Since I am adding system prompt and instruction tokens, do I need to setencode_special_tokens
anddecode_special_tokens
to True?Beta Was this translation helpful? Give feedback.
All reactions