Here you'll find common questions regarding WyvernChat, chatting, LLMs, and anything else in the world of AI roleplay.
If your character wasn't approved by our automod, it will be reviewed by our volunteer human moderation team for any rule-violating content. Reviews typically take no more than 12 hours, but can take longer if we're under heavy load.
Due to security reasons, we only allow WyvernChat to be open in one tab. Please close your new tabs and the website should function as normal.
This means there's an issue with the model's connection. Check the model status page from Featherless for the latest updates on WyvernChat's models.
Here is which model corresponds to which Free queue option:
This issue happens frequently in the beginning of chats. Your character may be doing any of these:
512
-1500
1
- 1.15
0.3
- 0.8
These solutions have been tested for LLAMA-3.1 finetunes, Magnum finetunes, and Mistral Finetunes.
<|im_end|>
?<|im_end|>
is a special token used to mark the end of a message or turn in a conversation. Model developers use these to train AI for turn-based chats, helping it understand where one message stops and another starts). These marks for conversation structuring are named instruct templates.
<|im_end|>
is a part of a chat template known as ChatML. ChatML is widely compatible with models like LLaMA, Qwen, and Mistral, even if they have their own templates. So ChatML is a default in many sharable presets on Wyvern.
1. You’re using DeepSeek
DeepSeek is a special cookie and sassy rebel. Technically, DeepSeek has no instruct template. It cuts responses when needed but sometimes leaves <|im_end|>
hanging in the end. No clear reason—it’s just DeepSeek being DeepSeek.
Solutions:
ON
2. There's a missing stop-sequence for <|im_end|>
If you’ve been experimenting with your chat template, you might’ve forgotten to add <|im_end|>
to the stop sequences. Without it, the model won’t know to cut that token from the output. The sight of it is LLM keep generating out of context (OOC) text after <|im_end|> with instructions or reasonings.
Solutions:
<|im_end|>
to the stop sequences.<|im_end|>
. It should be on the list.3. Are you certain It is exactly <|im_end|>
. not something else?
Some models pull raw text from web sources with HTML or XML tags like , ,
or The model sees these in your chat history and tries to mimic them, spitting out variations like , < | im_end | >
, or similar. Autocut only catches exactly <|im_end|>
matches, so these weird variants slip through.
Solutions:
Don't see your question answered? Ask your question in our Discord server's
#support
channel!