Models and Settings
Model, max tokens, temperature.
When you're in a chat, you can customise settings through the three-dot menu. These affect how the AI generates responses.
Model selection
The model dropdown lets you choose which LLM powers your character. Some models are better at following instructions precisely. Some are more creative and expressive. Some are faster but less detailed. If your character feels “off,” try a different model. Sometimes switching models is the single biggest improvement.
Max tokens
Controls how long responses can be. The slider goes from 64 to 5,000, with a default of 200.
For chat mode, the default (200) works well. For roleplay, you'll probably want 600+.
Temperature
Controls creativity versus consistency. Default is 0.8.
- Low (0.0 — 0.4): Very consistent. Can feel repetitive over long conversations.
- Medium (0.5 — 0.9): Good balance. Default zone.
- High (1.0 — 2.0): Very creative, unpredictable. Use carefully.
If your character feels robotic, try increasing temperature. If they feel random and unfocused, try lowering it.
Advanced: Top P and Top K
- Top P (default 1.0): Lower values make output more focused. Reducing to 0.9 can help with consistency.
- Top K (default 50): Lower values reduce randomness. Most users can leave this alone.