• 2 Posts
  • 10 Comments
Joined 5 months ago
cake
Cake day: August 7th, 2025

help-circle


  • Short answer: you are cooked. The current old language model (it’s in the process of being replaced, but we don’t know when) has a 4k context window token, aprox 2 pages of text. This is the what the model can remember. Best option is to use the non-fancy AI Chat (https://perchance.org/ai-chat) put less than 1k tokens in your character description or even better just put the essential, and put the rest of your lore in the Scenario & Lore and pray to the Old Gods you don’t exceed 3k-4k characters or you will experience a new universe of roleplaying…the goldfish memory experience. I love Perchance but I gave up months ago about doing roleplay with strong scenarios and world building because with 4k tokens it’s simply impossible. We all hope the new model will have at least 10k tokens window so you will have the bare minimum to create cool worlds and scenarios without having to describe each character in one line of text. More than 20k tokens will be fucking amazing…but one only can dream.




  • Well, then I have good news for you, the chat is down because it is being upgraded after 1.5 years with the new Llama 3. So you just need to wait and be happy because if we are lucky and it is Llama 3.3, we could have 128k tokens context instead of the shitty 4k tokens or at least 10k or 20k that would make it possible to finally have lores and world building, so goodbye to goldfish memory issues. But “let’s not get ahead of ourselves”. lol Speaking about running LLM locally. I have an RTX 3060, and it takes a minute to get an answer, which is why I generate images locally with comfy and use Perchance for roleplay…so you are fucked.



  • Well, I hope it’s 3.3, after two years of waiting, and with the image generators using Chroma and Flux. I can’t complain about a free service anyway. It will be a bittersweet victory; having 2x context will surely improve the roleplay experience, but it will still be impossible to create lore and world-building unless we have at least 50-60k context tokens. And yes, the actual model is amazing, but it’s being held down by the context window; you cannot create a deep lore or word building, and worse of all, the chat forgets the key events around 3-4 pages of convo, having to rely on summaries that miss A LOT of important elements.