

There is a better explanation for the behavior you are experiencing, and yes, it is one if not the biggest hurdle the new model has yet to overcome: You have hit a log long enough that the model is starting to make a word salad of its past inputs as it “inbreeds”.
What I mean by this is something explained before: For generators such as AI Chat and ACC, the input will be mostly 70% AI made and only 30% handwritten (95%-5% in AI RPG which crashes faster), because the whole log is an input for the next output. Of course, the shortest the log is, the less you’ll feel the effect of the model being insufferable because you still have the long instruction block “holding back” the manic behavior.
I agree, this is something that has to be worked on from the development side, otherwise generators such as AI Chat or Story Generator are rendered short-lived as the point of them is to grow progressively, and as today, instability can happen as soon as 150kB-200kB, being significantly lower that what this model was able to hold in the past. However, a temporary fix on our side of things is to just make a “partition” of your log/story. Meaning:
- Plan and start your run as usual.
- Save constantly, monitoring the size of the log.
- When you hit the 100kB mark, try to get to a point where you can “start over” as a point where you can keep moving without requiring the context prior.
- Make a copy, delete all prior to that desired state, load the save and continue pretending that nothing happened.
That will keep the model “fresh” at the cost of losing “memory”, which can be worked around as you can update the bios or instructions which will have better chances of working now under a clean slate.
It is not the best way to work around this, but it is better than wrestling with all the nonsense that the model will produce past the 250kB threshold.
Hope that helps and… also hole that a future update would make the model more stable rather than more unstable. At least something that was fixed and that the dev deserves more credit for making it work, is that at least now the English has improved significantly compared with the first release. In terms of grammar, content and consistency. I know, past the 250kB it is “allegories” or “crazy man ramblings”, but… it is good English! 😅










Correct, that’s what I implied, since otherwise, past the 1Mb you’ll experience “groundhog day” unable to escape the loop no matter what you do.
Now… let me tell you buddy, you just scratched the tip of the iceberg with the model new obsessions. Just to showcase a couple:
You are in the money with one thing, all this is product of the training data, and not even the one that comes pre-packed with DeepSeek (I still hold that this is the current model being used, if I’m wrong, I’ll gladly accept the failure on my prediction), this is product of the dataset being used to re-train the model into working for dev’s end. For example, the “knuckles turning white” phrase appeared rarely with the old Llama model, but it was a one in a hundred occurrence as the model didn’t care for that construction and rather focused on a different set of obsessions.
This is a never ending problem with all LLMs though, as in all languages, some constructions are more often than others, and since in both AI Chat and ACC the model is constrained by the “Make a story/roleplay” context, it produces those pseudo-catchphrases incredibly often. In the past we had to deal with “Let’s not get ahead of ourselves” or “We should tread carefully” appearing always no matter the situations, now “knuckles turning white” or similar are the new catchphrases in town.
In an older post I warned about this, since DeepSeek trying to be more “smart” will take everything to face value, so the “correct” answer for many situations tends to be any of these constructions cited, and performing extreme training will yield us a model as dumb and stubborn as Llama was, but with a new set of obsessions plus the inability to move forward which Llama could despite it being exasperating at times. There is progress with the new model, I won’t deny it, but the threshold from were we entered “groundhog day” has been reduced from 1Mb+ to barely 250-500kb and I suspect it will keep reducing if the training is done on top of the existing one, rendering the model pointless for AI Chat, AI RPG or ACC.
Then again, I could be wrong and a future update will allow the context window to hold further as Llama where 15Mb+ was possible and manageable without much maintenance. Some degree of obsession on any LLM is impossible to avoid, what is important is that the model doesn’t turn it into a word salad that goes nowhere. That I think is one of the biggest challenges the development of ai-text-plugin has.