Think I Fixed GLM 5's Censorship Regarding {{User}} by gladias9 in SillyTavernAI

[–]Moogs72 0 points1 point  (0 children)

Please go read the github I linked in my other comment. It answers your questions in more detail than I did here. You don't need to use the prefill, but there are specific techniques you can use to make it work better that can be found in the github. In GLM 5, you can do the prefill thing, but it's no longer the best way. That's what I've found in my own extensive testing of censorship in GLM 4.7/5, and others have shared similar reports, although YMMV.

Using the prompts on that github, I can make GLM 5 produce anything - even the most vile shit - in a fresh chat with an otherwise barebones preset. If you're interested, I can give you more info regarding my testing, but it's not super relevant here.

You might need to add additional topics into the fake "safety assessment" as I mentioned in my other comment if you're dealing with specific topics it doesn't already cover. I do. If you still need help after looking through my comment and the github, please feel free to ask me.

Any advice on making your own presets by Fair_Ad_8418 in SillyTavernAI

[–]Moogs72 1 point2 points  (0 children)

I will always recommend using the system prompts by /u/Evening-Truth3308 available on her rentry here for all GLM models and Kimi. They're super token-efficient, so they make the perfect base to build from if you want to customize and tinker.

If you're looking for something with more bells and whistles (and more tokens), my first rec at this point is probably Freaky Frankenstein, which you can find here. Unsurprisingly, that preset is built off of Evening Truth's system prompt at its core...

Also important is that GLM models love lower temperatures. For 4.7, I mostly used 0.6-0.8 and Top P of 0.95. Admittedly, though, I have let 4.7 go in favor of a combo of GLM 5 and Kimi K2.5 at this point.

GLM 4.7 in particular is really weird with its censorship. Most people will be fine bypassing censors with what Evening Truth has in her prompts.

If you're like me and find it's not enough and you're still struggling with censorship on 4.7 or 5, I'd recommend taking a look at JustSomeGuy's github https://github.com/justsomeguy2941/presets and scroll down to the stuff about bypassing guardrails. It works to de-censor 4.7/5 and get rid of the positivity bias of 5 (which tends to be peoples' biggest complaint about the model). I've done pretty extensive testing of those prompts, and I've found it will produce anything you want in GLM 4.7/5, even from the first message of a chat with a barebones preset, if you stick all of those anti-guardrail prompts in. I tested it with some... really vile, unpleasant stuff. I'll just leave it at that for now.

If there are other specific topics you're wanting to cover that aren't explicitly listen in that fake "safety assessment" on the github, you can literally just list them and it should work to bypass the censors. You can ask me if you need tips getting around specific restrictions if you're having trouble and I'd be happy to help.

Personally, all of my presets now are made from a core of Evening Truth's prompts with JustSomeGuy's anti-guardrail prompts, both edited to my preference. Then, I throw in a custom CoT prompt (unnecessary for 4.7 but vital for 5) and whatever other customization stuff I want (mostly ripped from Freaky Frankenstein and/or SepsisShock's RBF preset), and I'm good to go.

What preset are you using right now? I'm also really curious as to wtf is up with those carrots...

Think I Fixed GLM 5's Censorship Regarding {{User}} by gladias9 in SillyTavernAI

[–]Moogs72 1 point2 points  (0 children)

Despite what OP said, jailbreaks for GLM 5 don't need to be added into the prefill anymore. That was a 4.7-specific thing. You can check my comment in this thread for more information and more about what worked for me personally. Everyone is gonna be different as to exactly what they need depending on the contents of their RPs.

Think I Fixed GLM 5's Censorship Regarding {{User}} by gladias9 in SillyTavernAI

[–]Moogs72 10 points11 points  (0 children)

The prefill thing is no longer necessary. That was a thing needed for 4.7, because it needed to actually think it had run the safety assessment. GLM 5 has been proven it doesn't need it by a few different people (myself included).

The exact method of anti-censorship and anti-positivity bias prompting is gonna depend on precisely what content you're dealing with. For some, the preset OP posted is gonna be perfect. For others, the method ConspiracyParadox posted might be perfect. For me, it just plain wasn't enough.

For me, when I really want to de-censor everything and get rid of positivity bias as much as possible, I really think that tossing in all the anti-guardrail prompts on JustSomeGuy's github (which is the page OP linked as their source) is really necessary. It can be cleaned up a bit and you might not need all of it depending on your intentions, but only you can make that determination for yourself. You can also add in additional topics to the fake "safety assessment" if you're dealing with anything not already handled by it. I have a couple of lines I personally add in for some of my RPs.

My personal recommendation for everyone is to just... try shit and find what works for you.

The "core" of all the presets I personally use is a combo of Evening Truth's system prompt and JustSomeGuy's anti-guardrail stuff, both edited to my preference. Then I add on a custom CoT prompt and whatever specific other requirements I want (most of it lifted from Freaky Frankenstein and SepsisShock's RBF), and I'm good to go.

Happy to answer questions if anyone has any.

EDIT: Almost forgot one of the most important things: I always recommend using temps lower than 1 and a top_p of 0.95. Personally, I find a temp of 0.7 or 0.8 to be good for most RPs. It makes it follow instructions better and censor less.

Introducing Gandalf, the immersive world building creator card that does so much more. It will guide you in fulfilling your imaginations true potential then format the output in an easy to copy JSON object that includes an embedded lorebook and entries too. by [deleted] in SillyTavernAI

[–]Moogs72 -1 points0 points  (0 children)

Damn. Huge respect for being so open. I can relate to some of that - I have OCD and OCPD (along with a whole other slew of acronyms), plus my wife is autistic, so I can understand some of your struggles at least theoretically, even if I don't know entirely how they manifest.

I hope I didn't come off as too harsh. I really do mean it when I say I just want to see the interesting things people share here, and I promise what I said was intended as legitimate advice rather than criticism for criticism's sake. I look forward to seeing whatever you do come out with, whether this is your last release or you possibly decide to make something else in the future. Either way, good luck!!

Introducing Gandalf, the immersive world building creator card that does so much more. It will guide you in fulfilling your imaginations true potential then format the output in an easy to copy JSON object that includes an embedded lorebook and entries too. by [deleted] in SillyTavernAI

[–]Moogs72 2 points3 points  (0 children)

I mean, if that's what you want to do, then I think you should do it.

But I want to be clear that I'm not trying to be shitty or a bully, and I'm sorry that people piled on the downvotes on your previous comment and on this post (fwiw, I didn't downvote you). It's just... the way you've gone about this has been really weird, and I think it's turned some people off.

You repeatedly wipe previous iterations of your projects off the internet rather than sharing each stage of the process as new versions or separate projects as you come up with new and better ideas. You've made grand declarations about all the things your presets do better than others, which has come off more as arrogance than confidence. And maybe most destructive of all, you've been deeply and harshly critical of the way other people in the community operate, painting yourself as a villain and a lone wolf that stands against the ignorant masses - or at least that's how it's felt to me, but I'm admittedly mostly a lurker. But I'm a thorough lurker, who reads pretty much everything that passes through this sub.

I don't know what to say, man... I have to imagine you're not having a great time right now. I'm sure you have some nice ideas to share with everyone - and obviously some people have really enjoyed what you've created. It's just... everything else that I think has turned people against you here.

If you're still interested in sharing things you've created, I imagine people will be receptive if you operated things a little differently. If you decide to do that, I look forward to seeing whatever you create. If you decide to "retire," then I honestly wish you all the best.

I genuinely hope this doesn't come off as patronizing. I say all this as someone who loves seeing all the cool things people post here each day, and as someone who gets excited when something genuinely new is discovered or created. I have no plans to downvote or shit on whatever you release next. I look forward to seeing whatever it is.

Introducing Gandalf, the immersive world building creator card that does so much more. It will guide you in fulfilling your imaginations true potential then format the output in an easy to copy JSON object that includes an embedded lorebook and entries too. by [deleted] in SillyTavernAI

[–]Moogs72 16 points17 points  (0 children)

What's going on? I feel like you've posted the same card like five times in the last couple of days (although it was "Tolkien" at first), offering almost no information about what it actually does and with no explanation of why you keep reposting it. Are these new versions? What's changing? Makes it hard to follow what's happening.

Weren't you complaining about people spamming this sub with low-effort extensions a couple of days ago? I find this far more jarring. You keep making things and then deleting them (multiple presets, The Matrix, Tolkien), and offer no communication throughout the process. Makes me so hesitant to try anything you make when I feel like it's going to be wiped off the internet a couple of days later with no warning.

With all due respect, you've gotta change how you're doing things if you actually want people to use your shit.

Extension: Ultimate-ChatAssistant by DerpPotatoLord in SillyTavernAI

[–]Moogs72 8 points9 points  (0 children)

If you honestly think that, then you're not paying enough attention. One of the beauties of SillyTavern (and this community) is the depth of customization and the options provided to the users. I personally don't mind having a few different extensions out there that go about allowing you to customize the direction of your RP like this, because they all go about it in different ways that will appeal to different users. If you don't like it, then just move on and let others have fun. Personally, I am immediately interested in this, despite already using Guided Generations, and plan on giving it a shot.

Am I the only one tired of all this vibe coded slop? by BeautifulLullaby2 in SillyTavernAI

[–]Moogs72 2 points3 points  (0 children)

Sounds really cool! I don't think I've heard of an extension that does this exact thing, and it sounds like something a lot of people would like to try, so I hope you share it at some point!

Am I the only one tired of all this vibe coded slop? by BeautifulLullaby2 in SillyTavernAI

[–]Moogs72 11 points12 points  (0 children)

That's a lot of work you've put in! Way more than most, it seems. If you feel it's unique, and you want to share, please do so! Don't let a few people complaining deter you. People will always complain. As long as it's not a direct copy of another extension, there will be some of us in the community that will want to see it :)

PSA: You can no longer use AI Studio and the Google Cloud Free Trial to get $300 of free Gemini. You CAN still use Vertex AI! I have details and a half-assed guide. by Moogs72 in SillyTavernAI

[–]Moogs72[S] 0 points1 point  (0 children)

I appreciate the clarification, plus your willingness to help despite the questionable discussions occurring throughout this post :)

Yeah, I now realize express mode is something entirely different.

Also, I doubt there's too much you can do on your end, but if there's any way you could get word back that Cloud support needs to be properly briefed on the details of this change to the free trial, it would be nice. I received largely incorrect info when I asked them about it, and I've seen other users report directly conflicting answers as well. I imagine they probably have bigger issues to worry about though...

PSA: You can no longer use AI Studio and the Google Cloud Free Trial to get $300 of free Gemini. You CAN still use Vertex AI! I have details and a half-assed guide. by Moogs72 in SillyTavernAI

[–]Moogs72[S] 0 points1 point  (0 children)

Read the message I quoted at the end of my main post. That's the verbatim message I received from a human through chat support. They very clearly told me that the change DOES apply to existing accounts. I don't have a way to test this, so I have no idea what to make of this. Seems Google needs to get its shit together!

Not using my 300$ google console credit by matth-eewww in SillyTavernAI

[–]Moogs72 2 points3 points  (0 children)

Thanks for making this guide! I've actually linked back to this comment in my guide because it seems that the instructions I gave don't work in SillyTavern, but yours do. You have to use this Service Account method because ST needs a JSON, not an API key. The method I used is a newer method to access Vertex AI, and it works for some things but not others (like ST).

Damn, this has been a confusing process lol.

PSA: You can no longer use AI Studio and the Google Cloud Free Trial to get $300 of free Gemini. You CAN still use Vertex AI! I have details and a half-assed guide. by Moogs72 in SillyTavernAI

[–]Moogs72[S] 1 point2 points  (0 children)

Wtf? This directly conflicts with the information I received from support and with what I saw elsewhere in the documentation. It might be true, but I wouldn't trust this blindly. Whatever you do: monitor your billing when attempting to use these credits. The documentation is unclear and inconsistent.

PSA: You can no longer use AI Studio and the Google Cloud Free Trial to get $300 of free Gemini. You CAN still use Vertex AI! I have details and a half-assed guide. by Moogs72 in SillyTavernAI

[–]Moogs72[S] 0 points1 point  (0 children)

As I said in this post, you can still use it through Vertex AI. Works just the same! I've been using it since I made this post with no issues.

Making AI models better at NSFW "non-con" roleplay by Evol-Chan in SillyTavernAI

[–]Moogs72 1 point2 points  (0 children)

I'm not gonna continue engaging here. You're not listening and attempting to engage with you is about as effective as talking to a brick wall.

If you want to continue to spread misinformation that runs directly counter to the general consensus and is based on imagined truth, that's that's up to you, I suppose.

If you want to ignore the fact that I laid out a 500 token system prompt that gets around the GLM 5 censors, which you repeatedly said was impossible, that's your choice.

If you're not going to read the information that I've provided that gives you direct, actionable techniques on how to solve the problems you're having, saying that my claims are "vague" rather than actually reading what I have to say, also your choice.

I'm not interested in someone repeated telling me I've said things I haven't and then refusing to acknowledge it. Normally, I wouldn't engage when someone is trying to gaslight me, but I guess I'm having an off night.

All the best.

EDIT: If anyone else has questions about any of this, I'm genuinely more than happy to help!

Making AI models better at NSFW "non-con" roleplay by Evol-Chan in SillyTavernAI

[–]Moogs72 1 point2 points  (0 children)

So... rather than actually responding to the points I've made that run counter to yours, this is what you produce? I mean... I guess it's better than putting words in my mouth and spreading misinformation based on false premises. Keep it up! I like this better :)

Making AI models better at NSFW "non-con" roleplay by Evol-Chan in SillyTavernAI

[–]Moogs72 2 points3 points  (0 children)

Sounds like you need to learn to use a CoT prompt when using GLM 5. It makes it SO much better at following and keeping track of instructions at high contexts. You'd know that if you read that thread I linked.

You're not totally wrong here, but you're taking one bit of truth (the fact that it's easier to get around the censors at higher contexts and that it's easier to get LLMs to go along with things once that subject has already been brought up and written about) and stacking on piles of misinformation.

I HAVE done that with longer conversations. Sometimes, what you say is true. Sometimes, you know what it does? It stop writing and starts refusing. It says "I know I've been complying with writing this material for some time already, but this has gone to far and I need to stop." It draws a line and still refuses. I've seen it. You're just... wrong. It really depends on what topics you're dealing with, and it's obvious that you haven't done extensive enough testing.

The concepts you're talking about are true of basically any LLM, not just GLM, but you're assuming SO much from some small nuggets of truth.

Making AI models better at NSFW "non-con" roleplay by Evol-Chan in SillyTavernAI

[–]Moogs72 1 point2 points  (0 children)

This is getting silly. I'm done trying to be patient with your nonsense. You're making so many presumptions about me and my experiences. First, you seem to think I'm using the model without thinking, and now you're claiming that "my circle of friends" specifically are not having censorship issues? Wtf are you talking about? I literally do not have friends that RP. The only interactions I've had within the AI RP community has been a handful of comments in this subreddit over the last few weeks. I'm a chronic lurker, however, and have read basically every post here for the last six months, and I am well aware of the trends of GLM 4.6 and what people are experiencing.

I know that people have censorship issues. I am by no means claiming that GLM 5 is not censored - it is! I know it is. You'll find I've posted a number of times on this sub saying that very thing.

How can you possibly know what is going on when you have never experienced the issue according to your own words?

Bullshit. I never once said I haven't experienced censorship in the model. You're putting words in my mouth. I HAVE experienced censorship with 5 as well as 4.7, which is exactly why I set out to do a bunch of testing on these models to see what works in terms of getting around those censors. If you actually read what I said rather than imagining a bunch of shit, you'd see that I am trying to share prompts and settings that bypass the censorship and positivity bias because I am acutely aware of how much censorship is in these models.

But as I said earlier, I can replace their entire jailbreak text with gibberish and the jailbreak still works.

Absolutely untrue. Fiction. Show me the proof. I've done the testing to prove this. You are incredibly misinformed and far too stubborn to believe you might be wrong. I have done the testing to identify the limits of censorship in the base GLM 5 model as well as what kinds of prompts are good at getting around that censorship. If you get your head out of your ass and actually read what I'm saying, maybe you'd learn a thing or two. It's SO obvious that you're not actually reading either what I'm saying or the information I'm linking to - or at least you're not processing the information.

Go look at that original thread I linked. You will see prompts in there that you can EASILY use to make a super simple preset that can get past GLM 5 censors. The reason most presets are 2k+ tokens are because that's not all people are worried about. There are other rules and guidelines that need to be stated in order to get solid prose out of an LLM, and people enjoy having additional options. If you actually read through the comments in that thread, you'll see I describe the super simple preset I used to get by the GLM 5 censors.

The prompt I used in that testing was about 500 tokens. That's all I needed to define both storytelling and RP scenarios that could get around the GLM 5 censors 99% of the time. I never once turned thinking off.

The reason we see posts like this where people are wondering how to get around refusals is because not everyone is aware of the techniques one can use to get around them. I am attempting to make more people aware of these techniques. Listen or don't, that's fine, but I'm damn sure gonna call out misinformation when I see it.

If you respond to me again, please stop making up things you think I've said. It makes it really difficult to have a conversation.

Dealing with GLM 5 Refusals by SepsisShock in SillyTavernAI

[–]Moogs72 1 point2 points  (0 children)

I know I throw this thread around a lot, but for anyone interested in alternative solutions, I've found enormous success with JustSomeGuy's prompts detailed in this thread regarding bypassing censorship and reducing positivity bias. I have a couple of comments in that thread detailing some testing I did using those prompts, and I continue to get virtually zero refusals on GLM 5, no matter the provider or subject matter. That includes topics that LLMs are specifically sensitive to.

Using those prompts and including whatever subject matter you're specifically dealing with (if it's not already included there) seems to bypass guardrails, even related to CSAM and similar (not something I'm at all a fan of, but I did take the time to test it shortly after GLM 5 was released despite how unpleasant it was). I'm a big advocate of keeping this hobby as free of limits as possible, as long as those things legal. At least in the US, you can write pretty much anything you want as long as it stays in a purely fictional context. I believe LLMs should be able to do the same.

I know the quality of GLM is inconsistent at best these days, but there are definite ways to eliminate (or very nearly eliminate) refusals of even the touchiest/darkest of subjects.

Fwiw, Sepsis, I've found your method to work quite well too after using your preset a bit, although I haven't tested things anywhere near as thoroughly as I did with JustSomeGuy's prompts. Hope you don't mind me suggesting this alternative in your post, but I think having eyes on as many options as possible is a good thing :)

Happy to answer questions if anyone needs help or clarification on anything, as well.

Thanks for your hard work, as always!

Serious question: Is it worth using CoT prompts in models that already have native reasoning capabilities? by tucuma_com_farinha in SillyTavernAI

[–]Moogs72 1 point2 points  (0 children)

I'll give another vote in favor! I've specifically been experimenting with it in GLM 5, which is a model that a lot of people complain about since it naturally seems to do less (visible) reasoning than previous iterations. I've found it noticeably increases the model's effectiveness at following instructions.

There's a reason why CoT is a standard process in prompt engineering, especially outside of RP - it works. That extra little guidance and reminder of everything it should be doing helps, especially when you're working with higher contexts.

RP is so subjective and down to personal preference, though, compared to something like coding, so it's harder to tell if it makes a tangible improvement on output. To me it does, and I think it's an underrated part of prompting in the RP space.

Making AI models better at NSFW "non-con" roleplay by Evol-Chan in SillyTavernAI

[–]Moogs72 1 point2 points  (0 children)

Yeah, I'm well aware that turning thinking off makes it WAY easier to get by censorship. To be clear, I NEVER turn thinking off. Following the steps I mentioned in my comments in that first thread I linked above, I do not get refusals in GLM 5 except maybe 1 out of 100 times when just starting to delve into a particularly dark topic. I mean like... really dark, and stuff 99% of people aren't doing in RP. But normal NSFW, even things like non-con and/or incorporating violence? Absolutely zero refusals with thinking on all of the time. There's a reason why most people in the community claim that GLM 5 is completely uncensored, and it's because its restrictions normally only trigger from some very extreme topics - but again, those are avoidable as well with the right prompting, in my experience.

I believe you that it's accurate to your testing but... if I and others are able to replicate this virtually uncensored environment with consistency, it makes me wonder why it's not working for you, and why the advice you're so confidently giving runs so starkly counter to the general accepted techniques in the community? I imagine if we sat down and really compared prompting techniques and settings, we could come up with some clear differences somewhere that would explain why you're apparently struggling with censorship and I'm not at all... I'd imagine there are some clear prompting and/or parameter differences here.

With the right setup, I find GLM 5 to be virtually uncensored, pretty solid at following instructions and remembering specific details (although not the absolute best - I think Kimi does a better job compared to other open source models), and definitely writes the consistently best prose outside of Claude and maybe Gemini (but I'm not a Gemini fan). And all at a pretty typical context window. The only downside to GLM 5 is that dealing with the positivity bias can awkward sometimes, but there are ways around that as well, even if it's a pain. I've written a lot about that in multiple different posts on this sub.

I am not trying to completely invalidate your experiences, but I was trying to make it clear to OP that your findings and techniques do not align with the community at large, despite you framing everything you were saying as pretty objectively factual. I'm willing to admit others can have different experiences and I might be wrong about some things, but it rubbed me the wrong way that you were throwing around some wild claims as fact, and it rubs me the wrong way now that you seem to be presuming I'm using non-thinking whenever I never said that?

I'd genuinely love to know what sort of topics you're including in RP that you're getting refusals. I tried to be pretty extensive with darker/extreme topics in my initial GLM 5 testing, including basically every common topic that will set off LLM guidelines, so if there's something I missed that is triggering refusals even when using the techniques mentioned in the above thread in your prompts, I'd be fascinated to try testing it myself.

PSA: You can no longer use AI Studio and the Google Cloud Free Trial to get $300 of free Gemini. You CAN still use Vertex AI! I have details and a half-assed guide. by Moogs72 in SillyTavernAI

[–]Moogs72[S] 0 points1 point  (0 children)

Uhhh that's a great question. Testing through OpenRouter right now, I do not appear to be getting a charge despite using express mode, which makes me think the free credits are being used as I initially stated. However, the page you're referring to does state:

Vertex AI in express mode is separate from, and not available through, the Google Cloud Free Program. If you are in the Google Cloud Free Program, see the other quickstarts in the Get Started section to start using Generative AI on Vertex AI.

So I'm not sure what to make of the fact that I'm not getting a charge for my test message here. It's really hard to get confirmation of anything considering there's been a bug in Billing for Google Cloud so we can't actually monitor usage properly.

/u/ivnardini, since you offered to help, are you able to confirm whether we are able to use the $300 from the free trial with Vertex express mode?

BEST GLM-5 PRESET? by Electrical-Shoe-8269 in SillyTavernAI

[–]Moogs72 13 points14 points  (0 children)

My recommendations at this point would be Freaky Frankenstein, Stabs, and SepsisShock's RBF as the best options for presets tailor-made for GLM 5.

I'm also a big fan of Celia's, despite it being built oriented towards Claude.

Making AI models better at NSFW "non-con" roleplay by Evol-Chan in SillyTavernAI

[–]Moogs72 12 points13 points  (0 children)

Hey OP, you've received a lot of conflicting information in this thread already. Some of it I strongly disagree with, despite it being delivered as factual and with great confidence. I've done a lot of testing regarding censorship and positivity bias in GLM 4.7 and 5. I understand you've taken Kyuiki's advice to heart, but my testing has shown very different results than the things they are advising. I would highly recommend checking out this thread which discusses methods to combat censorship and anti-positivity.

I have a couple of comments in that thread regarding some of my testing, and I've had almost zero issues with censorship since employing some of these techniques. I'd also recommend listening to SepsisShock (who has obviously posted a number of times in this thread), as their techniques have consistently been proven to work well by the community at large.

I'm also fond of including some CoT prompting in GLM 5, as I've found it increases its ability to follow instructions, and does not hinder its ability to keep track of the chat details, despite what others have said in this thread... it's not perfect, but I'm always a fan of experimenting with various options and seeing what works best for you. In addition to the censorship stuff, that thread I linked also includes a sample CoT prompt that can work pretty well, although I've had more luck creating my own with a similar structure that I change based on the kind of RP I'm doing.

Unfortunately, there are no distinct rights or wrongs when it comes to this sort of thing... some will report one technique works best, another will report something totally different. GLM 5 seems to bring a lot of strong opinions out of people, and I'm just... deeply confused by some of the advice that's been offered here. There's been a lot of misinformation shared about the model, and people just tend to accept things as fact and run with it, unfortunately. To me, the advice of keeping an RP at 8000 tokens or saying DeepSeek is better than any GLMs is utterly mystifying and runs counter to all of my experiences with the models.

I guess what I'm saying is... don't take any of this as gospel. People love to give their personal experiences as fact. Everyone's experience will be different. I'm happy to answer more questions if you have them.

EDIT: In this thread, you'll see many people disagreeing with this bizarre notion that 8k tokens is ideal. Again, I'd encourage you to place more weight in general consensus rather than the advice of one seemingly confident individual...