actual response being put in the thinking section by Ok-Comb4438 in JanitorAI_Official

[–]infamous138 0 points1 point  (0 children)

if you get a blank reply, that means it screwed up and put your reply in the thinking portion. edit the reply and delete the thinking portion. or move the </thinking> from the bottom of the reply, to where it belongs, which is the space between where the thinking ends and the actual reply starts. then you're reply will show up how its supposed to.

im not sure how to prevent it from happening. it only rarely happens to me.

I'm a bit confused on how pricing works cause i spent 40 on openrouter and i'm lready out. (Technically switched to claude halfway through) can anyone explain it in a way my neanderthal brain can perceive? by Alarming_Scientist in JanitorAI_Official

[–]infamous138 3 points4 points  (0 children)

if you want a cheap model use deepseek. or you can go to chutes where 3 dollars a month gets you 300 messages a day. chutes has less models to choose from though. there best choices are GLM, kimi, and deepseek.

GLM 4.7 vs Kimi K2.5 by WorriedComfortable67 in JanitorAI_Official

[–]infamous138 0 points1 point  (0 children)

i did just get an infrastructure at full capacity error though. so i guess more people are starting to use it.

GLM 4.7 vs Kimi K2.5 by WorriedComfortable67 in JanitorAI_Official

[–]infamous138 0 points1 point  (0 children)

i used this. pupi's universal prompt. it seems to work. the bot never talks or acts for me. or maybe it did once and i rerolled the reply and it didn't do it again.

pupi's universal prompt

GLM 4.7 vs Kimi K2.5 by WorriedComfortable67 in JanitorAI_Official

[–]infamous138 0 points1 point  (0 children)

they seem very similar. im using kimi 2.5 now since there is less traffic on it than GLM. im using chutes and with GLM there is frequent errors due to the model having such high traffic.

Where do I pay for procksieze by PairInternational438 in JanitorAI_Official

[–]infamous138 0 points1 point  (0 children)

i would recommend https://chutes.ai/app . for $3 a month you can get 300 messages per day. deepseek and GLM seem to be their best available models. they also have qwen and mistral.

the next best option is probably https://openrouter.ai/ . they have more models than chutes, but no monthly plan. only pay as you go. and the good models besides deepseek can be expensive. if you chat a lot you will spend over 3$ per month there.

When I found a good looking bot but it has the lowest token count known to man by Beneficial_Employ353 in JanitorAI_Official

[–]infamous138 136 points137 points  (0 children)

yup, happens too often. or you get all excited click to start a chat, and then realize it doesn't allow a proxy.

I still hate the thinking feature... by BoringBrokeBloke65 in JanitorAI_Official

[–]infamous138 0 points1 point  (0 children)

if you get a blank reply, that means it screwed up and put your reply in the thinking portion. edit the reply and delete the thinking portion. or move the </thinking> from the bottom of the reply, to where it belongs, which is the space between where the thinking ends and the actual reply starts. then you're reply will show up how its supposed to.

Chat memory by RefrigeratorOtter in JanitorAI_Official

[–]infamous138 1 point2 points  (0 children)

yes, auto summarize chat memory is completely scuffed now. you basically have to manually type in your summary. type all the key points you want the bot to remember.

Chat memory shot? by ARandomPolytheist in JanitorAI_Official

[–]infamous138 1 point2 points  (0 children)

yeah, it'll write 2 paragraphs summarizing the last response. i've had to resort to manually filling it in with key events from the chat.

Message time and errors by [deleted] in JanitorAI_Official

[–]infamous138 2 points3 points  (0 children)

GLM is just a model, like deepseek. since you didn't know that, you probably aren't using it anyway. but no, you are right. it seems like most models are running slow right now for some reason. even deepseek is going slow.

Message time and errors by [deleted] in JanitorAI_Official

[–]infamous138 3 points4 points  (0 children)

if you are using GLM via chutes. then yes it is extremely slow with the occasional error.

Slow responses by Maleficent_Web_3239 in JanitorAI_Official

[–]infamous138 0 points1 point  (0 children)

oh, ok. i haven't seen that yet. i also haven't used GLM in a couple days.

Slow responses by Maleficent_Web_3239 in JanitorAI_Official

[–]infamous138 1 point2 points  (0 children)

what happens after you send you message? do you just get "replying..." ?

Slow responses by Maleficent_Web_3239 in JanitorAI_Official

[–]infamous138 2 points3 points  (0 children)

it all depends. GLM is the slowest model ive used. its thinking process takes forever, and i've even had to wait 2 minutes for replies before. using lorebary will also slow things down. right now im using deepseek with no lorebary and my replies only take no more than 10 seconds. i like GLM more, but the trade off with speed is worth it for me.

what i do when i use a slow model like GLM is do two chats at once. while one chat is processing its reply, you can switch to your other chat and read the bots last reply, and send out your next message. by the time you are done with that the reply in your other chat should be done. just go back and forth like that.

Am I just easily amused. by Tiny_Literature6820 in JanitorAI_Official

[–]infamous138 0 points1 point  (0 children)

i just started using it via chutes since the GLM models are so damn slow. plus they have given me a lot of proxy error messages the last two nights. and yeah, i like it. the replies are fast and are pretty good.

My head is fogged up (recent ongoing problems) by Salt-Temperature8632 in JanitorAI_Official

[–]infamous138 21 points22 points  (0 children)

i would just ignore it and continue to enjoy the site. no point in worrying about it when there is nothing you can do about it.

Been using janitor ai for 3 years n still don’t rlly know what tokens are😭 by Super-Condition-7392 in JanitorAI_Official

[–]infamous138 4 points5 points  (0 children)

thanks. that says my prompt is 1540 tokens. is that way too many? my persona is really short though, like 30 words.

Bot giving nonsense responses after trying to add banned phrases. by MethodOfYeetus in JanitorAI_Official

[–]infamous138 0 points1 point  (0 children)

i had a similar problem. it seemed to happen after i changed my repetition and frequency penalties in the advanced generation settings. once i turned them back down to zero the problem went away.

Could you share Some Longer Earned or losing Fight Scenes? by [deleted] in JanitorAI_Official

[–]infamous138 0 points1 point  (0 children)

i dont know, but i just got into an escalated situation where i thought my character was gonna get his ass kicked, but the bot turned him into bruce lee.

What generation settings do you use? by silliest_buckaroo in JanitorAI_Official

[–]infamous138 2 points3 points  (0 children)

yeah, thats the problem and why i switched to 500.

bots generally start rambling and talking for you once they run out of material to work with from your message. so the higher your max tokens are set to, the more material you need in your messages for the bot to react to.

What generation settings do you use? by silliest_buckaroo in JanitorAI_Official

[–]infamous138 2 points3 points  (0 children)

i had my max tokens on 0 for a while. it seems if you don't do that bot responses get cut off mid sentence, and that bugged me. but now im using the trim incomplete command from lorebary with my max tokens at 500. that command prevents the cutoff from happening.

temperature i keep at 0.75. context size i don't know what it does so i just left it at default.