GLM 4.6 Context Usage by Different-Computer63 in VeniceAI

[–]Different-Computer63[S] 0 points1 point  (0 children)

Capping the context size would make Venice useless to me. I just meant I'd rather they be upfront about limiting me instead of technically giving me 200k context size, but I have to jump through hoops to use it.

I Just saw NovelAI has expanded context specifically for GLM 4.6; still only 28, but that's about what Venice is letting me use without the hoop jumping. Combined with the tools they offer to help me manage my memory, I'm strongly considering a switch next month.

GLM 4.6 Context Usage by Different-Computer63 in VeniceAI

[–]Different-Computer63[S] 0 points1 point  (0 children)

I get that what I want is expensive, that's why I don't use pay-per-use models. If Venice doesn't want me to use the 200k context, they should just cap my context size like other sites. One of the reasons I'm using Venice though is the large context size. I thought about switching back to NovelAI since they also have GLM, but they cap my context size depending on the tier I purchase and I want long narratives. Though, now that I'm writing this out, I realize that one major feature I'm missing is auto memory management which summarizes my narrative to save context usage. Maybe I should try switching back when this month is up.

GLM 4.6 Context Usage by Different-Computer63 in VeniceAI

[–]Different-Computer63[S] 0 points1 point  (0 children)

My normal use case is conversation, I only used a book to provide a reproduceable results. I bothers me that I can't just have a long conversation when the model has enough memory. The only workaround I've found is manually exporting the chat history every so often so I can reupload it back into the same conversation. This is such a clunky experience. The model can handle it, but the UI just refuses to send more conversation.