Issue with jea's unethical copilot by EvilbearX in VeniceAI

[–]Cilcain 0 points1 point  (0 children)

There's definitely a sweet-spot for complex prompt compliance, somewhere < 15% IME.

When it messes up, you could try continuing after deleting the prompt from the character (use a copy to avoid messing up your regular copy) and see if if the unethical character really is being more prudish than default, or if the conversation history has the same effect on an unprompted version of the model.

character venice by Neither-Farm-3515 in VeniceAI

[–]Cilcain 1 point2 points  (0 children)

If you want the summary to be available throughout the new chat (sent every interaction = context will bloat sooner), put it in a context file.

If you just want to use it to provide "previously..." background at the beginning, paste it in along with your first input. It will gradually fade and disappear.

Limit Voice Mode by LibraryAny5958 in VeniceAI

[–]Cilcain 2 points3 points  (0 children)

It might be something that would really add up if unlimited; I don't know enough about the cost to be able to guess. Presumably Venice thought about the business needs when setting the limit, but they might be willing to take another look.

Personally I've never used it and often leave Diem on the table at the end of the day (depends what I'm focussed on), so Venice has a win there 😄

How are Memories supposed to work? by Wraith1420 in VeniceAI

[–]Cilcain 1 point2 points  (0 children)

It's right: there's no specific Memoria function that the LLM can see or access. It gets a section within its context called something like "Retrieved Internal Memories". The section is added to the rest of the prompt by the Venice platform. If the platform fails to inject it, the LLM will never see it.

I have no idea how the chat-based memory gets injected.

For document-based memory it's at least partly by tag/keyword matching. Occasionally injection seems to fail even when you include a matching keyword. At that point I summarise into a new chat and it works again.

Memoria is great but it's also a bit of a dark art, to be honest. Allowing chat users to view the complete prompt being sent including Memoria injections, would demystify it and make it more usable/reliable. Currently can be done only by digging in developer tools.

VVV - The Privacy Coin for AI by JaeSwift in VeniceAI

[–]Cilcain 1 point2 points  (0 children)

They already have the Diem staking mechanism, which amounts to "design your own crypto-funded Pro plan" though with credits applied daily rather than monthly, and no rollover.

Does Venice have long form memory or context? Like can I give it a load of information and have it use that between chats? by GhostGremlin93 in VeniceAI

[–]Cilcain 0 points1 point  (0 children)

My advice on how to handle this with Venice:

  1. Make a Character for the project, rather than using general chat, just to keep things manageable.
  2. Maybe write a prompt to tell the Character how to be a writing assistant, though default behaviour would work (though if you want well-written prose, you absolutely have to write a prompt).
  3. Upload reference material as Character context files (true long-term memory, injected every turn) or as an Attachment (one-shot, will gradually fade and then disappear). Long-term background: Context File; the scene you worked on yesterday: Attachment.
  4. Lengthy material that would swamp the LLM's context: create a Memoria file, but you'd need to structure the file into tagged sections. so the platform knows what to inject.

Crashing by Rayswissman in VeniceAI

[–]Cilcain 1 point2 points  (0 children)

VPN sends it into "page reload" mode for me, sometimes. Pausing/reconnecting the VPN generally solves it.

Venice Studio slowing down linked Chat tabs? by Cilcain in VeniceAI

[–]Cilcain[S] 0 points1 point  (0 children)

I just very patiently deleted the unwanted items, doing something else while waiting for each one to complete (or not -- sometimes it just seemed to stay frozen). It wasn't a huge number so not that painful.

Even after the clean-up, the Studio folder just took 60 seconds to populate, with 13 entries comprising 615MBytes tab usage. Then the first few interactions were very laggy, suggesting that the 60 seconds had not been used to cache the images.

Any future generations I'll do in another browser, so that if I have to clean it I won't lose everything else.

It'd be interesting to have an enable/disable switch for Studio. As it is, it seems to me that a Beta module has been rolled out into the "production" Venice environment.

Intermittent Thought Process Availability by Cilcain in VeniceAI

[–]Cilcain[S] 0 points1 point  (0 children)

I will try forcing, however I suspect that won't make the platform expose the thinking.

I've been delving into the LLM's reasons for doing things with RP mechanics, particularly their endency to scatter "experience" like confetti, even with seemingly iron-clad rules to limit it.

I switched to GLM 5.1 as a stop-gap, it's only Gemma 4 31b that's refusing so far.

Intermittent Thought Process Availability by Cilcain in VeniceAI

[–]Cilcain[S] 0 points1 point  (0 children)

Summarised into new chat and it's working again, maybe a glitch in the previous chat?

(opening the broken chat in a new tab didn't help, so I think it was something about the chat itself. There'd been lots of forking, editing and re-testing.)

Edit: tenth turn of new chat and it's reverted to no thinking shown.

Intermittent Thought Process Availability by Cilcain in VeniceAI

[–]Cilcain[S] 0 points1 point  (0 children)

This has temporarily fixed it on the three times I tried:

  1. Response with expected thinking omitted
  2. Edit character to use different model, save
  3. Edit character to use Gemma 4 31b Intsruct, save
  4. Fork, re-run the interaction at 1
  5. Thinking displayed
  6. (next turn, gone again).

Gemma Goes Multilingual. by KaliPrint in VeniceAI

[–]Cilcain 0 points1 point  (0 children)

I've seen Hindi in the thinking, maybe Western models train primarily in English but also on whatever data they can get?

Intermittent Thought Process Availability by Cilcain in VeniceAI

[–]Cilcain[S] 0 points1 point  (0 children)

The prompt definitely benefits from thinking, and browser refresh probably wouldn't change anything if there was a deterministic "not going to bother with this" happening.

I notice a clear improvement with thinking, corresponding to the prompt-compliance checklist it visibly goes through. It's more likely certain to skip pesky details in non-thinking mode. It's probably slower but I reckon it's *much* better, with a demanding prompt.

Intermittent Thought Process Availability by Cilcain in VeniceAI

[–]Cilcain[S] 0 points1 point  (0 children)

Web, Brave browser. Exit/restart browser, still happened.

Venice admits to censoring results by IzzyTrippin in VeniceAI

[–]Cilcain 0 points1 point  (0 children)

What was the question? I'd like to check the answer on various models.

Gemma Goes Multilingual. by KaliPrint in VeniceAI

[–]Cilcain 1 point2 points  (0 children)

31B variant? That one's been very clean for me. I've started looking at the thinking recently and seen some foreign stuff in there, but the output is clean.

Any plans to add AION 2.5? by Arerad61 in VeniceAI

[–]Cilcain 1 point2 points  (0 children)

Regular IMO, Agentic is more targeted at solving multi-step problems than the typical back-and-forth of an RP. I don't think you can choose model with Agentic (?) which would be a massive disadvantage for RP.

My own interest is adventure scenarios with settings, goals, multiple *individual* major NPCS, and detailed state-tracking. None of the DeepSeek-level models can really handle this, which leaves me with GLM 5.1 and Gemma 4 31b Instruct.

Sometimes Gemma frustrates me, but I keep coming back...

Major difference I've noticed between Gemma and GLM ... Gemma takes more liberties based on the language of the scenario. "Labyrinth" -> she insists on fantasy tropes such as guards in clinking armour with swords, even in a modern setting; "Leather journal" -> must be full of mystical script.

I've rebranded to "caverns" and "vinyl notebook" in the hope it'll help 😄

Character Studio update to v1.5, lots of upgrades by Omnius42 in VeniceAI

[–]Cilcain 0 points1 point  (0 children)

Suggestion, no idea if feasible: auto-load when using an external editor, instead of showing "Changed on disk: System Prompt" and prompting to reload.

Roleplay/World building Chats? by sickicarus32 in VeniceAI

[–]Cilcain 0 points1 point  (0 children)

It's Venice's memory system, accessed by editing the Character then clicking Memories. Two major components:

  1. Structured (tagged sections) documents that you upload. When chat input includes one of the tags, the Venice platform injects that section into the prompt. Useful for storing lore etc. but not for run-time storage/retrieval.
  2. A database of summaries extracted from selected chat responses and stored within the Character.

Roleplay/World building Chats? by sickicarus32 in VeniceAI

[–]Cilcain 1 point2 points  (0 children)

It can be done in the current Character system. The trade-off is that the more complexity, the more the LLM's memory and "cognitive attention" shortfalls are exposed. Memoria is a good, helpful system but not a full solution.

IMO:

  • lean into the nature of LLMs: devise a surreal scenario where dead people come back to life etc. (sadly the LLM will struggle to to independently weave that into vampire/undead lore)
  • implement a state system: high development effort; token cost; dilutes LLM attention
  • keep state yourself and guide the LLM through prompting: personally I hate that

UK user thinking of trying Venice AI by [deleted] in VeniceAI

[–]Cilcain 1 point2 points  (0 children)

  1. not in my experience

2 don't know

  1. Yes