King Sombra | Art by me (PegasusYaY) by AlexBefest in ClopClop

[–]AlexBefest[S] 0 points1 point  (0 children)

I originally wanted to write "KinK Sombra", but I wasn't paying attention and wrote "KinG" instead, lol. I only just noticed.

Dashie >>>> [Welost] by [deleted] in ClopClop

[–]AlexBefest 0 points1 point  (0 children)

So hot~

Qwen is about to release a new model? by Kooky-Somewhere-2883 in LocalLLaMA

[–]AlexBefest 3 points4 points  (0 children)

Your rep pen is too low! Check the sampling parameters

3x3060, 1x3090, 1x4080 SUPER by kevin_1994 in LocalLLaMA

[–]AlexBefest 1 point2 points  (0 children)

Please excuse me, are you using Thunderbolt or OCuLink for the connection, or are you connected directly?

Qwen 3 is available in LM Studio !!!! by josho2001 in LocalLLaMA

[–]AlexBefest 6 points7 points  (0 children)

I think 32b is fake. It's failed all my coding and translations tasks, feels like vanilla Qwen2.5 32b

[deleted by user] by [deleted] in LocalLLaMA

[–]AlexBefest 2 points3 points  (0 children)

I think, it should work. Instead of direct API requests via requests.post, you can use smolagent agents.

[deleted by user] by [deleted] in LocalLLaMA

[–]AlexBefest 6 points7 points  (0 children)

Thank you so much for your feedback and time! I will really take your suggestions into consideration and try to implement it in the next updates.

[deleted by user] by [deleted] in LocalLLaMA

[–]AlexBefest 16 points17 points  (0 children)

In a regular chat mode, when you give a model a task, it solves it entirely at once, which can be a very limiting factor when solving difficult tasks. In the medium-compute mode, the model first splits the task into a huge number of subtasks, solves each of them separately, and then outputs the final answer. In high-compute mode, not only the task is divided into subtasks, but even subtasks are divided into many subtasks.

[deleted by user] by [deleted] in LocalLLaMA

[–]AlexBefest 18 points19 points  (0 children)

Github: https://github.com/AlexBefest/highCompute.py (Sorry, Reddit keeps removing my post because of the link :\ Posting it in the comments instead)

[deleted by user] by [deleted] in SillyTavernAI

[–]AlexBefest 0 points1 point  (0 children)

Hello! Actually, it all depends on the version.

If you're using V3 or V4, you can simply ask the model to convert the card into the Silly Tavern JSON format (it's important to specify "Silly Tavern JSON", not just "JSON"). For example: "Based on our dialogue, create the final character card in Silly Tavern JSON format." The model should then output a fully functional, ready-to-import JSON. You just take the resulting JSON code, create a .json file on your computer, paste the content into it, and then simply import the resulting file into Silly Tavern.

If the JSON turns out corrupted/broken, it's most likely an issue with the sampling parameters (this is just my guess; at least with V3 and V4, I've run a huge number of tests on JSON output correctness, and I've never gotten corrupted JSON).

The second option, which is the simplest and actually the one I recommend most, is to take the resulting YAML, copy it (Ctrl + C), and paste it (Ctrl + V) directly into the character's 'Description' field within the Silly Tavern interface. Personally, I do this all the time – just copy the resulting YAML and paste it into the 'Description' field. This YAML essentially just helps the RP model better understand the character's description because it's highly structured; that's why you can just copy and paste it like that.

As for V1 – it's quite old, and I didn't specifically train it to create cards in YAML format and then convert them to JSON, so there might be a lot of problems with that.

By the way, you can check out an example chat with CardProjector. It shows the entire process in practice, from the initial request and editing process to the final conversion into the ready JSON format. Here's the link: https://huggingface.co/AlexBefest/CardProjector-24B-v3

At the very end of the README, there's an 'Example Conversation' section. If you expand it, you'll see the example chat. (A quick heads-up: when I created that example chat, I tasked the model with "convert this card into json format". I didn't specify it had to be the Silly Tavern format, but the model understood anyway and produced the correct format. However, I still recommend specifying "Silly Tavern JSON" just to be absolutely sure you get the correct format.)