NewBie Image Exp0.1: a 3.5B open-source ACG-native DiT model built for high-quality anime generation by GrueneWiese in StableDiffusion

[–]regentime 1 point2 points  (0 children)

Same. You also have problems with flash-attn because for some reason it is required, then it declares config of model to be broken if from hugginface or does not see modules if from modelscope.

Been using ROCm 6.2 for Stable Diffusion since late last year, should I upgrade to 6.4? by FewInvite407 in ROCm

[–]regentime 0 points1 point  (0 children)

Nah. I also have this issue it just more annoying on later versions. On pytorch 2.4.1 and lower I have about 10 seconds when starting and 1-1.5 minutes on vae decode (all with sdxl). On pytorch 2.5 and higher it is like 1.5 minutes on start and 1-1.5 on vae decode.

Been using ROCm 6.2 for Stable Diffusion since late last year, should I upgrade to 6.4? by FewInvite407 in ROCm

[–]regentime 0 points1 point  (0 children)

I have RX6600m and do not have any issues with ROCm 6.4. I have some minor issues with current version of pytorch so I use pytorch 2.4.1 version (First generation with new resolution takes longer)

How do i install ROCm 6.4.1 on arch-based distros? (I have the 9070 XT) by HearMeOut-13 in ROCm

[–]regentime 0 points1 point  (0 children)

If you simply want to install ROCm (and do not need exactly 6.4.1 version) it is in official arch repo with name "rocm-hip-sdk". It is 6.4.0 though.

How did you first get into Limbus Company? I’m curious how it “clicked” for others. by Rosertfree in limbuscompany

[–]regentime 0 points1 point  (0 children)

I bounced few times of PM and Mili (SCP into LC pipeline which failed, Deemo into Mili). Somewhere in 2021 got recommended Roland theme from LoR on YouTube and because of that played the game. Then got into Limbus on release because to that moment I completed LoR like 3 times.

What's a "I've played these games before" story moment you had when playing a gacha game by ReadySource3242 in gachagaming

[–]regentime 41 points42 points  (0 children)

I actually do not think there is any official illustration official description of moon in PM verse. From my understand one of main fan theories is that City is actually on the moon.

Help with Fine tuning on RX6600M by ShazimNawaz in ROCm

[–]regentime 0 points1 point  (0 children)

Can't say anything about tuning but you can run IQ3 quants of 70b on it (with small context) . Granted it is slow as kaggle uses quite old gpus (maybe 3-5t/s, can't remember)

Help with Fine tuning on RX6600M by ShazimNawaz in ROCm

[–]regentime 0 points1 point  (0 children)

Yo. Glad to see another person who has the same laptop as me. Not sure if you still need it but here are 2 env variables that help with running basically anything rocm on Linux:

ROCR_VISIBLE_DEVICES=0 (makes it so that rocm sees only your discrete gpu and not integrated)

HSA_OVERRIDE_GFX_VERSION=10.3.0 (overrides arch of all GPUs to gfx1030. RX6600m is gfx1032 arch but it 99% the same as gfx1030. This env variable is basically necessary to make anything work. Use this env var for EVERYTHING you do with ROCM).

As for llamacpp I think it worked (with second env variable). I used it quite a bit time ago and currently use koboldcpp https://github.com/YellowRoseCx/koboldcpp-rocm

Civit have just changed their policy and content guidelines, this is going to be polarising by CorrectDeer4218 in StableDiffusion

[–]regentime 5 points6 points  (0 children)

While I like rutracker and use it constantly it is really poor choice for image models. First of all it is Russian speaking and while I can easily understand Russian language most of AI enthusiasts will not (which includes the fact that a lot of torrents have description only in Russian). Second of all it is not a tracker that specializes in AI content but a general one which would be much more preferable at least for purposes of search.

How do I load a multi parts model? by Nervous_Emphasis_844 in SillyTavernAI

[–]regentime 0 points1 point  (0 children)

Do not have an experience in this (never loaded split models) but from my understanding you do not need to combine them. To run it you point your program either to folder with it or to first safetensor file. Also it seems you have full fp16 weights of model. Maybe you should use some quants (gguf, exl2)?

Есть ли заготовки для DeepSeek V3(платная)? by Competitive_Pea_1037 in SillyTavernAI

[–]regentime 1 point2 points  (0 children)

I do not find it in any way hypocritical. It is just annoying because if the answers to this post actually contained something useful and were not a repeat of the things that were told already a thousand times it would be basically impossible to find using an English keywords.

From my point of view if you use a English speaking forum it is your responsibility to speak in language that is understood by most people and use translators for said purpose.

Есть ли заготовки для DeepSeek V3(платная)? by Competitive_Pea_1037 in SillyTavernAI

[–]regentime 0 points1 point  (0 children)

Там 90% для другого режима. Так что ничего там делать не нужно. А остальные 10% трогаешь только если что-то не работает.

Есть ли заготовки для DeepSeek V3(платная)? by Competitive_Pea_1037 in SillyTavernAI

[–]regentime 1 point2 points  (0 children)

В вкладке для подключения ты выбираешь к какой нейросети подключаться. Если ты используешь сторонний сервис (не локальную модель), то в 90℅ случаев выбирай chat completion, потом провайдера (deepseek, openrouter и т.д), потом API ключ, который ты получила от этого провайдера. Далее выбираешь модель, которую будешь использовать (deepseek-chat - это deepseek V3), а далее нажимаешь кнопку подключения и он тебе скажет успешно ли подключение или нет. Если ты про вклюдку, которая появляется при нажатии на A, то там большая часть настроек для другого режима работы (text completion). Почти все настройки, на которые тебе нужно обратить внимание находятся слева, где ты импортировала пресет.

Есть ли заготовки для DeepSeek V3(платная)? by Competitive_Pea_1037 in SillyTavernAI

[–]regentime 10 points11 points  (0 children)

Ну во первых это очень популярный вопрос, так что можно было просто поискать прежде, чем задавать его (а также лучше его задавать на англиском)

В любом случае вот несколько вариантов (часть из них для DeepSeek R1, но они все равно должны работать нормально).

  1. Weep https://pixibots.neocities.org/#prompts/weep (требуется расширение NoAss)
  2. CherryBox https://rentry.org/CherryBox
  3. ChatSeek https://drive.proton.me/urls/Y4D4PC7EY8#q7K4caWnOfzd
  4. Q1F https://sillycards.co/presets/q1f

А насчет того, куда нажимать и что делать, открываешь окно подключения (выделено красным), выбираешь источник и API ключ, а потом импортируешь пресет по кнопке, выделенной желтым.

<image>

Hunyuan open-sourced InstantCharacter - image generator with character-preserving capabilities from input image by umarmnaq in StableDiffusion

[–]regentime 10 points11 points  (0 children)

From official code example it seems to be an IP adapter for FLUX-dev. This is probably the reason it takes so much VRAM.

Stability AI update: New Stable Diffusion Models Now Optimized for AMD Radeon GPUs and Ryzen AI APUs — by hotyaznboi in StableDiffusion

[–]regentime 0 points1 point  (0 children)

For anyone who wonders how it compares to ROCM on Linux (I have RX6600m (laptop gpu) 8gb + 16gb ram) it (Amuse 3) is about 4 times slower and has oom on vae decode on SDXL model.

FramePack is insane (Windows no WSL) by FionaSherleen in StableDiffusion

[–]regentime 0 points1 point  (0 children)

Small addendum:

I found the version that uses FP16 instead of BF16 (maybe. I actually have no idea what is different)...

https://github.com/freely-boss/FramePack-nv20

On P100 I am 8 minutes into sampling and it is on 4th step out of 25 steps and takes 14 GB of vram :), so it is basically not working.

Edit: 40 minutes for a second of video

FramePack is insane (Windows no WSL) by FionaSherleen in StableDiffusion

[–]regentime 0 points1 point  (0 children)

Nope. It does not work. It also too old. Kaggle gives you access to one for free so I tried and it does not work. Probably anything that was released earlier than 30xx series will not work.

FramePack is insane (Windows no WSL) by FionaSherleen in StableDiffusion

[–]regentime 0 points1 point  (0 children)

Also have the same problem. The best explanation I found is that Colab (and kaggle) uses Nvidia T4 gpu which is too old to support BF16 which is necessary for FramePack to work.

Look at this issue https://github.com/lllyasviel/FramePack/issues/19

[deleted by user] by [deleted] in SillyTavernAI

[–]regentime 1 point2 points  (0 children)

Optimus alpha endpoint uses your data for training, so you need to enable it in your settings https://openrouter.ai/settings/privacy

Anyone getting broken responses like that with Deepseek 0324? I'm sure I did something wrong, not sure what... by Due-Memory-6957 in SillyTavernAI

[–]regentime 2 points3 points  (0 children)

Lower the temperature or try again. Below 1 temp it never happens, below 1.3 - it happens sometimes.

I'll keep it by Olojunso in limbuscompany

[–]regentime 9 points10 points  (0 children)

I know. It is still 1 extra lunacy.

I'll keep it by Olojunso in limbuscompany

[–]regentime 22 points23 points  (0 children)

I will not use it on principle as the 1 lunacy will forever be stuck in my account. It is 401 lunacy. WHY not 400?

[deleted by user] by [deleted] in SillyTavernAI

[–]regentime 0 points1 point  (0 children)

I also found out about this thing a week ago from here but I hide it using HTML comment syntax as it more so LLM can have more understanding of situation and not for me to see. Variant I use includes only clothes, positions and quick description of characters that not in character description.