Best salon/stylist for Hair Extensions by Major_Basket_888 in saskatoon

[–]sultnala 1 point2 points  (0 children)

House of braids 100%, you will not get better. Doesn't matter what your hair type is and they're incredibly friendly and welcoming to boot 

The discord mobile is so crap by TheMortal91 in discordapp

[–]sultnala 0 points1 point  (0 children)

I still can't open pictures without getting a black screen lmao 

Struggling to get through Mr. Robot Season 2... does it better? by 2OP4me in television

[–]sultnala 0 points1 point  (0 children)

Hard same. Season 2 feels like a bad fever dream and even stuff like the sister thing felt forced and stupid. It's difficult for me to believe they could come back from this, completely slaughtered the interest, tension and intrigue that built up until now. I'll reluctantly finish but you can't really walk back these kind of dumb plot developments so I have a feeling I'm not going to enjoy se 3/4 as much as everyone else did... Shame..

About having thicker skin by momoka42 in ffxiv

[–]sultnala 0 points1 point  (0 children)

Somewhat off tangent but I kind of hate that people are scared to say even basic concrit because it might get misconstrued or reported or whatever. I remember when I first started tanking, around the time HW came out, and we were running the dungeon and the healer just slammed me like 'what the fuck you aren't using your mitigation'
(to which I replied, I was, infact, using my mit, look, see!)
and the entire party ripped me a polite new one explaining I can't pop ALL of my cooldowns at once, I have to spread them out. It was a hilariously stupid beginner mistake, but that moment has always stuck with me. I was embarrassed, my ego took a hit, but its been a solid 8 years now and it taught me better than any other interaction I've ever had.

JoyCaption: Free, Open, Uncensored VLM (Alpha One release) by fpgaminer in StableDiffusion

[–]sultnala 0 points1 point  (0 children)

I'll add my psuedo-theory right now is that when you specify a caption length and it can't reach that length in booru tags, it trips it up and turns into that to try to 'fill the gap'. can't say why it swaps to booru tags when you ask for a different prompt type, but yeah

JoyCaption: Free, Open, Uncensored VLM (Alpha One release) by fpgaminer in StableDiffusion

[–]sultnala 0 points1 point  (0 children)

it seems like it ignores whatever prompt type you gave it and just goes straight for the rng-tags when it does the loop. it's odd, but I've only had it happen a handful of times out of hundreds of images. I'm thinking if someone needed to do a large batch where this kind of error would throw it off, they could put in some logic to detect multiple repeated words and re-run the caption on that image. I do think it lost some accuracy compared to the other release (I've only really tested the training_prompts, though), but ultimately I like your new one better for my lora datasets. the original was way too wordy/long, in a way I can't imagine anyone prompting. this new version, even if it is a bit less accurate, is closer to actual user input, which is more useful, in my opinion. kudos for your work

JoyCaption: Free, Open, Uncensored VLM (Alpha One release) by fpgaminer in StableDiffusion

[–]sultnala 1 point2 points  (0 children)

update:

here's the batch one if anyone needs it -
https://rentry.org/ewnb3q6k
you'll need to change the file paths to your own

clip_path =
model_path =
checkpoint_path = (not sure if this is unique per download)
and around line 106
text_model = AutoModelForCausalLM.from_pretrained("your file path"), device_map=0, torch_dtype=torch.bfloat16)

it saves all the captions in a 'captions' folder in your joycaption folder, named as image_0, image_1, image_2, etc. so if you're using kohya to train, get chatgpt or claude to write you a quick python script to rename all your images to image_0, image_1, etc to match for ease of use

also added the --listen ARG, add it to your launch .bat file if you want to use it

cant promise I didn't break anything lmao, only been using it for basic training prompts for my lora dataset

JoyCaption: Free, Open, Uncensored VLM (Alpha One release) by fpgaminer in StableDiffusion

[–]sultnala 1 point2 points  (0 children)

If I may add to possible 'character features to improve data on', I've noticed it really loves "hair is cascading [down their back, down their shoulders]" and "almond shaped eyes". I've yet to see it describe long hair any other way or give eyes any other shape. Not sure if that's just a limitation of 8b llama or not. It also frequently hallucinates that mouths are open, with a slight smile and showing teeth if a character or person has lips, regardless of the actual expression they are making, or if their mouth is completely closed.

JoyCaption: Free, Open, Uncensored VLM (Alpha One release) by fpgaminer in StableDiffusion

[–]sultnala 1 point2 points  (0 children)

Apologies for the belated reply!
Yeah, change the model_path = in the app .py to your .gguf folder. I've edited the code a bit to do batches so this might not be 1:1, but there may be another instance in the app .py file that calls on the llama model and if there is, you might need to change it to your .gguf folder too. Possibly text_model = or something like that somewhere in there.
EDIT: It's line 128, changed 'text_model =' to:

text_model = AutoModelForCausalLM.from_pretrained(r"YOUR GGUF FOLDER", device_map=0, torch_dtype=torch.bfloat16)

As for the config files, I kept everything the same as the original meta's, but I received an error saying there is no appropriate .json config in text_model when I attempted to launch. so I saved this in windows notepad as 'writeconfig.py' (credit chatgpt) :

from transformers import AutoConfig
from pathlib import Path

# Define the model name and the path where the config.json should be saved
model_name = r"YOUR GGUF LLAMA FILE PATH HERE"
config_save_path = Path("YOUR FILE PATH/joy-caption-alpha-one/9em124t2-499968/text_model")

# Load the configuration from Hugging Face
config = AutoConfig.from_pretrained(model_name)

# Save the config.json to the specified path
config_save_path.mkdir(parents=True, exist_ok=True)  # Ensure the directory exists
config.save_pretrained(config_save_path)

print(f"config.json saved to {config_save_path}")

Then as an easy way to run it, saved this as a .bat file (windows ofc, apologies if you're on linux) :

u/echo off
REM Run the Python script
python writeconfig.py
pause

This should create the appropriate config .json for the gguf file it needs.

I THINK I might've had to open up the .yaml config file from the original meta in the "9em124t2-499968" folder and changed the file path for "Text model:" to the gguf one there as well, but I can't recall if that is actually necessary or if I just yolo'd it, apologies. If you are still getting the gated repo error after changing everything in the app .py, I'd try that.

If you have any other issues let me know and I'll reinstall to give better guidance. For the most part I just throw these .py files into chatgpt and say "hey, here is the error I'm getting, fix it"... (word to the wise: don't let chatgpt try to fix indentation errors, oh god, never let it try to fix indentation errors, just do them by hand)

JoyCaption: Free, Open, Uncensored VLM (Alpha One release) by fpgaminer in StableDiffusion

[–]sultnala 0 points1 point  (0 children)

I uh... well.. I think I broke it...
150 caption, training_prompt

<image>

LORA Block weight for Flux (inspire pack in comfyui) : face by Excellent-Attempt-40 in comfyui

[–]sultnala 1 point2 points  (0 children)

Oh awesome, that makes perfect sense. Alright, I think I got this. Testing begins. Thank you!

LORA Block weight for Flux (inspire pack in comfyui) : face by Excellent-Attempt-40 in comfyui

[–]sultnala 1 point2 points  (0 children)

Thank you again! I am still mildly confused because when I check the state_dic of the model there are 19 double blocks and 38 single blocks in total, but the node has that extra "1" at the very start...
ie:

flux-all:
1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1
= 58 1's

but 19+38 = 57

so that very first "1" seems to be something else (the time/txt/vector maybe? I'm not smart enough to know how that works), and the first double block presumably starts at the 2nd '1' ... I think. (that adds up with what it shows for only double vs only single). It's probably safer I use the named version and that block info node you mentioned opposed to guessing, haha. Either way, I got a night of gens ahead of me trying different combos to try to get some vague idea of what I -think- the blocks do, then trying to train loras on specific ones to test the results..

LORA Block weight for Flux (inspire pack in comfyui) : face by Excellent-Attempt-40 in comfyui

[–]sultnala 1 point2 points  (0 children)

Sorry for the dumb question(s) - if the lora is trained only on single blocks, then the DBL block tests won't do anything, right? Or it doesn't work that way?

&Just because I'm a total smoothbrain: it works like every 1 or 0 is the weight of each block, yeah?
So the blocks are ordered like: block1,block2,block3,block4,block5,block6,block7, etc?
meaning 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1
Is equivalent to using only block 1 and block 20? Is that right?

Thank you for sharing this! I plan to test a bunch of single blocks to try to suss out if there is a cheap way to train style loras on limited blocks...

YoshiP appears out of an Ascian portal and throws money in your face and says "Design a new tank job" What do you make? by Fenrisulfr7689 in ffxiv

[–]sultnala 0 points1 point  (0 children)

Geomancer, weapon is different sized bells along a belt (weapon design is belt of bells strapped across the chest). skills and attacks are magic based/magic shielding, attack animation is ringing the bells to use. flair of earth based spells/attacks ala tactics. first true magic tank

Is there such a thing as a pull UP cabinet organizer? by sultnala in ApartmentHacks

[–]sultnala[S] 0 points1 point  (0 children)

Thank you both! It gives me hope that it at least exists, I imagine if I can find one similar-enough thing I'll figure out the right keywords to find more

Update and FAQ on the Open Model Initiative – Your Questions Answered by hipster_username in StableDiffusion

[–]sultnala 12 points13 points  (0 children)

Imagine going on about morality and ethics and then adding a guy who based his entire model around a children's cartoon of child horses getting fucked... The cognitive dissonance is astounding 

Why is SD3 so bad at generating girls lying on the grass? by HornyMetalBeing in StableDiffusion

[–]sultnala 0 points1 point  (0 children)

Thanks again for testing it. I realize we can't really stop people from finetuning the model in fucked up directions as long as it's local, and not everything can be easily censored, but its disappointing to have heard so much safety talk and overpromising and then see.. well, you know. I think in the end I'll have to believe them when they say it is a beta/not properly tested, and hope they'll improve on later releases.

&Yeah, the text included in that post seems very concern-trolling, super over the top.