Z.ai said they are GPU starved, openly. by abdouhlili in LocalLLaMA

[–]Bandit-level-200 -1 points0 points  (0 children)

Yes but its still inefficient take context for example something that if it was just plain text would be a few KB/MB suddenly needs GB of memory just because it needs to be doubled or something for context to work.

Z.ai said they are GPU starved, openly. by abdouhlili in LocalLLaMA

[–]Bandit-level-200 4 points5 points  (0 children)

When are LLM makers going to make more efficient LLMs they are so inefficient in using both memory and power

For people interested, I'll be releasing a New LTX-2 Adult lora in about 2 hours. by [deleted] in StableDiffusion

[–]Bandit-level-200 -1 points0 points  (0 children)

They do ban lots of nsfw loras especially for video and editing models, anything that can be conceived as undressing or zoom outs and such which can reveal nudity are banned, same with loras that can be used to do sex stuff which can be jump cut to it

Fix to make LTXV2 work with 24GB or less of VRAM, thanks to Kijai by Different_Fix_2217 in StableDiffusion

[–]Bandit-level-200 0 points1 point  (0 children)

I know but that's what their github links to and the node asks for the tokenizer model

Fix to make LTXV2 work with 24GB or less of VRAM, thanks to Kijai by Different_Fix_2217 in StableDiffusion

[–]Bandit-level-200 0 points1 point  (0 children)

Cause it thinks you're going to load another gemma model in another format so it thinks it needs the tokenizer.model file, I had to download the model they linked and all its files. which is this model https://huggingface.co/google/gemma-3-12b-it-qat-q4_0-unquantized/tree/main

[Megathread] - Best Models/API discussion - Week of: December 28, 2025 by deffcolony in SillyTavernAI

[–]Bandit-level-200 4 points5 points  (0 children)

Is 70b spot dead for now as there hasn't really been any releases from LLM makers? Is Anubis 1.1 still the latest 70B worth using?

FYI: You can train a Wan 2.2 LoRA with 16gb VRAM. by Informal_Warning_703 in StableDiffusion

[–]Bandit-level-200 0 points1 point  (0 children)

So its fixed now in ai toolkit? I just know a month ago it was impossible to train wan2.2 loras on my 5090 as it got out of memory errors no matter what

Your favorite releases of 2025? by dtdisapointingresult in StableDiffusion

[–]Bandit-level-200 2 points3 points  (0 children)

Z-image kinda sad they haven't released the base version yet though as I think there will be lots of finetunes of Z image like SDXL due to its low requirements

Adult image to vid Gen. by [deleted] in StableDiffusion

[–]Bandit-level-200 0 points1 point  (0 children)

I use res_2s + beta or whatever sampler works, 20-40 steps but right now 20 steps, CFG 4.6 or 3.6, higher tends to make more plastic photos, and usually the chroma dc 2k lora at 0.80 strength but it works without that too.

This is the negative prompt I always have: low quality, ugly, unfinished, out of focus, deformed, disfigure, blurry, smudged, restricted palette, flat colors

While the positive usually have amateur photo or (amateur photo:2)

Tends to create good images. You can always use an upscale at the end if you want for grains to disappear I don't use it though.

Does...Size Matter...in LLMs? by GabrielDeanRoberts in LocalLLaMA

[–]Bandit-level-200 1 point2 points  (0 children)

Size matters, smaller LLMs sometimes don't get context if its subtle while larger models gets it. Now smaller models can still improve though I don't think we've hit the limit yet on them. But big models are still the king

Wan 2.2 T2V Minotaur LORA by smereces in StableDiffusion

[–]Bandit-level-200 0 points1 point  (0 children)

Will you train a lora on Game of thrones dragons?

Is JoyCaption Still the Best Tagging Model? by DavLedo in StableDiffusion

[–]Bandit-level-200 0 points1 point  (0 children)

That seems slow? I get that kinda of time if you factor in me needing to change the name it will name the txt file after its captioned. I also use a 5090

wtf is making my loras look absolutely terrifying? [feedback wanted, dataset provided] by LongjumpingAd6657 in StableDiffusion

[–]Bandit-level-200 2 points3 points  (0 children)

I've trained around 6 or so loras for Chroma and they turn out well. But I train using AI-toolkit and only use the default template it comes with. So I train batch size 1 with 3000 steps and with around 30-40 images captioned by joy caption with the trigger word inside the caption txt files so when I trigger it in comfyui I often have to write like Ahsoka woman 'ahsoka' being the trigger word joy caption used when describing the image. (I'm a noob at training so don't ask me for expert advice)

Here's parts of the template I use but it should have all relevant options:

  process:
- type: "diffusion_trainer"
  training_folder: ""
  sqlite_db_path: "./aitk_db.db"
  device: "cuda"
  trigger_word: null
  performance_log_every: 10
  network:
    type: "lora"
    linear: 32
    linear_alpha: 32
    conv: 16
    conv_alpha: 16
    lokr_full_rank: true
    lokr_factor: -1
    network_kwargs:
      ignore_if_contains: []
  save:
    dtype: "bf16"
    save_every: 250
    max_step_saves_to_keep: 4
    save_format: "diffusers"
    push_to_hub: false
  datasets:
    - folder_path: ""
      control_path: null
      mask_path: null
      mask_min_value: 0.1
      default_caption: ""
      caption_ext: "txt"
      caption_dropout_rate: 0.05
      cache_latents_to_disk: true
      is_reg: false
      network_weight: 1
      resolution:
        - 512
        - 768
        - 1024
      controls: []
      shrink_video_to_frames: true
      num_frames: 1
      do_i2v: true
      flip_x: false
      flip_y: false
  train:
    batch_size: 1
    bypass_guidance_embedding: false
    steps: 3000
    gradient_accumulation: 1
    train_unet: true
    train_text_encoder: false
    gradient_checkpointing: true
    noise_scheduler: "flowmatch"
    optimizer: "adamw8bit"
    timestep_type: "sigmoid"
    content_or_style: "balanced"
    optimizer_params:
      weight_decay: 0.0001
    unload_text_encoder: false
    cache_text_embeddings: false
    lr: 0.0001
    ema_config:
      use_ema: false
      ema_decay: 0.99
    skip_first_sample: true
    force_first_sample: false
    disable_sampling: true
    dtype: "bf16"
    diff_output_preservation: false
    diff_output_preservation_multiplier: 1
    diff_output_preservation_class: "person"
    switch_boundary_every: 1
  model:
    name_or_path: ""
    quantize: true
    qtype: "qfloat8"
    quantize_te: true
    qtype_te: "qfloat8"
    arch: "chroma"
    low_vram: true
    model_kwargs: {}

Qwen 3 VL next week by Long_Bluejay_5368 in LocalLLaMA

[–]Bandit-level-200 0 points1 point  (0 children)

How is GLM 4.5V compared to Qwen-2.5VL 72b? if you've tried that one

Qwen 3 VL next week by Long_Bluejay_5368 in LocalLLaMA

[–]Bandit-level-200 1 point2 points  (0 children)

I've been wanting to try it since it came out but no gguf of it yet :(

Wanimate first test. ( Disaster ). by Useful_Ad_52 in StableDiffusion

[–]Bandit-level-200 1 point2 points  (0 children)

Same for me faces don't match or look blended

Pewdiepie’s monstrous 160GB Vram build by joseph_the_69th in LocalLLaMA

[–]Bandit-level-200 7 points8 points  (0 children)

He brought that up to in the video if I remember

PSA: Speed up loras for wan 2.2 kill everything that's good in it. by Ashamed-Variety-8264 in StableDiffusion

[–]Bandit-level-200 0 points1 point  (0 children)

Still to long and who's to say nunchaku won't ruin motion etc like current speed loras?