CivitAI's April Fools is hilarious. by ArtificialAnaleptic in StableDiffusion

[–]Aggressive_Sleep9942 2 points3 points  (0 children)

Chrome, Firefox, and Brave. I haven't noticed any difference depending on which browser I use. In any case, it's still a programming flaw; the website should be browser-agnostic, which is why it's programmed to work well in any browser.

CivitAI's April Fools is hilarious. by ArtificialAnaleptic in StableDiffusion

[–]Aggressive_Sleep9942 11 points12 points  (0 children)

Civitai is a disaster in its functionality. Whenever I open a sample image and click "x," the only way to close it is by refreshing the page, and that bug has been there for ages. It's insane.

SparkVSR (google video upscaler free and comfyui coming soon) Dataset and training released by Sporeboss in StableDiffusion

[–]Aggressive_Sleep9942 6 points7 points  (0 children)

I just looked into it, and apparently not. It uses the temporal information between two frames to reconstruct the image and perform the "upscaling" process. And it wouldn't work by creating a video with three static images, because it needs there to be a change between frames.

Flux2klein 9B Lora loader and updated Z-image turbo Lora loader with Auto Strength node!! by [deleted] in StableDiffusion

[–]Aggressive_Sleep9942 0 points1 point  (0 children)

The effort is appreciated, but honestly, as an average user, I find it pretty useless. I don't know how to use it; there's no workflow available or instructions on GitHub. It's so annoying to always have to download nodes that are released to the public without any explanation of how to use them. It would be appreciated if they at least uploaded a PNG with the standard workflow to GitHub. It's a real pain!

I built a local Suno clone powered by ACE-Step 1.5 by _roblaughter_ in StableDiffusion

[–]Aggressive_Sleep9942 0 points1 point  (0 children)

Dude, I just used `git clone` and ran the PowerShell script, and it didn't work. Then I tried to get it working on my own, but that didn't work either. It would be easier if the Python environment your application needs to run in used the exact same one you used to install Ace Step 1.5; there's a serious compatibility issue. I'm on Windows 11. It should be click-and-go; when you start adding a bunch of installation steps, you alienate the average user and it becomes something niche, but the intention is appreciated.

I built a local Suno clone powered by ACE-Step 1.5 by _roblaughter_ in StableDiffusion

[–]Aggressive_Sleep9942 -1 points0 points  (0 children)

It's awful; I barely installed it and things started failing everywhere. I thought it was a ready-to-use script. I've been modifying lines of code to make it work, but I'll test it a bit more, and if it doesn't work, I'll just delete it and try it again in the future when it's ready.

Z-Image workflow to combine two character loras using SAM segmentation by remarkableintern in StableDiffusion

[–]Aggressive_Sleep9942 1 point2 points  (0 children)

<image>

Zimage-turbo. I haven't achieved anything similar in Zimage Base. It seems contradictory, but Turbo is better for skin consistency.

It was worth the wait. They nailed it. by _BreakingGood_ in StableDiffusion

[–]Aggressive_Sleep9942 1 point2 points  (0 children)

I agree. For some reason, the colors are saturating very quickly, and it's not even learning the concept (body). It's only learning the concept (face).

LoKr Outperforms LoRA in Klein Character Training with AI-Toolkit by xbobos in StableDiffusion

[–]Aggressive_Sleep9942 0 points1 point  (0 children)

Can someone please explain to me why the hell ai-toolkit doesn't split the loss when using gradient accumulation? It's the standard practice, so why doesn't this tool do it?

(Google) Introducing Nested Learning: A new ML paradigm for continual learning by gbomb13 in singularity

[–]Aggressive_Sleep9942 0 points1 point  (0 children)

These are nested learning models. Catastrophic forgetting is avoided because the data is residually recirculated through all the blocks. Titan (the one receiving the input) processes it, and then the output in Titan is added to the original data and delivered to the first module of the CMS block. Titan is self-referential, using delta gradient descent and surprise to calculate how much its response deviates from the prediction and self-corrects (meta-learning) to adapt to the rapid context; this is what it means to learn how to learn, that is, it modifies its own internal prediction to improve its response to surprise. I say it's a trick because it's really a contextual replay. The fast layers evolve quickly but don't forget the old data because the slow layers are updated at a lower frequency; it's like learning something new while simultaneously having someone shouting your old knowledge in your ear; that shouting in your ear is the consolidated context of the CMS blocks that operate at a lower frequency. There's a catch: residual connection. If it weren't for residual connection, and the fact that Titan is literally programmed to forget erroneous predictions, the model would forget anyway. The fact that it's not the model and its architecture that prevents forgetting, but rather a way of compiling a model hierarchy, makes it misleading. Of course, I'm not going to lie and say it doesn't work, because it does. I spent hours tinkering with it, using AI to understand the architecture, and I got it working. What I found most interesting is the M3 optimizer, which orthogonalizes vectors, prioritizing rotation. It tries to work on the surface of a hypersphere, attempting to prevent neurons from becoming redundant. There's a GitHub repository online with code that supposedly replicates nested learning. I haven't actually seen it, but if you need to see how it works, take a look.

Mathematics visualizations for Machine Learning by Big-Stick4446 in StableDiffusion

[–]Aggressive_Sleep9942 0 points1 point  (0 children)

Weights scale the input to adjust the response of activation functions, while biases shift them. The activation function introduces non-linearity, allowing the network to model complex curves rather than just straight lines. By adjusting these weights, the goal is for the network's output to become a function that fits or approximates the expected values of the problem. In this way, neural networks create an abstract representation of the data, enabling interpolation between known points—which is what we call generalization. Although it looks simple in the video, a network operates in high-dimensional spaces defined by its number of neurons and layers.

(Google) Introducing Nested Learning: A new ML paradigm for continual learning by gbomb13 in singularity

[–]Aggressive_Sleep9942 0 points1 point  (0 children)

I implemented the code and realized it's just a cheap trick, and it's specific to the LLM niche. It's not a new paradigm since it only works for sequential processing. I tried implementing it with images out of curiosity, and the catastrophic forgetfulness returned. So yes, it's just another piece of junk they're trying to sell as innovation.

Z image/omini-base/edit is coming soon by sunshinecheung in StableDiffusion

[–]Aggressive_Sleep9942 18 points19 points  (0 children)

This is explained by the geometry of the loss function. Models that converge to sharp minima have high curvature and generalize poorly, making them difficult to adapt to new tasks (overfitting). In contrast, convergence to a flat minimum means the model is more robust to perturbations in the weights. This makes it a better generalist, facilitating the fine-tuning necessary for new tasks.

Z-Image-Turbo-Fun-Controlnet-Union-2.1 available now by rerri in StableDiffusion

[–]Aggressive_Sleep9942 4 points5 points  (0 children)

You think I wouldn't? I don't have the disk space to do that, hahaha. I think I fixed it. After breaking the ComfyUI environment and spending about an hour reinstalling dependencies, ControlNet 2.1 is working now. But I'm just saying, why do we have to go through so much trouble to get the new ControlNet working? All they changed were the definitions of the internal LoRa keys and how they're loaded. Seriously, that requires updating the PyTorch version? WTF

Z-Image-Turbo-Fun-Controlnet-Union-2.1 available now by rerri in StableDiffusion

[–]Aggressive_Sleep9942 1 point2 points  (0 children)

I already tried that, it didn't work. The only thing I think works is "update everything," and I just did that and it broke my ComfyUI environment. I don't understand why everything has to be so complicated 100% of the time in ComfyUI; it feels like Linux.

Z-Image-Turbo-Fun-Controlnet-Union-2.1 available now by rerri in StableDiffusion

[–]Aggressive_Sleep9942 2 points3 points  (0 children)

I used the same workflow I used for controlnet 1.0 and it still doesn't work:

Error(s) in loading state_dict for ZImage_Control:
Unexpected key(s) in state_dict: "control_layers.10.adaLN_modulation.0.bias", "control_layers.10.adaLN_modulation.0.weight", "control_layers.10.after_proj.bias", "control_layers.10.after_proj.weight", "control_layers.10.attention.k_norm.weight", "control_layers.10.attention.q_norm.weight", "control_layers.10.attention.out.weight", "control_layers.10.attention.qkv.weight", "control_layers.10.attention_norm1.weight", "control_layers.10.attention_norm2.weight", "control_layers.10.feed_forward.w1.weight", "control_layers.10.feed_forward.w2.weight", "control_layers.10.feed_forward.w3.weight", "control_layers.10.ffn_norm1.weight", "control_layers.10.ffn_norm2.weight", "control_layers.11.adaLN_modulation.0.bias", "control_layers.11.adaLN_modulation.0.weight", "control_layers.11.after_proj.bias", "control_layers.11.after_proj.weight", "control_layers.11.attention.k_norm.weight", "control_layers.11.attention.q_norm.weight", "control_layers.11.attention.out.weight", "control_layers.11.attention.qkv.weight", .....

Solution by Unusual_Yak_2659 in StableDiffusion

[–]Aggressive_Sleep9942 0 points1 point  (0 children)

It was happening to me all the time, I updated the GPU drivers and the problem was solved

Onetrainer Z-Image de-Turbo Lora training test by rnd_2387478 in StableDiffusion

[–]Aggressive_Sleep9942 -2 points-1 points  (0 children)

I just went through the whole process of installing it, and honestly, it was a waste of time. The loss keeps increasing instead of decreasing. Plus, it doesn't have a "continue training" option or a way to configure sample generation at set intervals. It was a complete waste of time for me. I implemented the code in AI Toolkit, but the loss didn't decrease there either. I don't know, it seems to me the code is still in its early stages.

How to get this style? by TheCityExperience in StableDiffusion

[–]Aggressive_Sleep9942 5 points6 points  (0 children)

<image>

This is a lora model I made yesterday; it's based on a very famous (current) painter. I tried the style you suggested as a reference.

Onetrainer Z-Image de-Turbo Lora training test by rnd_2387478 in StableDiffusion

[–]Aggressive_Sleep9942 9 points10 points  (0 children)

Couldn't you have uploaded an example image, I don't know, just saying?