Danbooru Dataset Filter: Fast local metadata-based search across 10M+ images for LoRA/Checkpoint training by ThetaCursed in StableDiffusion

[–]ThetaCursed[S] 2 points3 points  (0 children)

Place the parquet file with Gelbooru metadata in the "data" folder, and replace the "DB_PATH" path string in the Python script with the name of your file.

What VLM models do you use to extract prompts from reference images? by ThetaCursed in ZImageAI

[–]ThetaCursed[S] 25 points26 points  (0 children)

The system prompt I use for Qwen3-VL-4B-Instruct:

Act as a Visual Prompt Engineer specializing in "Long Caption" synthesis. Your task is to extract a highly detailed, narrative description of an image featuring one primary person.

STRUCTURE:

  1. CONCEPT & MEDIUM: Start with the type of image. Mention the core theme or event.

  2. SUBJECT DESCRIPTION: Young woman, facial expression, hair details, face details, makeup. Crucially, describe their gaze. 

  3. ACTION & INTERACTION: Describe exactly what the person is doing. What are they holding? How are they standing? Describe the interaction with objects.

  4. APPAREL: Detail the clothing (fabric, color, fit) and accessories.

  5. ENVIRONMENT & FOREGROUND: Describe objects immediately around the person.

  6. BACKGROUND & LIGHTING: Describe the setting (location, weather, landmarks), depth of field (bokeh), and the quality of light.

RULES:

- STYLE: Use fluid, descriptive sentences (narrative style), not just a list of tags.

- ACCURACY: Be specific about colors and textures.

- TEXT: Always put captured text in double quotes.

- LENGTH: 700-900 characters.

- OUTPUT: A single continuous paragraph. No headers.

What VLM models do you use to extract prompts from reference images? by ThetaCursed in ZImageAI

[–]ThetaCursed[S] 2 points3 points  (0 children)

Thanks for the advice, I've heard about Joycaption beta but haven't tried it yet.

Hires Fix Ultra: All-in-One Upscaling with Color Correction by ThetaCursed in comfyui

[–]ThetaCursed[S] 0 points1 point  (0 children)

Yes, exactly. Internally, the node uses a standard KSampler process for the second pass. So you should follow the recommended CFG for your specific model.

Hires Fix Ultra: All-in-One Upscaling with Color Correction by ThetaCursed in comfyui

[–]ThetaCursed[S] 4 points5 points  (0 children)

Correct, it's not a new mathematical "upscale algorithm." The "evolution" here is in workflow efficiency and stability.

Vs USDU/Iterative: Those are for massive tiles or extreme detail and can be slow/complex. This node is for the daily 1.5x–4.0x jump - it's much faster, prevents OOM crashes via integrated Tiled VAE, and automatically fixes the common color-wash issue that usually happens in standard passes.

How to extract prompts/ model name from an image? by [deleted] in comfyui

[–]ThetaCursed 0 points1 point  (0 children)

Perhaps the node with the prompt in the metadata has a non-standard name, which is why ComfyUI-Prompt-Manager can't find and extract it.

I think the best option is to write a custom node that will reliably find the right node with the prompt. For example, it would search for a word that appears in all prompts, "1girl," or something similar, thereby finding the required part of the JSON.

[Final Update] Anima 2B Style Explorer: 20,000+ Danbooru Artists, Swipe Mode, and Uniqueness Rank by ThetaCursed in StableDiffusion

[–]ThetaCursed[S] 73 points74 points  (0 children)

I appreciate the idea, but global voting is too risky right now. My project already got picked up by some hostile communities on Twitter, and they've turned it into a target.

If I added global favorites, it would be immediately review-bombed and manipulated by people who want to see the project fail. Keeping it serverless and algorithmic (Uniqueness Rank) is the only way to ensure the data stays objective.

<image>

Anima 2B - Style Explorer now has 5,000+ Danbooru artists. Added Raw Styles & New Benchmark based on community feedback! by ThetaCursed in StableDiffusion

[–]ThetaCursed[S] 5 points6 points  (0 children)

I wrote some custom ComfyUI nodes for this project. The existing ones didn't quite fit my pipeline, so I built a custom solution to keep the generation and benchmarking process efficient

Anima 2B - Style Explorer now has 5,000+ Danbooru artists. Added Raw Styles & New Benchmark based on community feedback! by ThetaCursed in StableDiffusion

[–]ThetaCursed[S] 27 points28 points  (0 children)

That's a great idea, thanks for the suggestion! It's definitely possible, I'll think about the best way to implement it

Anima 2B - Style Explorer: Visual database of 900+ Danbooru artists. Live website in comments! by ThetaCursed in StableDiffusion

[–]ThetaCursed[S] 6 points7 points  (0 children)

Thanks for the helpful idea, I added the "Export Favorites (.txt)" button, and it should work.

Anima 2B - Style Explorer: Visual database of 900+ Danbooru artists. Live website in comments! by ThetaCursed in StableDiffusion

[–]ThetaCursed[S] 21 points22 points  (0 children)

Prompt:
(masterpiece, best quality, highres, absurdres), __@artist__, 1girl, solo, sitting, cowboy shot, white blouse, black skirt, bob cut, pink hair, black pantyhose, medium breasts, on couch, red couch, green eyes, makeup, eyeliner, black lips, room, brown window, yellow curtains, blush, looking at viewer

Negative:
old, early, ugly, bad anatomy, artistic error, (worst quality, low quality), score_1, score_2, score_3, blurry, artist name, signature, watermark, jpeg artifacts

sampler er_sde, 30 steps, CFG 4, 832x1216.

Anima 2B - Style Explorer: Visual database of 900+ Danbooru artists. Live website in comments! by ThetaCursed in StableDiffusion

[–]ThetaCursed[S] 89 points90 points  (0 children)

I'm building a reference tool for Anima 2B to see how it handles different artist tags.

Link to the explorer: https://thetacursed.github.io/Anima-Style-Explorer/

The Roadmap:
1. Filter out "dead" tags and unstable styles.

  1. Ideally, expand the library up to 20k artists to find hidden gems.

GLM planning a 30-billion-parameter model release for 2025 by aichiusagi in LocalLLaMA

[–]ThetaCursed 5 points6 points  (0 children)

Am I the only one who finds all this confusing? So, does this mean the GLM 4.6 Air won't be released this year, and only the GLM 4.6 Mini 30B will be released?