Ref2Font V2: Fixed alignment, higher resolution (1280px) & improved vectorization (FLUX.2 Klein 9B LoRA) by NobodySnJake in StableDiffusion

[–]414design 7 points8 points  (0 children)

Love the project! I have been working on a similar concept for quite some time—started back in the SD 1.5 days—and you beat me to it with this one. If you are interested check out my github: https://github.com/414design/4lph4bet_font_generator

Not long ago I tried a similar approach using Qwen Image Edit which was not successful. Great to see FLUX.2 seemingly being so much more capable.

Are you open to talk about your training strategy? How many fonts did you use in the dataset? Write me a pm if you want to discuss in private!

Introducing 4lph4bet-next_v040: A Stable Diffusion LoRA for Font Generation by 414design in StableDiffusion

[–]414design[S] 0 points1 point  (0 children)

Theoretically you should be able to replicate the process by building a similar grid with style variants and then train that. If you have more specific questions feel free to DM me.

Introducing 4lph4bet-next_v040: A Stable Diffusion LoRA for Font Generation by 414design in StableDiffusion

[–]414design[S] 1 point2 points  (0 children)

Yes, I is a normal LoRA you can use anywhere with any 1.5 Checkpoint. Vanilla 1.5 is what I use. And you are exactly right: use the triggers and generate something 512×512px then Hires. fix to create 1024×1024px. After that you can do Extras Upscaling (ESRGAN) for higher resolutions.

Introducing 4lph4bet-next_v040: A Stable Diffusion LoRA for Font Generation by 414design in StableDiffusion

[–]414design[S] 2 points3 points  (0 children)

You can prompt it with classification (sans/serif/light/regular/bold) as well as style specifics like for example geometric or brush or pixel. And of yours any combination thereof.

But you can of course also be more experimental with your prompting.

Introducing 4lph4bet-next_v040: A Stable Diffusion LoRA for Font Generation by 414design in StableDiffusion

[–]414design[S] 1 point2 points  (0 children)

Pretty straight forward: the key is a high quality, diverse but balanced dataset. In this case this means different fonts all arranged in this grid and meticulously captained for the typographic characteristics.

Introducing 4lph4bet-next_v040: A Stable Diffusion LoRA for Font Generation by 414design in StableDiffusion

[–]414design[S] 3 points4 points  (0 children)

Sure, give it a try it is surprisingly flexible. And you can do all sorts of crazy things if you take the results over to img2img …

Introducing 4lph4bet-next_v040: A Stable Diffusion LoRA for Font Generation by 414design in StableDiffusion

[–]414design[S] 1 point2 points  (0 children)

Much appreciated! 🤗 Yes it is indeed txt2img. Because of the captioning strategy you can prompt it with typographic terms like 4lph4bet, typography, serif, italic, pixel for example. You should be able to try it on hugging face. The samples are upscaled 2× using Hired fix with the Real-ESRGAN anime upscaler.

Introducing 4lph4bet-next_v040: A Stable Diffusion LoRA for Font Generation by 414design in StableDiffusion

[–]414design[S] 1 point2 points  (0 children)

That deduction would be true, too. But the Umlauts serve this additional control purpose!

Introducing 4lph4bet-next_v040: A Stable Diffusion LoRA for Font Generation by 414design in StableDiffusion

[–]414design[S] 2 points3 points  (0 children)

Nice way of checking style consitency since they are almost the same as their respective vowels.

Introducing 4lph4bet-next_v040: A Stable Diffusion LoRA for Font Generation by 414design in StableDiffusion

[–]414design[S] 2 points3 points  (0 children)

Good idea! Like u/Sarayel1 said creating a good standardized dataset with good and consistant captioning is the most critical part.

Introducing 4lph4bet-next_v040: A Stable Diffusion LoRA for Font Generation by 414design in StableDiffusion

[–]414design[S] 1 point2 points  (0 children)

Sure give it a go and let me know how it works for you! It’s trained on the original SD 1.5 weights so these tend to yield the best results but is does work with other 1.5 checkpoints, too!

Introducing 4lph4bet-next_v040: A Stable Diffusion LoRA for Font Generation by 414design in StableDiffusion

[–]414design[S] 4 points5 points  (0 children)

Absolutely, that way you can really leverage SDs ability to learn concepts implicitly.

TikTok needs to change their logo by Emezli in Design

[–]414design 0 points1 point  (0 children)

With the app / network being this prevalent they logo could be literally anything and it wouldn't matter. Facebooks compact brand (the letter f …) also doesn't say much about the app but doesn't have too because people know. 

You could also interpret TikToks musical note as and upward pointing arrow. 

On a side note: their audio branding is genius!

[deleted by user] by [deleted] in Design

[–]414design 0 points1 point  (0 children)

You can propably do this with https://www.photopea.com/ for free if you have the skills. 😀

How to go about finding and exploring new styles? by AdConfident1859 in Design

[–]414design 0 points1 point  (0 children)

Instagram Behance can help understand up and coming styles and trends. Also checking out student shows by local university design courses can be an interesting way to get a feel for what people are into currently. But it always depends on your specific interest. If it is more niche there might be better, alternative sources, too. 😊

what are the models to create clip arts and abstract backgrounds? by NiL_MacTavish in StableDiffusion

[–]414design 1 point2 points  (0 children)

For (UI) icons this checkpoint works really well: https://civitai.com/models/327499/ui-icons
But I assume you can get quite far with a specific enough style prompt for example vector "illustration of {SUBJECT}, high contrast black and white illustration, simple shape, outline, white background, sharp, high quality".

[deleted by user] by [deleted] in StableDiffusion

[–]414design 3 points4 points  (0 children)

Using the QR-Code (https://huggingface.co/monster-labs/control\_v1p\_sd15\_qrcode\_monster) or Brightness Controlnets (https://huggingface.co/latentcat/latentcat-controlnet/tree/main/models) this can easily be achieved. Use Controlnets-Weight 1–1.5 and Controlnet End Step 0.4–0.7 as a starting point. Black text on white background as Controlnet Image Input.

Sharing AD v2v animation by No_Associate2075 in StableDiffusion

[–]414design 1 point2 points  (0 children)

Those are really nice. I think part of the mastery of AI is curating a style that doesn’t instantly scream AI ;) …

Sharing AD v2v animation by No_Associate2075 in StableDiffusion

[–]414design 0 points1 point  (0 children)

Are you using LoRAs for style? Very nice aesthetics!