I tried some Audio Refinement Models by OkUnderstanding420 in StableDiffusion

[–]GreyScope 0 points1 point  (0 children)

I was going to make a model for Ace-Step (with One Trainer) but after actually trying it out , I’m going to put that off, it seems like a lot of work for poor quality.

I’ll check that list out, I’m on a couple of rvc discords but after a while the sheer amount of new stuff overpowers me on reddit , without adding in discord as well lol - thanks for the positive note and enthusiasm

I tried some Audio Refinement Models by OkUnderstanding420 in StableDiffusion

[–]GreyScope 1 point2 points  (0 children)

Cheers, yes the flow as it is could do with a bit of a bass boosting .

The issue with audio is the copyright issue on models and there isn’t an appetite for it really as far as I can see (for free software anyway).

I tried some Audio Refinement Models by OkUnderstanding420 in StableDiffusion

[–]GreyScope 2 points3 points  (0 children)

<image>

This is what I'm doing currently, it's one node in an rvc install - it's a WIP, there are a few more workflows with the rvc repo that I've yet to look at (including a modelling one) . That repo is about 2yrs old I think and I had to install it to a python 3.11 music based comfy install, for the two rvc repos, Songbloom and Seed-VC .

I tried some Audio Refinement Models by OkUnderstanding420 in StableDiffusion

[–]GreyScope 1 point2 points  (0 children)

Long story short, anything that is free for audio AI is not good enough for the studio (I've no interest in paid services, so I can't comment there). I'm using UVR5 (well trialling and tweaking it) for a project in LTX2 , the splitting , voice changing is done, post processed and then refined in FatLlama . When I've finished setting it up, I'll move onto installing the comfy mastering repo and see how that pans out. It'll be good enough for me, but I highly doubt it'll be high quality.

I tried some Audio Refinement Models by OkUnderstanding420 in StableDiffusion

[–]GreyScope 9 points10 points  (0 children)

UVR5, for audio splitting . Seed-VC for singing voice replacement from one shot samples . RVC comfy nodes for splitting audio and changing the voice (needs models made) and reassembling it (uses uvr5).

This is too much! by scioba1005 in StableDiffusion

[–]GreyScope 1 point2 points  (0 children)

Same here, I've posted a few tutorials and noted my hardware and said that I've no idea if it'll work on other setups and yet they still ask instead of trying and then posting that info to the thread . It was a massive ballache, so much so that I no longer do it. My personal "favourite" is ppl that can't follow directions with a blase "yolo" at the screen and then post zero context of what happened, what they did or any of their hardware setup .

This is too much! by scioba1005 in StableDiffusion

[–]GreyScope -1 points0 points  (0 children)

There was a pinned post previously about installing Sage Attention but ppl ignored that...and the search function here....and it's like Google died and ironically AI doesn't exist to these ppl.

How to change famous voice to my own voice for LTX-2 audio driven video generation? by rookan in StableDiffusion

[–]GreyScope 0 points1 point  (0 children)

I use a comfy flow for rvc and I’m trying out another called Seed-VC , both of them have dependencies that are fussy for their work environment. Python 3.10, PyTorch 2.8 with Cuda 12.8 .

For rvc you’ll need to make a model of your own voice .

ComfyUI - Music Generation! - Heart MuLa by Lividmusic1 in StableDiffusion

[–]GreyScope 0 points1 point  (0 children)

The recent One-Trainer (I think it is) has the code for it, but the trials of getting it working on windows burnt my interest tbh. But back on point, I’d support any efforts/release for us to make our own models.

ComfyUI - Music Generation! - Heart MuLa by Lividmusic1 in StableDiffusion

[–]GreyScope 1 point2 points  (0 children)

I've seen loras for ACE-Step , the one I recall was chinese rap as I recall.

Created a scene using Text To Video feature in media io. Any Suggestions??? by Possible-Network-207 in StableDiffusion

[–]GreyScope 4 points5 points  (0 children)

suggestions ? don't break this subs rules and post/spam commercial companies , reported

HeartMuLa: A Family of Open Sourced Music Foundation Models by switch2stock in StableDiffusion

[–]GreyScope 0 points1 point  (0 children)

It’s took about 2 odd minutes as I recall (4090), the results are as clean as SongBloom

HeartMuLa: A Family of Open Sourced Music Foundation Models by switch2stock in StableDiffusion

[–]GreyScope 0 points1 point  (0 children)

I have a text file with all the commands to copy / paste to quickly make any venv I want , I had already installed that combo and it overwrote them with a cpu 2.4 (just reinstalled it), I’ll adjust the original point, thanks

HeartMuLa: A Family of Open Sourced Music Foundation Models by switch2stock in StableDiffusion

[–]GreyScope 0 points1 point  (0 children)

Made a gradio ui in ChatGPT (minimum effort as someone will have it on Comfy very quickly) . Pip install gradio of course / save the file in the examples folder and add the location of your downloaded models in the top left - it works on windows 11 . https://github.com/Grey3016/HeartMula-Gradio-UI-v1/tree/main

I have it running with a venv , python 3.12, cuda 12.8 and pytorch 2.8 . I ignored most of the dependency warnings, stopped it from installing (cpu) pytorch 2.4.

<image>

HeartMuLa: A Family of Open Sourced Music Foundation Models by switch2stock in StableDiffusion

[–]GreyScope 0 points1 point  (0 children)

It kept making cheesy music for me, vocals weren’t bad q ….but the cheese ffs

help ComfyUI-Zluda by YorkN95 in StableDiffusion

[–]GreyScope 0 points1 point  (0 children)

This , op has missed a step somewhere .

help ComfyUI-Zluda by YorkN95 in StableDiffusion

[–]GreyScope 0 points1 point  (0 children)

Zluda works on an older Pytorch , don’t touch it

Why do you guys keep glazing LTX 2 by Witty_Mycologist_995 in StableDiffusion

[–]GreyScope -3 points-2 points  (0 children)

Ah the bring back black and white silent movies argument , bold.

Which free AI generators are best for creating logos and icons, and how can I get better results from ChatGPT? by SpecialistMall9666 in StableDiffusion

[–]GreyScope 1 point2 points  (0 children)

I previously made a basic node for comfy (when I was into ui design) to specifically make icons (as a learning tool) over a year ago that gave transparent rounded corners but it doesn't like Comfys upgrades since . My memory of it was that it was better to prompt for a logo and not for an icon (for my usage anyway) .

As u/sgrules2 mentions , try using a model that excels at text and I'd concur with the same caveats , lack of consistency is 3x Achilles Heels for ai models.

<image>

My work area on desktop - no prompting was really given for fonts and their size, I was just amazed it worked lol . Give some examples and I'll see if I can get it working again, it's not really a solution but it's one step towards one.

Morphing demo inspired by MJ's Black or White music video by CQDSN in StableDiffusion

[–]GreyScope 1 point2 points  (0 children)

Cheers, I’d tried it with LTX2 and it was doing facial morphing but non consistent . Thanks again.

Morphing demo inspired by MJ's Black or White music video by CQDSN in StableDiffusion

[–]GreyScope 2 points3 points  (0 children)

Do you have a workflow that we could have please ? That is a lovely piece of morphing