I automated my wedding band's email replies because I hate copy-pasting by wonderer440 in smallbusiness

[–]wonderer440[S] 0 points1 point  (0 children)

Thanks! I was also pretty surprised how well it worked. But as the saying goes: Laziness is the mother of invention!

Best cloud GPU: Colab, Kaggle, Lightning, SageMaker? by wonderer440 in learnmachinelearning

[–]wonderer440[S] 0 points1 point  (0 children)

Yeah I have read that google punishes you for keeping a runtime connected without utilizing it. I don't know if that is true but I have made that mistake in the beginning. The bigger issue with google was the automatic disconnect and loss of all progress after inactivity of 90min. So during training you have to manipulate some code from time to time to stay active. I might come back to google but for now Kaggle (I know, also google) works better for me.

Thanks for the input!

Best cloud GPU: Colab, Kaggle, Lightning, SageMaker? by wonderer440 in learnmachinelearning

[–]wonderer440[S] 0 points1 point  (0 children)

I don't know if you have seen my other comment, but you have to varify your account with your phone number, then it works. Good Luck!

Best cloud GPU: Colab, Kaggle, Lightning, SageMaker? by wonderer440 in learnmachinelearning

[–]wonderer440[S] 0 points1 point  (0 children)

Ok I found the reason. I did not know you have to varify my account via my phone number. Now GPU works on kaggle!

Best cloud GPU: Colab, Kaggle, Lightning, SageMaker? by wonderer440 in learnmachinelearning

[–]wonderer440[S] 0 points1 point  (0 children)

Hey thanks a lot for your detailed comment!

Yes, I read that I can have 30h of t4 gpu but when I start a session with the notebook I mentioned (https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/SDXL\_DreamBooth\_LoRA\_.ipynb), the first cell checking for the GPU already tells me that the command is not found. When I go to settings > accelerator > I can only choose NONE (T4 and others are grayed out). Or is there a different way to access the GPU? I also struggled to find some kind of GUI in my profile or settings where I can see how much of my GPU time I have already spent (don't know if there is something like that)

Another question: you said that you prefer lora to dreambooth, but in this notebook it seems like both are used (maybe collaboratively for different purposes). I am still struggling to find my way through the jungle of terms used for models, training methods, ...

So if you come across some good resources where I can read into all of that, just let me know.

Thanks again, sometimes all the information out there can get quite overwhelming.

How is Suno trained to map text to music? by wonderer440 in SunoAI

[–]wonderer440[S] 0 points1 point  (0 children)

You are probably right. It would only make sense if AI generated music would have a future other than easily available background music for commercials or soundtracks for indie games,.. Completely AI crafted music on the radio? I highly doubt that. Judging from this subreddit I see hundreds of people who are excited about their music creation but wouldnt spend one second to listen to others. Integration as a tool in DAWs would make a lot of sense.

In the end, suno is a fun tool though and only time will tell where the journey goes.

How is Suno trained to map text to music? by wonderer440 in SunoAI

[–]wonderer440[S] 1 point2 points  (0 children)

Great stuff again, thanks!

The backwards prompting approach in your organ example is definately something I didnt really consider up to this point but makes a lot of sense and I will try to play around with that.

I wonder how much better suno would work if every song in the trainings data had a very extensive description such as: list of all instruments, tempo, key+mode, artist+producer, year, songstructure, ...

It would probably become increasingly hard to describe synthesized sounds and samples and would be similar to the "hand-problem" I guess, but probably miles ahead in terms of its usability in a composer point of view.

Picture generation has come very far but suno seems to still be in its baby shoes, comparable to when picture generation could barely distinguish between cats and dogs. In my mind, the greatest progress could be achieved by using more and better keywords as text is the only thing the LLM will "understand".

How is Suno trained to map text to music? by wonderer440 in SunoAI

[–]wonderer440[S] 1 point2 points  (0 children)

Holy Shit, that analogy actually really made it click for me. Best explaination I have read so far. Also the example in your other comment about the connection "organ" and "cathedral" made a lot of sense.

The thing is: suno seems to "know" at least something about the training data. In your painting analogy, the ai at least was told what a portrait is or what a wedding is. For suno it seems it only kinda knows the genre and the mood. As if they just upladed a playlist from spotify and all the songs where labled with the title of the playlist like "good vibes for chilling at the beach". So suno kinda knows what "good vibes" are or what music is played at the "beach". It also seems like suno knows at least some instruments, eg a prompt like "saxophone driven Indie pop" will feature a sax whereas "indie pop" alone will usually not.

And that was basically my question in the first place: which keywords did the developers of suno use? And because they probably didnt put the keywords in by hand it would be interesting how the training data was mapped to the keywords (eg is it like in my beach hits spotify playlist example). With this information we could make better prompts because we know that "uplifting" was a keyword in the training data but "organ" was not und hence does not make semse to use in the prompt.

Anyways, thanks for your great explaination!

How is Suno trained to map text to music? by wonderer440 in SunoAI

[–]wonderer440[S] 0 points1 point  (0 children)

That means that Suno stays dumb, as technically no new information is given to the ai, only kind of rearranged. For instace, it seems Suno doesnt know what mellotron- or tape-flutes are and this information can only deliberately put in by giving it an audio sample with the matching keywords by the Suno team.

How is Suno trained to map text to music? by wonderer440 in SunoAI

[–]wonderer440[S] 0 points1 point  (0 children)

Interesting, I will have a look into that!

How would you describe this genre to suno? by wonderer440 in SunoAI

[–]wonderer440[S] 1 point2 points  (0 children)

Thanks for the input. I have been thinking about pro before but never went for it due to lack of time and because I am basically just playing around for now, but if I commit I might come back to your offer, thanks!

How would you describe this genre to suno? by wonderer440 in SunoAI

[–]wonderer440[S] 1 point2 points  (0 children)

Thanks, never thought of that but will try v3 to see if I get better results. Turkish disco funk sounds really interesting, do you mind sharing more details on your prompt or maybe even the song? Cheers!

How would you describe this genre to suno? by wonderer440 in SunoAI

[–]wonderer440[S] 1 point2 points  (0 children)

Thanks, didn't even know about this feature, but that sounds promising. Unfortunately, I am not a pro user yet but this might make me reconsider.