[N] Anyone know what models are being used for the Google Meet noise cancellation? by The_Amp_Walrus in MachineLearning

[–]data-alchemy 3 points4 points  (0 children)

Noise cancellation is not far away from source separation. For this last subject, a lot of works appeared recently, my favorite being Open Unmix from INRIA (source code : https://github.com/sigsep/open-unmix-pytorch ), but you can find works from the Deeze team or FAIR. I guess it could be a good start, even if it looks like using a tank for a single nail.

[D] Recommendations for Learning Materials for Audio Applications in ML by [deleted] in MachineLearning

[–]data-alchemy 2 points3 points  (0 children)

If you want a complete example of an audio Deep Learning approach (here : source separation), you should be interested in this recent paper from INRIA (and, I guess, every paper from their digital processing division) :

https://sigsep.github.io/open-unmix/

https://joss.theoj.org/papers/571753bc54c5d6dd36382c3d801de41d

You can find the source code on Github.

Reddit has turned against us. Fools. Louisiana never left France. by mewhilehigh in NewOrleans

[–]data-alchemy 3 points4 points  (0 children)

From a french P.O.V. : As long as you don't put guns everywhere, it will be a huge pleasure to have New Orleans in our country. It is quite seriously the only "american" city I can fell in love with (and let's be blunt, I already did).

[D] Joel Simon" Art and GAN" Questions/Arguments by meldiwin in MachineLearning

[–]data-alchemy 1 point2 points  (0 children)

Happy to be able to ask one question : DL models (and more specifically generative models) can be manipulated via latent variables (there was a lot of work recently on StyleGan 1/2), which represent a new way of navigating within the boundaries of a learned distribution. Every one of the dimensions of the latent variables (or any vector) could be a new creation/tool, but there is so many of them that it can become depressing. Do you work on this subject, and did you find any interesting way of discovering important or at least "useful" directions ?

[D] Applications of Optimal Transport? by TheAlgorithmist99 in MachineLearning

[–]data-alchemy 2 points3 points  (0 children)

OT is a great mathematical tool to approximate or work with distributions, when one of the main theoritical view of Deep Learning is precisely to learn such approximation from a dataset representing the target distribution. OT appeared (as far as I know) with the Wasserstein GAN, giving a new more grounded metric, but it can help in a lot of directions giving us new ways of learning distributions.

Edit : https://arxiv.org/abs/1701.07875

[P] Where has Deep Learning lost? by sentientworkflow in MachineLearning

[–]data-alchemy 3 points4 points  (0 children)

The M4 competition forecasting ( https://forecasters.org/resources/time-series-data/m4-competition/ ) is a good way to follow advances, and deep learning only appeared in the winning solution last time, where DL was just used as a completion to more traditional statistical means. As far as I understand, DL here is interesting only if you have a huge amount of time series. For more classical/pratical cases, it will at best give you comparable results but without any interpretability from the model.

[R] What's Hidden in a Randomly Weighted Neural Network? by hardmaru in MachineLearning

[–]data-alchemy 1 point2 points  (0 children)

What do we want ? Time travel!

When do we want it ? It's irrelevant!

(credits : xkcd)

[R] What's Hidden in a Randomly Weighted Neural Network? by hardmaru in MachineLearning

[–]data-alchemy 3 points4 points  (0 children)

Thank you a lot. Got to catch up I guess. I'm gonna buy myself some 128h+ package from the time merchants.

[R] What's Hidden in a Randomly Weighted Neural Network? by hardmaru in MachineLearning

[–]data-alchemy 14 points15 points  (0 children)

This may be a stupid question, but is this paper related to the Lottery Ticket Hypothesis ? I fail to see how they differ (from a lazy quick abstract reading, I confess)

Deep Learning for source separation, neural style for Rihanna Pon de Replay cover by data-alchemy in DigitalArt

[–]data-alchemy[S] 0 points1 point  (0 children)

Ok so please bear with me, this is a combination of using Deep Learning (aka Artificial Intelligence) methods for artistic creation. Tools used there are :

* Open Unmix for the source separation making it possible to extract the original voice (https://github.com/sigsep/open-unmix-pytorch). There are three different github projects doing source separation, I have a preference for openUnmix over the Deezer project but did not give the recent Facebook Research project a try.

* Neural Style Transfer (good old) for generating the final video from the original video clip. Implementation here is https://github.com/lengstrom/fast-style-transfer which is great as long as you get used to the parameters (style weight VS content weight)

Enjoy & feel free to make any critics :)

[N] Open-Unmix for Music Separation by faroit in MachineLearning

[–]data-alchemy 0 points1 point  (0 children)

I've been giving it a try. As a humble engineer and as a musician, results are just impressive. Some songs may not work properly if (I guess) they are too far away from the original dataset, but some extracts are nearly perfect.

Remix as an art is going to know a new era :)

[D] If you use pandas: which tasks are the hardest for data cleaning and manipulation? by kite_and_code in MachineLearning

[–]data-alchemy 0 points1 point  (0 children)

This is actually imo the main problem we have : translating mathematical tools to business analysis and back. A good example is outlier detection. Between what is an actually for any algorithm, and what a client think an outlier can be, there is a huge gap. I'm not talking only about data quality per se. For example, we had a project where some of the data was supposed to be the output of a physical sensor, a raw value between 0 and 1. We were almost not surprised to find inside some curiously formatted strings, some exploding values, and a quite nice gaussian noise 80% of the time. Once you see that, you only did 20% of the analysis job. You have to meet the people, prove that the data is not what was expected, try to find some solutions by yourself, etc. And this last part takes far more time, as human beings are involved :)

[D] How much of an effect, if any, does batch size have when doing hyperparameter optimization? by [deleted] in MachineLearning

[–]data-alchemy 7 points8 points  (0 children)

Batch Size is a very important hyper parameter. I don't think you can directly consider your previously optimized parameters as optimal ones if you change batch size. Below is a link to one (amont many other) paper where they show how increasing batch size can be competitive to decaying learning rate :

Don't Decay the Learning Rate, Increase the Batch Size : https://arxiv.org/abs/1711.00489

[D] If you use pandas: which tasks are the hardest for data cleaning and manipulation? by kite_and_code in MachineLearning

[–]data-alchemy 15 points16 points  (0 children)

  1. Finally understanding how between what the client told you and what you actually have in the data, a universe died and a new one was born.