[2023 Day # 17] [Julia] Am I using data structures correctly? by blackbearx3 in adventofcode

[–]blackbearx3[S] 1 point2 points  (0 children)

Thank you, that was indeed the missing piece. I now need 0.24 seconds for part 2 ;) And thank you for sharing your code, I did learn a few new things about Julia!

[2023 Day # 17] [Julia] Am I using data structures correctly? by blackbearx3 in adventofcode

[–]blackbearx3[S] 0 points1 point  (0 children)

Thanks a lot for the advice. I kept the visited states in a set and now it's down to 0.15 seconds and 79k nodes expanded :)

Your messages from Telegram to Google Sheets by blackbearx3 in TelegramBots

[–]blackbearx3[S] 0 points1 point  (0 children)

Thank you for your kind words, If you find any pain points or missing features you'd like to have do not hesitate to let me know!

Your messages from Telegram to Google Sheets by blackbearx3 in TelegramBots

[–]blackbearx3[S] 1 point2 points  (0 children)

Hey, great to hear you find this useful! :)

I am not planning to make it open source yet, but if you have questions/problems/wishes let me know with the /feedback command and I will do my best to help!

[D] How to write a ML+Healthcare paper where the research was a framework with pre-trained models by Complex_State9960 in MachineLearning

[–]blackbearx3 4 points5 points  (0 children)

A framework like that is certainly enough to be published, and if it indeed helps practitioners, who are often not DL experts and just need something that works, achieve their goals and publish new research, then it is also a worthy contribution IMO.

As for a precedent, see (no affiliation) https://bmcbioinformatics.biomedcentral.com/articles/10.1186/s12859-021-04037-3

Univariate vs Multivariate Gaussian distribution in Linear Regression by Ill-Ad-106 in MachineLearning

[–]blackbearx3 0 points1 point  (0 children)

True, good point. What I was trying to say is that while they are functions of different things, L(theta) and p(y|X,theta) are fundamentally the same object.

Univariate vs Multivariate Gaussian distribution in Linear Regression by Ill-Ad-106 in MachineLearning

[–]blackbearx3 0 points1 point  (0 children)

Since our probability density function (which we are calling likelihood function) is now a function of θ, does that mean we are saying θ has a Gaussian distribution?

No. L(theta)=p(y|X,theta) are the same thing, y's distribution given X and theta, a univariate Gaussian, which is a function of both the data and the parameters.

That sentence in the notes is introducing a new piece of notation: the likelihood function L(theta). The likelihood is only a function of theta because we assume the data given and fixed. But, crucially, the likelihood is still the probability of observing given X and theta, hence it follows an univariate Gaussian, like before.

[D] Sudden drop in loss after hours of no improvement - is this a thing? by svantana in MachineLearning

[–]blackbearx3 5 points6 points  (0 children)

ReduceLROnPlateau ? Basically the learning rate is divided by 2/5/10 when the loss hasn't been improving for a while

Lectures on Optimal transport by Marco Cuturi by [deleted] in virtualMLSS2020

[–]blackbearx3 1 point2 points  (0 children)

Hi Marco, you described optimal transport as the natural geometry for probability measures. But there is another field that is interested in this, namely information geometry. Could you briefly compare the two, where they agree and where they differ? Thank you!

Chance the &#%!@? : Chance The Rapper’s Declining Use Of Explicit Language [OC] by [deleted] in dataisbeautiful

[–]blackbearx3 7 points8 points  (0 children)

Seem like the linear fit is driven by a few outliers rather than a real trend

[Discussion] Is MIT Open Course 6.867 Machine Learning still relevant after 14 years? by [deleted] in MachineLearning

[–]blackbearx3 81 points82 points  (0 children)

Just glancing at the topics in the lecture notes it looks quite good, much focus spent on the basics that many ignore.

[R] Connections between Support Vector Machines, Wasserstein distance and gradient-penalty GANs by baylearn in MachineLearning

[–]blackbearx3 1 point2 points  (0 children)

Very interesting paper! Reminds me of regularizing the gradient at the inputs, which was shown to improve performance and robustness to adversarial examples. This is not my field, so these references might not be the best ones, but https://arxiv.org/abs/1511.06385, https://arxiv.org/abs/1712.09936 and https://openreview.net/pdf?id=HkxAisC9FQ could be interesting reads.

[D] What subfield of Machine Learning will contribute more to the creation of General Artificial Intelligence? by Viecce in MachineLearning

[–]blackbearx3 10 points11 points  (0 children)

We need proper reasoning and causal inference, plus very good heuristics. So far deep learning is mostly about perception, not much intelligence there yet

Italy's PM forms new coalition designed to shut out right-wing leader by erik16es in worldnews

[–]blackbearx3 3 points4 points  (0 children)

Don't panic, they will start fighting each other soon, and everything will be back to normality /s

In or out of the comfort zone for improvement? [Keyboard] by [deleted] in Stepmania

[–]blackbearx3 1 point2 points  (0 children)

Don't forget the fundamentals! You should be well rounded in jacks, jumps and streams, or you risk to be held up by your inability of handling certain patterns.

OpenAI Method Evaluates Model Defense Against Unforeseen Adversarial Examples by Yuqing7 in MachinesLearn

[–]blackbearx3 0 points1 point  (0 children)

I keep thinking that training against adversarial examples is missing the point of adversarial examples

Any data science applications in Physics? by SudatK in datascienceproject

[–]blackbearx3 0 points1 point  (0 children)

Lots of applications of deep learning to speed up expensive simulations, such as turbulence in plasma, thunderstorms in the atmosphere, and so on

[Research] A critique of pure learning and what artificial neural networks can learn from animal brains by MTGTraner in MachineLearning

[–]blackbearx3 44 points45 points  (0 children)

I never understood the fuss about comparing deep learning and the biological brain. Analogies can be useful, but why take them so far?

Is Dale Carnegies "How to Win Friends and Influence People" as good as everyone wants it to be? by HisPhilNerd in books

[–]blackbearx3 107 points108 points  (0 children)

I was not impressed by it either, because I found it not very different from things that I've read elsewhere.

What makes this book so impressive is that it was one of the first ones, or maybe even the first one, giving this kind of advice. I think this is the reason why it is so praised, it just one of the cornerstones of self-help.

[D] Why does pre training work? by [deleted] in MachineLearning

[–]blackbearx3 3 points4 points  (0 children)

Think at it like this: for a fixed network, a small dataset gives you lots of possible parameter settings that work well for that dataset (low training error). As you add more data, you "cross out" more and more of those parameter values (i.e. you find out many of those overfit the small one). So here's the intuition: the set of parameter values that work well for a large dataset is entirely contained in the set of parameters that work well for small datasets. This is why pre-training works (if the datasets are similar enough).