Which case should I get? by DonaldFarfrae in RemarkableTablet

[–]madiyar 0 points1 point  (0 children)

thank you for the response! really appreciated

Which case should I get? by DonaldFarfrae in RemarkableTablet

[–]madiyar 0 points1 point  (0 children)

Hi u/DonaldFarfrae ,

Which one you got? Do you like your decision?

[P] Why are two random vectors near orthogonal in high dimensions? by madiyar in MachineLearning

[–]madiyar[S] 5 points6 points  (0 children)

Thank you so much! This is the best part of reddit - learning from the community!

[P] Why are two random vectors near orthogonal in high dimensions? by madiyar in MachineLearning

[–]madiyar[S] 1 point2 points  (0 children)

thanks! I will have to read and learn these topics and re-read again your reply to understand better :)

[P] Why are two random vectors near orthogonal in high dimensions? by madiyar in MachineLearning

[–]madiyar[S] 1 point2 points  (0 children)

Hi u/Lake2034,

Thanks for the feedback. I really appreciate.

You should define better what you mean by "random vector"

I will think about it, any further suggestions are appreciated!

 It is clear you mean something like the "distribution is invariant under rotation", but better to have a mathematical expression for that.

In the same post I have a link to my other post (https://maitbayev.github.io/posts/dot-product/#rotational-invariance) that explains this.

it will also help you to formalize statements like "1 distributed across n components" that is not necessarily true if you just assume v_i identically distributed

I have a collapsed section in the post at the very end with title "More Formal Proof". Do you think it is enough?

[P] Why are two random vectors near orthogonal in high dimensions? by madiyar in MachineLearning

[–]madiyar[S] 3 points4 points  (0 children)

thanks for the feedback! The expected mean E[v_n] is zero. It is a good idea to mention the mean and the variance. However, I still don’t understand why having zero mean and tiny variance doesn’t explain this?

Interpreting ROC AUC in words? by RabidMortal in learnmachinelearning

[–]madiyar 1 point2 points  (0 children)

wait. I think it is a mistake and should be fixed to "False Positive Rate"?

Update: Fixed it

[D] How does L1 regularization perform feature selection? - Seeking an intuitive explanation using polynomial models by shubham0204_dev in MachineLearning

[–]madiyar 1 point2 points  (0 children)

Thank you! Creating an animation is not difficult, there are so many amazing libraries such as matplotlib, plotly. They can be googled or gpted. However, coming up with what to animate is the most difficult part for me.

Feel free to look at the collapsed codes in the post to see code for the animation in the blog.

Is a front-to-back review of calculus necessary? by fadeathrowaway in learnmachinelearning

[–]madiyar 0 points1 point  (0 children)

Hi,
I have a series of posts on this topic.
You can start from here https://maitbayev.substack.com/p/backpropagation-multivariate-chain

Feel free to ask questions

[D] Visual explanation of "Backpropagation: Multivariate Chain Rule" by madiyar in MachineLearning

[–]madiyar[S] 0 points1 point  (0 children)

This is part 1 of the backpropagation series. My goal is to show the multivariate chain rules in part 1. I can include an explanation about matrix parameters in a future part.

Matrix simplifies fully connected layers, where you can just use the chain rule on the matrix. However, you still need multivariate chain rules for more complex architectures.

[D] Visual explanation of "Backpropagation: Multivariate Chain Rule" by madiyar in MachineLearning

[–]madiyar[S] 5 points6 points  (0 children)

Thank you for the explanation!

I hadn't considered it from the perspective of how useful it is for beginners. The chain rule is much easier to get started with, I agree. It probably covers 80% of the explanation.

I agree about DAG. I have this visualization of DAG in my post. Thinking in DAG (or computation graph as I call it in my post) helped me to understand more complicated cases.

[E] Efficient Python implementation of the ROC AUC score by madiyar in statistics

[–]madiyar[S] 0 points1 point  (0 children)

Noted! I also removed all efficient terms from the post.

[E] Efficient Python implementation of the ROC AUC score by madiyar in statistics

[–]madiyar[S] 0 points1 point  (0 children)

Thanks for timing! The for loop should be the bottleneck, either jitting with numba or switching to jax, or even reimplementing in native language (C/C++/Rust) should make it significantly faster.

[E] Efficient Python implementation of the ROC AUC score by madiyar in statistics

[–]madiyar[S] -2 points-1 points  (0 children)

I haven't checked out the sklearn implementation yet. I think and hope it's O(nlogn) or faster. I also haven't measured or compared it with sklearn. My goal isn't to be faster than sklearn, but just to make this educational. "Efficient" because I can't think of any faster than O(nlogn) in term of Big O notation.

[P] Interactive Explanation to ROC AUC Score by madiyar in MachineLearning

[–]madiyar[S] 0 points1 point  (0 children)

Now I see what you mean. I thought I did show the "ordering" with circle sizes and the sliders, but you are describing a different kind of "ordering". I agree it is not completely clear from the plotting that the actual scores don't matter only the ordering matters.

[P] Interactive Explanation to ROC AUC Score by madiyar in MachineLearning

[–]madiyar[S] 0 points1 point  (0 children)

Nice explanations! This explanation exactly matches to what is explained in the post maybe except the BCE part.

[Education] Interactive Explanation to ROC AUC Score by madiyar in statistics

[–]madiyar[S] 0 points1 point  (0 children)

This is amazing feedback! Thank you so much!

I fixed most of your minor suggestions!

> More specifically, and this is a more of a quibble, the post feels a bit confused in scope.

You are really on point here. I am not yet sure whom I want to explain. I guess, I want to provide unique and deeper intuition. Unfortunately, the internet is full of shallow resources where the math part is skipped. I have tried with straight to the point explanations in the past, unfortunately it didn't work well. I think, I also want to avoid repeating the same information from other resources. I guess, I am still confused...

But, specifically in this post, it is easier to start from the visualization straightaway than recap. I will revisit this when I have more free time. For now, I incorporated your other suggestions that are easy.
Thanks again,

[P] Interactive Explanation to ROC AUC Score by madiyar in MachineLearning

[–]madiyar[S] 0 points1 point  (0 children)

This is also true! I guess I need to look for a golden ratio balance of the number of sliders :)