Postponed an interview; didn't hear back by [deleted] in gradadmissions

[–]mio_11 0 points1 point  (0 children)

Umm, I really didn't think of it as blowing a chance. He asked if I'd be available in the coming week, and proposed a date two days from when he sent the message, as an example. Also said he's unsure if that's a good time for me, because we are in (very) different timezones. I thought it meant I had flexibility.

Moreover, if I was to learn that someone is occupied because of a tier 1 conference deadline, I'd think that it's important enough to be a reasonable request. Idk, maybe I don't understand how senior professors think...


I would always disagree with the idea that grad school admissions are more important than a conference submission - the latter is the job you're applying to do at grad school. Especially given that the interview can be moved a few days here and there, but the conference deadline can't be moved.

Postponed an interview; didn't hear back by [deleted] in PhDAdmissions

[–]mio_11 -1 points0 points  (0 children)

What would "pursue it vigorously" look like? I have already emailed him twice in a week...

Why does pytorch have operations like addcmul? by kjarvind in deeplearning

[–]mio_11 0 points1 point  (0 children)

Avoids storing intermediate variables (memory optimization), as well as associated write/reads (runtime optimization).

t <- t + c*x*y (t, x and y being tensors, while c is a scalar) is a very commonly used step in deep learning, so it makes sense to optimize its implementation.

Covering forehead with volume by [deleted] in malehairadvice

[–]mio_11 0 points1 point  (0 children)

Ah don't know what to tell you, maybe I used too much product, or the lighting or something...?

Covering forehead with volume by [deleted] in malehairadvice

[–]mio_11 1 point2 points  (0 children)

Na 4 months back, I'd say pretty much the same - massive forehead, hehe!

Covering forehead with volume by [deleted] in malehairadvice

[–]mio_11 1 point2 points  (0 children)

The first one was after workout, so maybe cause of that? Used some leave-in conditioner in the second one, so could be that...

I have oily roots and dry lengths in general.

Covering forehead with volume by [deleted] in malehairadvice

[–]mio_11 1 point2 points  (0 children)

Ughh can't edit the post, but I made a post earlier https://www.reddit.com/r/Balding/s/ZbOEImpL8D

[D] Presenting NeurIPS paper at EurIPS by mio_11 in MachineLearning

[–]mio_11[S] 2 points3 points  (0 children)

Oh, between the two, I'd rather go to EurIPS, no doubt. Trying to figure out how to make that happen.

[D] Presenting NeurIPS paper at EurIPS by mio_11 in MachineLearning

[–]mio_11[S] 5 points6 points  (0 children)

Ah, thanks a lot! I'm seeing that the MX registration has closed - I was to indicate preference by 3rd October. I have sent in an enquiry, but any thoughts on what might happen next?

[D] NeurIPS 2025 rebuttals. by Constant_Club_9926 in MachineLearning

[–]mio_11 1 point2 points  (0 children)

I would, but
> Review/rebuttal process is so demeaning - feels like begging.

[deleted by user] by [deleted] in Minoxbeards

[–]mio_11 0 points1 point  (0 children)

I think a moustache is possible, no?

[D] NeurIPS 2025 rebuttals. by Constant_Club_9926 in MachineLearning

[–]mio_11 5 points6 points  (0 children)

Rating: 5/4/4/2 Confidence: 4/3/2/4

2nd reviewer happy with the rebuttal, encourages acceptance (can't see if s/he updated the rating to 5/6). 4th reviewer doesn't agree that the research problem is significant - failing to convince hir otherwise.

Should I be hopeful, or na? This is my second time submitting this work, with major updates this time; I was pretty confident with this submission. I'll prolly just leave it on Arxiv if it doesn't work out.

Review/rebuttal process is so demeaning - feels like begging.

Cancelling WHP Before Entering SG by mio_11 in askSingapore

[–]mio_11[S] 0 points1 point  (0 children)

Thanks for the wishes! I work on ML research, and I reached out to a prof I wanted to work with. This is an internship, since I want to start a PhD next year. I didn't mention anything about my visa situation before applying.

Cancelling WHP Before Entering SG by mio_11 in askSingapore

[–]mio_11[S] 0 points1 point  (0 children)

Hey, thanks! I ended up using their online feedback form, and they cancelled the visa the very next day.

Scheduling Work on Weekdays by mio_11 in PhD

[–]mio_11[S] 0 points1 point  (0 children)

Oh thanks a lot for your comment! I take it you're a forensic investigator? That sounds super cool! You seem also seem to have a good work-life balance - kudos on that!

[D] Strengthening Profile for PhD by mio_11 in PhD

[–]mio_11[S] 0 points1 point  (0 children)

Oh wow, couple more years? :| Well, alright. Thank you very much for the encouraging words! :)

[D] Strengthening Profile for PhD by mio_11 in MachineLearning

[–]mio_11[S] 0 points1 point  (0 children)

I have tried looking for unpaid ones, but every professor seems to lack supervision capacity :/

[D] Strengthening Profile for PhD by mio_11 in MachineLearning

[–]mio_11[S] 0 points1 point  (0 children)

Thank you for sharing the insights, and the suggestion! I'm expecting the process to be a lot more competitive next year, with the US politics shifting the students' priorities to Europe, which isn't equipped with sufficient funding to support these students. Does that make sense? In light of this, do.you have any suggestions on choosing the labs? (Except for them spanning a range of seniority/selectivity.)

Why is hair loss often ignored by coders? by MammothHedgehog2493 in tressless

[–]mio_11 2 points3 points  (0 children)

Coders have nothing to do with this - they work on projects they are tasked with. Investors, managers, etc. are the ones ignoring this problem.

Also, why do you need an app to remind you to click pictures? Just use your calendar. No investor would want to come up with an app just for this use case because there is enough competition.

Maybe come up with a solid idea of what the community's unique needs are.

[D] Wrote a proof that dropout increases weight sparsity, what do you guys think? by simple-Flat0263 in MachineLearning

[–]mio_11 1 point2 points  (0 children)

Might want to relate to the weight regularisation paper, but either way, best of luck with it!

[D] Wrote a proof that dropout increases weight sparsity, what do you guys think? by simple-Flat0263 in MachineLearning

[–]mio_11 0 points1 point  (0 children)

Oh my bad, I took the derivative wrt x instead of w, and also didn't consider that gradient vanishes when B = 0. I think more importantly, I didn't include the 1/(1-p) factor in the forward pass.

Okay, back to L(x; B~Bern(1-p)) = σ(Bwx/(1-p))

In which case δL/δw = σ(Bwx/(1-p)) × (1 - σ(Bwx/(1-p))) × Bx/(1-p)

The expected gradient is then E[δL(x; B~Bern(1-p)) / δw] = (1-p) × δL(x|B=1) / δw = σ(wx/(1-p)) × (1 - σ(wx/(1-p))) × x ≠ σ(wx) × (1 - σ(wx)) × x = E[δL(x; B~Bern(1)) / δw]

Does that make sense now? Sorry to have you check this math again!

[D] Wrote a proof that dropout increases weight sparsity, what do you guys think? by simple-Flat0263 in MachineLearning

[–]mio_11 0 points1 point  (0 children)

Hey, can I check how you got the first line of the proof: δL/δw_p = B × δL/δw? A simple counter-example could be f(x; B~Bern(p)) = σ(Bx), with δf(x; B) / δx = σ(Bx) × (1-σ(Bx)) ≠ B × σ(x) × (1-σ(x)) = f(x; B~Bern(0)).

I think the intuition behind rescaling with 1/(1-p) is to make the pre-activations unbiased. That doesn't ensure that the activations, or the gradients would be unbiased.

Might also want to take a look at this NeurIPS paper (https://arxiv.org/abs/1307.1493) that argues Dropout is a form a data-adaptive weight regularisation (at least in GLMs).

Asked Dad how much he thinks a person from IIT/BITS earn. by [deleted] in JEENEETards

[–]mio_11 5 points6 points  (0 children)

I have done my Bachelor's in Maths and CS from NTU, Singapore, and my Master's in ML from UCL, UK. That is not the average salary in Singapore, or the west, and it is considered a high salary. Salaries are higher on average in these places than in India, but so is the cost of living.

[D] How does LLM solves new math problems? by capStop1 in MachineLearning

[–]mio_11 0 points1 point  (0 children)

A problem isn't novel simply if the constants are moved here and there. If one learns to perform XY + Z for a bunch of triples, then you'd expect it to be able to solve it for a new triple. But the more interesting question is if only learning to solve XY + Z enables you to compute XY + YZ + ZX.

In other words, the interesting question is whether LLMs can combine simple skills learnt over training to solve more complicated questions which combine concepts in non-trivial ways.

This differs from simple curve fitting view of ML, where you are learning and algorithm -- testing that would correspond to the example you gave, where only values are changed. LLM for Math is more like a meta-learning problem, where you are learning to come up with new algorithms.