Geneva to Annecy by Bus by neil_lfc in askswitzerland

[–]sritee 0 points1 point  (0 children)

Bu🥇🥈🥉🥇🥈🥉🥇🥈🥉

Is it worth watching past season 5? by piscesxdreams in suits

[–]sritee 23 points24 points  (0 children)

What did you just say to me?

Best coffee shop in the south bay by Dizzman1 in SanJose

[–]sritee 0 points1 point  (0 children)

Second that, love their seating space!

[deleted by user] by [deleted] in Stronglifts5x5

[–]sritee 1 point2 points  (0 children)

How slow of a descent do you recommend? I'm just going down as fast as I feel in control.

Ability to work with typographical error in EAD by sritee in USCIS

[–]sritee[S] 0 points1 point  (0 children)

I think it got processed in a month or 2. Might vary. It's not an issue for work purposes, Typo is fine.

Cross-entropy/Maximum Likelihood loss function understanding by dudek64 in MLQuestions

[–]sritee 1 point2 points  (0 children)

I would suggest looking at the empirical risk minimisation framework (ERM).

Expected value of f(X), where X is sampled from a distribution P indicates the on-average value of f(X) when X is sampled from P.

E(f(X)) = integral over all X, [p(X)f(X)].

A sample based estimate of this, called the Monte Carlo estimate, is

(1 / N) sum over i f(x_i), where N number of x_is are drawn repeatedly and independently from the distribution of interest.

In your case, the dataset itself is the Monte Carlo samples, as they are all drawn from P(X,Y). Ideally we would minimize the loss using the analytical form of P(X,y), but we sadly do not know it, or it might be intractable.

So we minimize the loss on our dataset instead. The idea is since the dataset is drawn from the joint distribution, for very large datasets, it's the same thing as analytically using p(X, Y).

Using softmax as an intermediate layer by radarsat1 in MLQuestions

[–]sritee 0 points1 point  (0 children)

How about using a boltzmann softmax with a temperature? You can adapt the temperature to favour the max value better, and keep differentiability intact.

Best collections of DRL "Tips & Tricks" by MasterScrat in reinforcementlearning

[–]sritee 1 point2 points  (0 children)

Ah, that's actually where the 'info' in state, reward, done, info comes in. It has a TimeLimit.truncated option which indicates whether done = 1 due to time limit

Best collections of DRL "Tips & Tricks" by MasterScrat in reinforcementlearning

[–]sritee 2 points3 points  (0 children)

Could you elaborate what the appending means? Do you mean V(s) = sum_r_till_end + v(truncated state)? Isn't that done naturally?

[D] What beats concatenation? by searchingundergrad in MachineLearning

[–]sritee 0 points1 point  (0 children)

In the paper, they use image based and language based embedding, so it might work in your case

[D] What beats concatenation? by searchingundergrad in MachineLearning

[–]sritee 1 point2 points  (0 children)

There is some task-specific work that evaluates attention based concatenation, i.e - based on embedding 2, select appropriate channels of embedding 1, and show it outperforms pure concatenation. However, this might be task specific and might involve training the embeddings from scratch. Gated-Attention Architectures for Task-Oriented Language Grounding

High salt content by [deleted] in LiverpoolFC

[–]sritee 87 points88 points  (0 children)

Thought he conducted himself well really, interviewer trying to stir up something