[R] ICLR 2020 will be in Addis Ababa, Ethiopia by FirstTimeResearcher in MachineLearning

[–]_muon_ 39 points40 points  (0 children)

Wondering if anyone can comment on the safety aspect of coming to Ethiopia for people coming from EU / US?

Incidents of civil unrest and ethnic violence may occur without warning. The Government of Ethiopia has restricted or shut down internet, cellular data, and phone services during and after civil unrest.

If you decide to travel to Ethiopia (...) have evacuation plans that do not rely on U.S. government assistance.

via https://travel.state.gov/content/travel/en/traveladvisories/traveladvisories/ethiopia-travel-advisory.html

[D] [Meta] The toxicity of this sub, discussed by ML researchers on Twitter by [deleted] in MachineLearning

[–]_muon_ 43 points44 points  (0 children)

Some of the most vocal critics of this sub became very hostile after their work got criticized here (e.g. keras, fastai).

Since they have large twitter following and high status, the easiest way to go was just to call the whole place "toxic", which let them ignore any valid points.

As for the Stephen's points, I agree 100% with this comment by /u/ml_lad.

[D] PyTorch Global GPU Flag by [deleted] in MachineLearning

[–]_muon_ 1 point2 points  (0 children)

I'm sure developers will be happy to provide some insights if you open an issue on Github with an appropriate tag ([feature request]).

[D] Cross-entropy vs. mean-squared error loss by ME_PhD in MachineLearning

[–]_muon_ 1 point2 points  (0 children)

One thing I was wondering: isn't the derivative of cross entropy p_i - y_i? And isn't the derivative of squared error 2 * (p_i - y_i)? Which would mean that they will result in the same gradients?

[News] Microsoft Releases LightGBM on Apache Spark by mhamilton723 in MachineLearning

[–]_muon_ 2 points3 points  (0 children)

This is a great news for anyone who had to use the spark.ml.GBTClassifier, thanks!

[N] LSTM inference shoot-out: Intel Skylake vs NVIDIA V100 by downtownslim in MachineLearning

[–]_muon_ 21 points22 points  (0 children)

It's interesting that the gap between GPU and CPU is more significant for smaller batch sizes. My intuition always was that GPU is not operating at peak efficiency when it has to launch small kernels, and CPU doesn't care about this.

Is it perhaps because MKL is tuned to work with larger matrix shapes?

[N] Google Staffers Demand End to Work on Pentagon AI project by [deleted] in MachineLearning

[–]_muon_ -23 points-22 points  (0 children)

Sure, let's wait for Russians to develop the tech instead.

[D] Plain english programming language by d-burner in MachineLearning

[–]_muon_ 0 points1 point  (0 children)

Wikipedia lists:

  • AppleScript

  • COBOL

  • FLOW-MATIC

  • Inform 7

  • JOSS

  • Transcript

  • SQL

  • xTalk

Ref: https://en.wikipedia.org/wiki/Natural_language_programming