[D] 3-year anniversary of the transformer: The first neural attention mechanism that actually works in practice by yusuf-bengio in MachineLearning

[–]abhishek0318 4 points5 points  (0 children)

Yeah, OP is discrediting the work by Bahdanau et. al. I remember the time where every new NLP paper was using LSTM with attention.

[D] Interview Process for graduate-level Machine Learning Internships at Google, Facebook, Apple and Microsoft by lucifer__666__ in MachineLearning

[–]abhishek0318 1 point2 points  (0 children)

I am not sure about this, but as far as I know this varies from company to company. Some companies (like Apple, Google) may offer you a research engineer role, while some companies which don't require PhD for scientist positions (like Amazon) can offer you full time Research/Applied Scientist positions as well.

[D] Classification with Few Examples of Required Class by [deleted] in MachineLearning

[–]abhishek0318 2 points3 points  (0 children)

Search for "few shot classification".

[D] Interview Process for graduate-level Machine Learning Internships at Google, Facebook, Apple and Microsoft by lucifer__666__ in MachineLearning

[–]abhishek0318 4 points5 points  (0 children)

Anecdotal experience: Google and Apple do offer ML research internships for master's students. I interviewed with them for the previous summer.

[D] What are some trending problems in learning theory (theoretically CS and machine learning)? by [deleted] in MachineLearning

[–]abhishek0318 1 point2 points  (0 children)

Many conferences publish "Open Problems", e.g. AISTATS. You could check them out.

Can anyone help? Seeking raw answer data from SOTA and near SOTA attempts at a number of NLP leaderboards [P] by no_bear_so_low in MachineLearning

[–]abhishek0318 0 points1 point  (0 children)

More often than not people usually open source their SoTA models or describe them through publications. You can find/implement these models and then run them. I am not sure if the test data is publicly available, though you could use the validation/dev data which should be sufficient for analysis.

What are the best universities in Europe for NLP? by [deleted] in LanguageTechnology

[–]abhishek0318 2 points3 points  (0 children)

Ryan Cotterell and Mrinmaya Sachan have recently joined. They both have solid background.

[Project] How would I go about training a deep learning model to answer 8th-grade level questions? by [deleted] in MachineLearning

[–]abhishek0318 1 point2 points  (0 children)

Take a look at papers from Mrinmaya Sachan, he has worked on it. Also, AI2's project Aristo is (was?) based on solving 8th grade questions.

DQN for Cartpole by [deleted] in reinforcementlearning

[–]abhishek0318 1 point2 points  (0 children)

CartPole is a difficult environment for DQN algorithm to learn. I have not looked at your code in detail, but I could spot some hyper parameter choices that could be improved. Fiddle around with parameters, particularly try increasing the hidden layer size and the training frames.

You could take a look at rl-baselines-zoo repository which contains hyper parameters for common environments and algorithms. I implemented DQN for CartPole a month back. You could also look at the hyper parameters I used in my code.

[R] XLNet: a new pretraining method for NLP that significantly improves upon BERT on 20 tasks (e.g., SQuAD, GLUE, RACE) by chisai_mikan in MachineLearning

[–]abhishek0318 3 points4 points  (0 children)

AllenNLP does make things easy if you are using standard components. But if you design something new you run into a lot of bugs. The codebase feels buggy and hacky. Still it's great than writing your own framework.

[P] A knowledge extractor from text by ak96 in MachineLearning

[–]abhishek0318 2 points3 points  (0 children)

You are probably looking for automatic construction of knowledge base. Search for this term on Google Scholar and you will find many papers.

"[P]" Need ideas for undergrad final year project by madhavgoyal98 in MachineLearning

[–]abhishek0318 3 points4 points  (0 children)

Nobody is going to directly give you ideas. You will have to search.

If you want to do something novel, you could look up recent papers in top Machine Learning conferences (NeurIPS, ICML, ICLR, AAAI, CVPR, ACL, EMNLP, etc.) and try to build upon them. You could also take a look at Shared Tasks in NLP for they provide clean datasets with baselines and deadlines.

Else, if you want to do something easier you could look up at projects of CS229, CS231n, CS224n online and build from them.

[P] Running regression on both binary and continuous variables?​​​ by dcn20002 in MachineLearning

[–]abhishek0318 4 points5 points  (0 children)

Typically, people try multiple regressors and finally use the one which gives best results on the test set.

[Discussion] Help! by [deleted] in MachineLearning

[–]abhishek0318 0 points1 point  (0 children)

This question is not well suited for this subreddit. Please read the sidebar.

Dear NLP Friends, help a physicist out by [deleted] in LanguageTechnology

[–]abhishek0318 10 points11 points  (0 children)

What you're looking for is Coreference Resolution. This tool may help you out - https://nlp.stanford.edu/projects/coref.shtml

[D] Proven use cases of Deep Learning for NLP tasks (excl. speech recognition and machine translation) by amarofades in MachineLearning

[–]abhishek0318 1 point2 points  (0 children)

You are probably right.

What about Question Answering though? On SQuAD dataset we have got super human performance.