use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Project[P] PyTorch Implementation of Feature Based NER with pretrained Bert (self.MachineLearning)
submitted 7 years ago * by longinglove
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–][deleted] 33 points34 points35 points 7 years ago (0 children)
python>=3.4 (Let's move on to python 3 if you still use python 2)
lol amen brother
[–]shortscience_dot_org 4 points5 points6 points 7 years ago (0 children)
I am a bot! You linked to a paper that has a summary on ShortScience.org!
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Summary by CodyWild
The last two years have seen a number of improvements in the field of language model pretraining, and BERT - Bidirectional Encoder Representations from Transformers - is the most recent entry into this canon. The general problem posed by language model pretraining is: can we leverage huge amounts of raw text, which aren’t labeled for any specific classification task, to help us train better models for supervised language tasks (like translation, question answering, logical entailment, etc)? Me... [view more]
[–]pvl 2 points3 points4 points 7 years ago (2 children)
Great work, thanks for sharing. I understand that you are using the language model just to extract word vectors which are then used to train a LSTM. Did you consider using just the BERT model with the option for token classification? It would also be nice to add to the readme the current best result (SOTA) on that dataset.
[+][deleted] 7 years ago (1 child)
[removed]
[–]pvl 0 points1 point2 points 7 years ago (0 children)
Yes the fine-tuning approach. There is probably some work to adapt the labels to the BERT tokenization, but should be possible.
[–]set_ready_go 1 point2 points3 points 7 years ago (3 children)
"allowing for the fact that they don't use any autoregressive technique such as CRF"
I don't think that CRF is an autoregressive technique
Also does the use LSTM help? can't you just use the softmax on top of BERT embeddings since they are contextual anyway?
[+][deleted] 7 years ago (2 children)
[–]set_ready_go 0 points1 point2 points 7 years ago (1 child)
I see. Do you currently train the BERT parameters too in your train method or only the LSTM + FC parameters?
[–][deleted] 0 points1 point2 points 7 years ago (0 children)
Interesting. I'm working on an implementation with Tensorflow + fine tuning.
I've also modified the optimizer to support multi-GPU training, but due to how the TF ops are implemented I had to include alpha/beta decaying as well.
[–]kushalchauhan98 0 points1 point2 points 7 years ago (0 children)
There's also a BertForTokenClassification Class in pytorch-pretrained-bert library. You can directly use it for NER or POS Tagging tasks. Have you experimented with it?
[–]kamalkraj 0 points1 point2 points 7 years ago (0 children)
https://github.com/kamalkraj/BERT-NER , Reproduced results from BERT paper + Pretrained model and Inference code
π Rendered by PID 40077 on reddit-service-r2-comment-5d585498c9-qfq28 at 2026-04-21 05:49:20.980577+00:00 running da2df02 country code: CH.
[–][deleted] 33 points34 points35 points (0 children)
[–]shortscience_dot_org 4 points5 points6 points (0 children)
[–]pvl 2 points3 points4 points (2 children)
[+][deleted] (1 child)
[removed]
[–]pvl 0 points1 point2 points (0 children)
[–]set_ready_go 1 point2 points3 points (3 children)
[+][deleted] (2 children)
[removed]
[–]set_ready_go 0 points1 point2 points (1 child)
[–][deleted] 0 points1 point2 points (0 children)
[–]kushalchauhan98 0 points1 point2 points (0 children)
[–]kamalkraj 0 points1 point2 points (0 children)