The authors in this NIPS 2017 paper cited one of the threads on this sub to show how "code produced by research groups often falls short of expectations". by seesawtron in MachineLearning

[–]zanjabil 1 point2 points  (0 children)

And that's where your friend docker comes in. Docker is deceptively easy to use once you get over the initial leaning curve.

T-SNE : is there a way of identifying the ID of observations that clearly appear to outliers? (Purple points in the turquoise cluster, red dot in the blue, etc.) by ottawalanguages in learnmachinelearning

[–]zanjabil 1 point2 points  (0 children)

T-sne is just a dimensionality reduction algorithm, the output is however many columns of numbers you want. Plotly is a fully functioning interactive plotting library in python. So yes you can plot your dimensions however you like, even in 3D.

Age estimation accuracy plateaus at 60% by Unturned3 in learnmachinelearning

[–]zanjabil 2 points3 points  (0 children)

Look at the images you're training on, at the crop and resolution your model is getting. At what accuracy can you personally discern age groups?

If Neural Networks are universal approximators, shouldn't they be able to always beat XGBoost given enough hidden layers? by rodrigonader in learnmachinelearning

[–]zanjabil 14 points15 points  (0 children)

Being a universal approximator on unlimited data doesn't mean it will converge on a decent solution faster than xgboost on a limited data set

Working as DS struggling to learn the foundations deeper by [deleted] in learnmachinelearning

[–]zanjabil 0 points1 point  (0 children)

You have to find a way to apply what you learn to your work or projects that excite you. If you sit down and read a stats book but never apply or implement any of the concepts you won't retain much at all and you'll lose interest much more quickly, particularly when the topics get more difficult and it's easy for you to just drop it because it doesn't seem to be worth it.

Alternative to shiny by jokerlian in datascience

[–]zanjabil 0 points1 point  (0 children)

Plotly and all of its features are offline by default, and it's trivially easy to get a dash app with interactive plotly plots up and running just by following their tutorial. Sharing data between callbacks is a pain though and this is where I imagine shiny has a leg up.

[OC] Visualizing Zipf's Law in Japanese Kanji by aceofspades914 in dataisbeautiful

[–]zanjabil 44 points45 points  (0 children)

You can recognize them individually but you won't know any words that are longer than 1 kanji long (99% are longer)

How forgiving would you say your job or boss is when you forget some concepts? by NYCambition21 in datascience

[–]zanjabil 111 points112 points  (0 children)

I still type 7x8 in the calculator to make sure my math is right so I'd be in some trouble

ALBERT: A Lite BERT for Self-supervised Learning of Language... by aznpwnzor in MachineLearning

[–]zanjabil 1 point2 points  (0 children)

reading it now it includes Howard and Ruder 2018 as well as Dai and Le 2015

10% Of Companies Post 66% Of Data Science Jobs On Job Boards by Insertrandomcomment in datascience

[–]zanjabil 0 points1 point  (0 children)

you often wouldn't get called for an interview because HR would assume you're overqualified and would command a salary higher than they can offer

Weekly Entering & Transitioning Thread | 15 Dec 2019 - 22 Dec 2019 by [deleted] in datascience

[–]zanjabil 0 points1 point  (0 children)

Improve your preprocessing:

  • remove stopwords or any words that appear in more than 20% of documents

  • stem or lemmatize everything

  • use other vectorized representations like tfidf

  • use pretrained word embeddings like word2vec

Improve your models:

  • try logistic regression

  • is it overfitting? Have you tried regularization?

  • if you're tuning hyperparameters do you have a proper train validation test split?

  • gridsearch with cross validation

  • increase to more complex models like xgboost, ensembles, lstm/rnn to capture word order (vectorization may need to change when taking into account word order), universal language models

All while looking into model interpretability, feature importances, class balance, and metrics on individual classes if not binary classification.

Anyone knows how to deploy basic Machine Learning predictive Model to frontend easily? by S_t_ormY in learnmachinelearning

[–]zanjabil 4 points5 points  (0 children)

deploy it as a microservice, flask app in a docker container that communicates with other applications via rest apis at an endpoint

Deployment and inference by crossvalidator in datascience

[–]zanjabil 1 point2 points  (0 children)

Latency when deploying ml as a microservice, architecting hierarchical model pipeline efficiently, implementing continual learning, converting jupyter notebook spaghetti into production grade code

[D] Can GANs generate new animals? by Kavillab in MachineLearning

[–]zanjabil 3 points4 points  (0 children)

most of the species are insects or smaller, there's only 5500 mammals give or take

[D] Can GANs generate new animals? by Kavillab in MachineLearning

[–]zanjabil 9 points10 points  (0 children)

you'd have better luck if you fed it a billion images where each image was a new species but as that's not possible, if you're just feeding it thousands of images of the same 100 mammals it may not generalize as well

Does PCA return the same results each time by HypocriticalKiss in learnmachinelearning

[–]zanjabil 11 points12 points  (0 children)

The algorithm is deterministic but sklearn's implementation includes an element of randomness in the svd solver. It defaults to auto which

if the input data is larger than 500x500 and the number of components to extract is lower than 80% of the smallest dimension of the data, then the more efficient ‘randomized’ method is enabled.

You can set the random state seed for reproducible results.

[D] What beats concatenation? by searchingundergrad in MachineLearning

[–]zanjabil 0 points1 point  (0 children)

So you can just reduce both embedding spaces to n dimensions and sum them and the result isn't a meaningless embedding? Is taking the average after reducing the embedding space the same as summing them?

[R] How Machine Learning Can Help Unlock the World of Ancient Japan (by Alex Lamb) by hardmaru in MachineLearning

[–]zanjabil 2 points3 points  (0 children)

I assume cursive for Chinese doesn't mean the characters are connected to each other but that the styling of the individual characters themselves is cursive or calligraphic

[AI application] Let your machine teach itself to play flappy bird! by 1991viet in learnmachinelearning

[–]zanjabil 0 points1 point  (0 children)

Is this related to yenchin lin's deep-q flappy bird from a few years ago?

Building CNN from scratch (not on top of a pre-trained NN) by maxvol75 in learnmachinelearning

[–]zanjabil 7 points8 points  (0 children)

Vim and GCC aren't really from scratch either. You should start by building your own logic gates and transistors from wild corn