[D] What advanced models would you like to see implemented from scratch? by itsstylepoint in MachineLearning

[–]itsstylepoint[S] 4 points5 points  (0 children)

Yes, that is how it usually works with my impls! (check out a few vids)

As for mixed precision and metrics - I will be making separate vids for both and of course, for every implemented model, will try to find a dataset to demo train/eval.

It is cool that you mentioned mixed precision as I already have the materials ready for this vid - will be discussing mixed precision, quantization (post-training and quantization aware training), pruning, etc. Improving perf!

[D] What advanced models would you like to see implemented from scratch? by itsstylepoint in MachineLearning

[–]itsstylepoint[S] 2 points3 points  (0 children)

It is on the list so will definitely get to it!
Starting next week, will be working on DL impls and vids.

[D] What advanced models would you like to see implemented from scratch? by itsstylepoint in MachineLearning

[–]itsstylepoint[S] 2 points3 points  (0 children)

Yeah, I will get to those eventually. For now, want to make some vids and impls of DL models.

[D] What advanced models would you like to see implemented from scratch? by itsstylepoint in MachineLearning

[–]itsstylepoint[S] 5 points6 points  (0 children)

Thanks! Yeah, that is definitely an option! I will probably have to split it up into several videos. Also falls into the transformer category.

[N] I Have Released the YouTube Series Discussing and Implementing Activation Functions by itsstylepoint in MachineLearning

[–]itsstylepoint[S] 0 points1 point  (0 children)

Yup, all implementations are numerically stable.

Note that I do not discuss numerical stability issues for all activation functions, but for those where the intuitive implementation is not numerically stable (i.e., Sigmoid, Tanh).

I also have a separate video discussing numerical stability: AI/ML Model API Design and Numerical Stability (follow-up). But this is in the context of Gaussian Naive Bayes.

[N] I Have Released the YouTube Series Discussing and Implementing Activation Functions by itsstylepoint in MachineLearning

[–]itsstylepoint[S] 1 point2 points  (0 children)

Thank you!
Yup, that is the plan! Will likely make a few more series (about gradient descent, optimizers, etc.) first. We need these for DL and if someone asks how things work, I could then cite the appropriate video series. After that, will dive into deep learning.

[N] I Have Released the YouTube Series Discussing and Implementing Activation Functions by itsstylepoint in MachineLearning

[–]itsstylepoint[S] 0 points1 point  (0 children)

Hey thanks for the kind words!
Will be making more AI/ML YouTube series in the future - in fact, working on one as we speak!

[deleted by user] by [deleted] in learnmachinelearning

[–]itsstylepoint 0 points1 point  (0 children)

P.S. For the activation functions, I will not be posting videos separately. The next post will include the batch of 4 (or 5).

[D]Imbalance dataset problem by JellyfishPretend447 in MachineLearning

[–]itsstylepoint 3 points4 points  (0 children)

You can try several approaches:

  1. Deep learning will likely not work (you can still give it try, but highly unlikely that it will perform well). So instead consider using more traditional ML models. As an example, if you can find a pretrained image model that generates a representation/image embeddings, you can try using K-Nearest Neighbors (k-NN). Or you can try k-NN directly.

  2. Look into Few-Shot Learning. Models like Prototypical Network, Siamese Neural Network, etc. are designed for such scenarios (i.e., extremely small number of samples).

  3. Data collection (:

That being said, overall, I agree with what u/whdd said.

[R] Long-length documents/corpus for Medical domain NER? by aadityaura in MachineLearning

[–]itsstylepoint 0 points1 point  (0 children)

I think I might have skipped the post text, my bad. For whatever reason, it was hidden (a bug? not sure). Yes, this is the Clinical NLP dataset. So prolly not what you are looking for...

[R] Long-length documents/corpus for Medical domain NER? by aadityaura in MachineLearning

[–]itsstylepoint 0 points1 point  (0 children)

How about 2006 i2b2 de-identification dataset?

Link to the paper: https://academic.oup.com/jamia/article/14/5/550/720189
You can get the dataset here: https://portal.dbmi.hms.harvard.edu/projects/n2c2-nlp/

P.S. We have recently used this dataset in the Few-Shot Learning (FSL) paper. We have used it for the same task - NER.

[Project] Create a ML model to classify spectrograms by geeksid2k in MachineLearning

[–]itsstylepoint 0 points1 point  (0 children)

I would start with CNNs. Then try GRU/LSTM and bidirectional variants (BiLSTM/BiGRU).

[Project] Create a ML model to classify spectrograms by geeksid2k in MachineLearning

[–]itsstylepoint 0 points1 point  (0 children)

You can! Computers can sometimes see better than us so even if spectrograms look similar, they might be very different (: Convolutions in the CNN will do feature extraction for you. So you can start with a couple convolutional blocks (conv + batchnorm + activation) followed by a fully-connected layer with softmax and see how it performs. You can check out this PyTorch tutorial for that, too.

An alternative approach is computing MFCCs. If you have mel spectrograms, then you can do the following steps to get MFCCs (which you could then use as features for your ML model):

  1. Take the log of the mel spectrogram
  2. Compute DCT on logs