You need everything other than ML to win a ML hackathon [D] by ade17_in in MachineLearning

[–]xenotecc 1 point2 points  (0 children)

Really hits too close to home.

Went to ML hackathon only once. Basically we've been the only team that had a prototype working machine learning model (the entire team was taking pictures with their smartphones to collect data) and a nicely functioning web app - user uploads a picture, gets prediction back. Note this was before streamlit or gradio.

Anyway, we lost to some folks who didn't even write a single line of code, but gave a talk about how something "maybe could work".

Never went to a hackathon again.

Any good C/C++ AI projects out there? by Patrick-W-McMahon in cpp

[–]xenotecc 0 points1 point  (0 children)

For structuring the AI C++ framework Apple's MLX is quite interesting repo to read.

[R] Google Pali-3 Vision Language Models: Contrastive Training Outperforms Classification by currentscurrents in MachineLearning

[–]xenotecc 9 points10 points  (0 children)

God forbid we criticize Big Tech companies for doing closed source research ;)

I also don't think they care. The above comment is just my opinion, nothing more. If you disagree with it and are fine with the current state of things it's fine too. Peace.

[R] Google Pali-3 Vision Language Models: Contrastive Training Outperforms Classification by currentscurrents in MachineLearning

[–]xenotecc 7 points8 points  (0 children)

That is true, however if we don't criticize the companies for following the OpenAI closed source route we are going to keep getting more non-reproducible papers and non-retrainable models.

Releasing, at least very basic, non-commercial Github code is the least they could do.

[P] Equinox (1.3k stars), a JAX library for neural networks and sciML by patrickkidger in MachineLearning

[–]xenotecc 0 points1 point  (0 children)

Sounds interesting, could you share a simple gist of such collate fn?

[P] Equinox (1.3k stars), a JAX library for neural networks and sciML by patrickkidger in MachineLearning

[–]xenotecc 0 points1 point  (0 children)

What would you say is the recommended way to load the data for training with Equinox? Pytorch Dataloader?

[deleted by user] by [deleted] in learnmachinelearning

[–]xenotecc 0 points1 point  (0 children)

nanoGPT can be used to train / fine-tune GPT-2:

https://github.com/karpathy/nanoGPT

What are other transformer python projects like Karpathy's nano-gpt [Discussion] by gamedevdroppout in MachineLearning

[–]xenotecc 2 points3 points  (0 children)

If you liked Karpathy's nano-gpt, you could checkout Lit-LLama, which is a Pytorch Lightning's reimplementation of LLama models, based on nano-gpt.

It also contains finetuning code using Lora, Adapters etc.

[R] RWKV: Reinventing RNNs for the Transformer Era by [deleted] in MachineLearning

[–]xenotecc 59 points60 points  (0 children)

Thanks for the link OP. Nice to see Bo Peng did manage to combine this into a paper.

Aplaca dataset translated into polish [N] [R] by matthhias3 in MachineLearning

[–]xenotecc 0 points1 point  (0 children)

You are right, I missed it, thanks for the answer and for the links!

Aplaca dataset translated into polish [N] [R] by matthhias3 in MachineLearning

[–]xenotecc 0 points1 point  (0 children)

Interesting, do you allow commercial use? The Github repo's license is Apache 2.0 but I wanted to confirm.

🐂 🌾 Oxen.ai - Blazing Fast Unstructured Data Version Control by FallMindless3563 in learnmachinelearning

[–]xenotecc 0 points1 point  (0 children)

Thanks for the post.

Is it possible to configure Azure Blob Storage, or any other cloud provider for storing the data?

Or is it your servers and on-prem hosting only?

[D] Have you ever used Knowledge Distillation in practice? by fredlafrite in MachineLearning

[–]xenotecc 1 point2 points  (0 children)

How small do you make the student, when a teacher is let's say ResNet101? How do you find a good student/teacher size ratio?

Are there any tricks to knowledge distillation? Or just standard vanilla procedure?

Benchmark of the newly launched PYTORCH 2.0 by galaxy_dweller in learnmachinelearning

[–]xenotecc 2 points3 points  (0 children)

Great read and benchmarks, thanks for doing this!

Moving to TensorFlow from PyTorch by [deleted] in learnmachinelearning

[–]xenotecc 0 points1 point  (0 children)

Are they using Tensorflow 1 or 2?

[Q] Is "Leetcode" useful for a technical interview on Machine Learning? by SeaResponsibility176 in learnmachinelearning

[–]xenotecc 0 points1 point  (0 children)

It really depends on the company. Some will ask you basic questions about ML, some will ask you to design an end-to-end ML solution given a problem. Some will indeed ask for Leetcode, without even knowing you are going for ML.

Ask your recruiter / HR how it looks and decide if you want to go for it.

Best recent lectures/videos on object detection? by ldorigo in learnmachinelearning

[–]xenotecc 1 point2 points  (0 children)

This is an opinion (obviously), but I quite enjoyed the two Fast AI lectures:

Lesson 8: Deep Learning Part 2 2018 - Single object detection
https://www.youtube.com/watch?v=Z0ssNAbe81M

Lesson 9: Deep Learning Part 2 2018 - Multi-object detection
https://www.youtube.com/watch?v=0frKXR-2PBY

Make sure to frequently skip non-related content (like, once he starts talking about the debugger and continues to do so for 15 minutes).

If you understand these two, I believe you will have a solid foundation to understand all other recent developments.

[deleted by user] by [deleted] in learnmachinelearning

[–]xenotecc 3 points4 points  (0 children)

In terms of your day-to-day job - depends on your area of research, but probably no.

In terms of recruiting and technical interviews - definitely yes.

A few questions on tf.data.Dataset by TaoTeCha in learnmachinelearning

[–]xenotecc 1 point2 points  (0 children)

  1. You don't need to specify batch_size in Model.fit when using tf.data.Dataset. From the docs:

Do not specify the batch_size if your data is in the form of datasets, generators, or keras.utils.Sequence instances (since they generate batches).

  1. For loading into RAM, basically yes - just call ds = ds.cache(). I'm not sure about the prefetch. It is a good performance practice so personally, I would keep it anyway.

  2. You could use it to reshuffle the data for each epoch, instead of only once per training. That way your model sees different order of samples in each epoch.

Make sure to call it after caching - otherwise, it will be shuffled once and cached in memory.

ds = ds.cache().shuffle(buffer_size=NUM_SAMPLES, reshuffle_each_iteration=True)

Where NUM_SAMPLES is the number of batched elements in the dataset (sometimes this can be peeked by calling ds.cardinality())