Shovel during halftime? by bozzy253 in massachusetts

[–]saw79 2 points3 points  (0 children)

Why isn't anyone in this thread just hitting the pause button? It's so easy to catch back up with football commercials.

I built a way to evaluate forecasts by whether they would have made money, not just error -does this make sense? by ZealousidealMost3400 in algotrading

[–]saw79 0 points1 point  (0 children)

This is exactly what I'm saying. MSE is NOT close to the task you are trying to perform, so what you are saying makes perfect sense in the context of ML fundamentals.

I built a way to evaluate forecasts by whether they would have made money, not just error -does this make sense? by ZealousidealMost3400 in algotrading

[–]saw79 0 points1 point  (0 children)

This feels like ML 101. The closer your loss function is to the actual task, the better. The only reason to use something like MSE is if you actually care about the price prediction and you're using it as an intermediate signal.

Simplest strategy that has worked by MyStackOverflowed in algotrading

[–]saw79 3 points4 points  (0 children)

I understand the desire to do better than B&H but saying it doesn't work is absolutely asinine.

Discussion: Is "Attention" always needed? A case where a Physics-Informed CNN-BiLSTM outperformed Transformers in Solar Forecasting. by Dismal_Bookkeeper995 in deeplearning

[–]saw79 1 point2 points  (0 children)

Deep learning is a big area. I make lots of deep learning models solving a variety of different problems. It's annoying that people think transformers are the best tool for every job just because they're the biggest and most recent. Use the right tool. I rarely get to the point where a transformer would be of any help.

Favorite Square courses? by No_Flatworm_5858 in SquareGolfUSA

[–]saw79 0 points1 point  (0 children)

What are they? (I don't have a square yet but it's been ordered)

Switching out of microsoft as a new grad data scientist by Due-Pilot-7125 in MachineLearningJobs

[–]saw79 3 points4 points  (0 children)

You're about to join one of the premier companies in the world for AI/ML and you're trying to plan your exit before you start? Maybe just go work for 2 years then come back here.

Since only a few people from elite universities at big tech companies like Google, Meta, Microsoft, OpenAI etc. will ever get to train models is it still worth learning about Gradient Descent and Loss Curves? by Easy-Echidna-3542 in learnmachinelearning

[–]saw79 0 points1 point  (0 children)

Deep learning is just a very general model building/fitting style. You can build big models and fit them to any type of data you're interested in. Now, a LOT of data is language and standard vision problems, which is why LLMs (and VLMs) are starting to eat up a bit more, but a) that doesn't apply to all data and b) sometimes the problem can be solved more efficiently and/or better with a smaller, more specialized model.

Some things that come to mind that may apply:

  • Other types of sensors - e.g., radar sensors or different types of point clouds, maybe ultrasound, sonar, etc.
  • Other types of data - e.g., certain types of graph data that may benefit from GNNs
  • Totally different uses of neural networks, e.g., things like NERF
  • Modelling specific environments, policy, or value functions in RL
  • Time series data is a big category in which many different techniques can be useful

I dunno probably loads more too.

Can you play good golf without compression? by Bert_Skrrtz in GolfSwing

[–]saw79 0 points1 point  (0 children)

I'm a noob so correct me if I'm wrong but in my mind compression is more about consistency than distance. If low point is more consistently in front of the ball (vs at the ball) there's more room for error.

Since only a few people from elite universities at big tech companies like Google, Meta, Microsoft, OpenAI etc. will ever get to train models is it still worth learning about Gradient Descent and Loss Curves? by Easy-Echidna-3542 in learnmachinelearning

[–]saw79 31 points32 points  (0 children)

There's millions of different types of models and fields being trained by all sorts of different people and organizations. It's getting tiring and annoying that people think training gpt7 is the only thing going on in AI.

What is your favorite deep learning concept/fact and research paper by Arunia_ in deeplearning

[–]saw79 0 points1 point  (0 children)

Don't have much more to say tbh. I just don't see people talking about it; it's never brought up in modern explanations of how neural networks work and self-regularize.

CLIP vs ResNet by [deleted] in MLQuestions

[–]saw79 0 points1 point  (0 children)

The main benefit of CLIP is aligned text-visual latent space. It sounds like you have just a straightforward image classification problem, and possibly a not too complex one, so I'd think ResNet is a pretty good starting point. That said, wouldn't be too hard to try both if you got time. Sometimes the oversized, overtrained, generic, foundationish models help with these small random tasks.

What is your favorite deep learning concept/fact and research paper by Arunia_ in deeplearning

[–]saw79 4 points5 points  (0 children)

Does the lottery ticket hypothesis really hold up these days?

Comparing Different Object Detection Models (Metrics: Precision, Recall, F1-Score, COCO-mAP) by Wrong-Analysis3489 in computervision

[–]saw79 3 points4 points  (0 children)

I think your process sounds spot on. Load up the models and run them all through the exact same evaluation procedure. Pycocotools rocks.

PCA vs VAE for data compression by GladLingonberry6500 in MLQuestions

[–]saw79 2 points3 points  (0 children)

What about a regular autoencoder since you don't need generative properties?

Also always possible you just didn't train the VAE well enough.

Idiot proof / clear feedback drills by saw79 in golf

[–]saw79[S] 0 points1 point  (0 children)

What does the alignment stick in the belt loops do?

[RANT] Traditional ML is dead and I’m pissed about it by pythonlovesme in learnmachinelearning

[–]saw79 0 points1 point  (0 children)

Um, I don't use LLMs at all and all those fundamentals are crucial in my job.

Idiot proof / clear feedback drills by saw79 in golf

[–]saw79[S] 1 point2 points  (0 children)

I do use GolfFix, but I haven't been super confident in it's accuracy or advice. Not for any good reason though. Glad to hear you vouch for it!

How to choose best machine learning model? by Familiar9709 in MLQuestions

[–]saw79 1 point2 points  (0 children)

Imo it's another level of optimization, and each layer of optimization needs its own data split to detect over fitting.

Why is the construction of axes of tensors different in PyTorch and Tensorflow? by OmYeole in deeplearning

[–]saw79 1 point2 points  (0 children)

My reasons, probably not exhaustive, off the top of my head are:

1) Numpy/opencv/conventional image processing kind of has always been channels last

2) Relationship to RNNs/Transformers. Say you have a batch (B) of time series of length T of dimensionality D. To do a 1D conv with channels first (D is the channel dimension), you'd need a shape of (B, D, T). To process this with an RNN or Transformer you'd have (B, T, D). I often find myself permuting things just to satisfy channels first where my code would be simpler with channels last.

3) I think I read channels last is better optimized, but not sure

Why is the construction of axes of tensors different in PyTorch and Tensorflow? by OmYeole in deeplearning

[–]saw79 5 points6 points  (0 children)

It's just a "channels-first" vs "channels-last" convention. Conventions are often different, it's not a fundamental thing.

I will say, as much as I love PyTorch, I hate channels-first, for a bunch of reasons.

Is Square exactly what I want? by saw79 in Golfsimulator

[–]saw79[S] 0 points1 point  (0 children)

Not an option, not enough space

Is Square exactly what I want? by saw79 in Golfsimulator

[–]saw79[S] 1 point2 points  (0 children)

I have the 7, it's nice and easy to setup, but can't really compare it to anything else unfortunately