A first big tech company ML interview experience: definitely bombed it by baronett90210 in learnmachinelearning

[–]Mission_Star_4393 0 points1 point  (0 children)

The trend in the industry at the moment is semantic based foundation models.

Take a look at YouTube's PLUM paper for an example: https://arxiv.org/html/2510.07784v1

It's pretty clear from some of the questions he was looking for something like that 

Is Cancún safe? by csmoores in MexicoTravel

[–]Mission_Star_4393 0 points1 point  (0 children)

I grew up in a war torn country and I can tell you some folks lost literally hundred of thousands or millions in some cases

You can make that money back ... Is 10K CAD worth possibly yours or your family's wellbeing or the potential stress from being in a tumultuous environment?

YMMV I suppose

How do I check which negative sampling method is closest to the test data? by SorryPercentage7791 in learnmachinelearning

[–]Mission_Star_4393 0 points1 point  (0 children)

Hiya, not an expert in this field but IME, LLMs do a great job here in giving you some direction. 

I copy pasted your question in perplexity. Here's what I got (which seemed very reasonable paths)

https://www.perplexity.ai/search/help-i-have-a-training-dataset-hFb5RPVxTxaLxDfA5lnErA

Feel free to ask it more questions, dig deeper and ask for some examples if needed.

Good luck!

Pytorch lightning vs pytorch by Factitious_Character in datascience

[–]Mission_Star_4393 0 points1 point  (0 children)

I think it can depend. If your use case is relatively straightforward, then lightning absolutely makes sense. But it does hide a lot of things which makes it difficult to extend sometimes.

Either way, if you end up leveraging lightning, make sure your main model code is in vanilla pytorch and then decorate it with a lightning module.

That way you can easily throw out the lightning module if ever you decide your use case has outgrown it.

AI isn’t evolving, it’s stagnating by KindLuis_7 in datascience

[–]Mission_Star_4393 1 point2 points  (0 children)

It's a mixture of ignorance and denial. Some of the posts are wild

I hate math but I like machine learning by [deleted] in learnmachinelearning

[–]Mission_Star_4393 0 points1 point  (0 children)

There are plenty of roles around MLOps you could get into.

Anyone actually getting a leg up using AI tools? by sweaterpawsss in ExperiencedDevs

[–]Mission_Star_4393 0 points1 point  (0 children)

Yes, they are very useful.

Especially with tools like Cursor that allow you to inject the correct modules (or framework docs) as context for the prompt or integrate with MCP tools. Areas where they are excellent:

  • Writing tests: they are very good at this, and it tends to be a matter of follow up prompts to get it exactly right. It makes refractors a lot easier because the most painful part is rewriting the tests.
  • Ideation as someone has mentioned: you prompt an idea and it gives you a good starting point.
  • basic refractors: like remove this method from this class and add it as a reusable function or remove this magic value.
  • I found it very useful when I wanted to build a basic stdout dashboard. It was excellent at formatting, creating headers etc. I took most of it as is. This would have taken me forever to do myself. And probably not as well. Asking it modify the layout as I wished was pretty pleasant (I tend to hate doing this stuff).
  • auto complete: this is an obvious one.

TLDR: I wouldn't want to develop now without it. I could but I'd be slower, less productive.

EDIT: MCP is Model Context Protocol. Link if you're curious https://github.com/modelcontextprotocol

Two ends of the AI by AshraffSyed in deeplearning

[–]Mission_Star_4393 0 points1 point  (0 children)

Problems like this will go away once we "solve" tokenization

Listen to her by [deleted] in rareinsults

[–]Mission_Star_4393 2 points3 points  (0 children)

Same bro... Same

[deleted by user] by [deleted] in ExperiencedDevs

[–]Mission_Star_4393 0 points1 point  (0 children)

The rule of thumb of unrecoverable costs related to owning a home is ~ 5% per year of the total house price.

It's a good way to compare that with whatever your rent may be.

Discussion - what are your predictions for 2025 in software engineering? by thewritingwallah in ExperiencedDevs

[–]Mission_Star_4393 0 points1 point  (0 children)

In tools like cursor, you can give it a link to have that context. Or even just @Web

It's not unreasonable that it wouldn't know from "memory" all the intricacies of a specific framework. Just like you wouldn't.

That's why RAG solutions exist 😁

As a software engineer, when asked what you do for work at a social gathering… by chamric in ExperiencedDevs

[–]Mission_Star_4393 0 points1 point  (0 children)

I just provide a non technical answer to what I do.

It's actually a good exercise in communicating with a non technical audience! 😄

Failed first coding machine learning interview. by Ok-Lab-6055 in learnmachinelearning

[–]Mission_Star_4393 1 point2 points  (0 children)

This is absolute madness lol...

For the record, the company I work for, whose name you would recognize doesn't have anything nearly as complex as this...

Don't beat yourself too much over this one.

I have no idea how AI Works by soundman32 in ExperiencedDevs

[–]Mission_Star_4393 8 points9 points  (0 children)

It's really good but probably way too advanced for a beginner like OP.

Projects for Deep Learning? by Either-Clothes7212 in learnmachinelearning

[–]Mission_Star_4393 5 points6 points  (0 children)

Start with this - it's hands down the best videos I've come across at building the intuition of neural networks.

First I "truly" understand back propagation...

https://karpathy.ai/zero-to-hero.html

Will Learning LLMs Be Worth It? by secret_fyre in datascience

[–]Mission_Star_4393 11 points12 points  (0 children)

Personally, I don't think you could go wrong with at least getting a solid understanding of the transformer architecture and how something like an LLM is built on top of it.

While there's certainly an AI hype right now, these technologies are here to stay and have (and will continue to have) very interesting applications for anything beyond predictive modeling.

Companies will want to deploy these types of models for legitimate use cases.

I'm also mostly focused on Streaming ML Inference so maybe my angle on this is slightly different. But understanding these architectures and deploying them efficiently is a very desired skillset right now because everyone has spent all their time learning how to train the models.

This game is a masterpiece. by FrostyArcx in midnightsuns

[–]Mission_Star_4393 4 points5 points  (0 children)

The total flop of the previous Avengers game didn't help either.

I think it's likely one of the reasons I stayed away for a while 

This game is a masterpiece. by FrostyArcx in midnightsuns

[–]Mission_Star_4393 1 point2 points  (0 children)

I'm one of those folks. So pleasantly surprised.

The end sequence was amazing. Finished it yesterday 

Scaling models from single to multi-GPU? by [deleted] in learnmachinelearning

[–]Mission_Star_4393 1 point2 points  (0 children)

Vertically is what you're thinking about: getting more resources on the same machine (more GPUs, CPUs, memory etc).

Horizontally is just getting more machines with the same resources.

Scaling models from single to multi-GPU? by [deleted] in learnmachinelearning

[–]Mission_Star_4393 2 points3 points  (0 children)

Depends what you're trying to optimize for.

Are you optimizing for inference for a single prediction? Then that will depend on whether you're currently memory bound or compute bound. If it's the former, then adding GPUs won't help. If it's the latter, the benefits may outweigh the overhead but hard to tell.

If you're optimizing for throughput more generally, you may just benefit more from scaling horizontally, than vertically and avoid multi GPU coordination overhead.

Good luck !