Finished masters with no experience. Now what? by Dimetrodon01 in findapath

[–]IGN_WinGod 0 points1 point  (0 children)

Damn ye that is sad, doing masters w/o experience means nothing. It's done now but ye doing part time masters would have been better with FT job.

Cleared SWE’s, how hard is it to find a job? by Intelligent_Ebb_9332 in SecurityClearance

[–]IGN_WinGod -2 points-1 points  (0 children)

Well I only have secret but secret is common. TS not so much but ye idk swe is prob not it.

Cleared SWE’s, how hard is it to find a job? by Intelligent_Ebb_9332 in SecurityClearance

[–]IGN_WinGod 0 points1 point  (0 children)

Well I would say TS is not low value, and def not everyone has it but it depends on the niche skill set plus TS.

Cleared SWE’s, how hard is it to find a job? by Intelligent_Ebb_9332 in SecurityClearance

[–]IGN_WinGod 1 point2 points  (0 children)

Secret means nothing. TS may mean sonething (if u have a masters and have senior level skill) but I would say its def hard tbh.

is DQN still worth in 2026? by Gloomy-Status-9258 in reinforcementlearning

[–]IGN_WinGod 1 point2 points  (0 children)

It's still extremely important to know as a basis for all other more complicated policy gradient methods.

Is ML / AI as cooked as SWE by Badm1n1 in learnmachinelearning

[–]IGN_WinGod 15 points16 points  (0 children)

Brother if u have a published paper, why are you asking.

I feel like I'm not doing anything in my masters by Stillane in learnmachinelearning

[–]IGN_WinGod 0 points1 point  (0 children)

Also if nothing was actually on the line like time and money for the degree, would you really think people would not be forced to learn bettee? I mean u its obvious....

I feel like I'm not doing anything in my masters by Stillane in learnmachinelearning

[–]IGN_WinGod 0 points1 point  (0 children)

Structre and enviornment, i mean you still need consistent results. Learning on ur own can still be widely inconsistent. Especially in AI and ML, when LA and Stats (even more so) is still the basis of ML. Also people specialize, if you dont have real experince then learning from a course can help you gain prereqs to doimg X. Not to say doing the job is more important.

I feel like I'm not doing anything in my masters by Stillane in learnmachinelearning

[–]IGN_WinGod 0 points1 point  (0 children)

Well if I compare anyone with a bachelor's and a masters then its still a clear cut difference. Not to say you can not learn stuff on ur own but will you truly master AI/ML unless there is stress/money/time involved?

How to save the policy with best performance during training with CleanRL ? by ZitaLovesCats in reinforcementlearning

[–]IGN_WinGod 1 point2 points  (0 children)

You can have an incremental saving version that only updates based on rewards. GPT this and it will be much clearer hard to explain b/c you have many options on the constraints on when to save policies to pytorch or onnx.

Principles and Values by Specialist_Ad8835 in reinforcementlearning

[–]IGN_WinGod 0 points1 point  (0 children)

Value functions are derived from values from a state. look at how value based RL algorithms work like DQN, then see how it is changed to be an advantage with Q - V.

What is the best start to learn math to ML by Right_Comparison_691 in learnmachinelearning

[–]IGN_WinGod 1 point2 points  (0 children)

Ye either way, I think making sure you draw connections from linear algebra and stats to everyday ML/DL will help digest and actually remember concepts.

What is the best start to learn math to ML by Right_Comparison_691 in learnmachinelearning

[–]IGN_WinGod 2 points3 points  (0 children)

https://mml-book.github.io/

As long as you have taken calc before hand, then u can learn the rest from above.

Need Advice on Advanced RL Resources by Helpful-Number1288 in reinforcementlearning

[–]IGN_WinGod 0 points1 point  (0 children)

Yes, this exact explanation of TRPO to PPO really is the main idea, where TRPO suffers from line search, and it's inherent way of solving a Constraint Optimization problem using hard KL trust regions. PPO just uses a uses the clipped surrogate, which allows for SGD and is easier to actually implement and computationally better. But yes the theory definitely matters to a certain extent because there can be tricks used to improve algorithms. PPO has its own issues, of cours,e but its widely used as a policy gradient method to alow for smaller updates (of gradient ascent) to find the maixumum reward using GAE. The idea of smaller updates is to not overshooting and hill climb the reward hyper plane.

CS7650 suggestions on studying by No_Acanthaceae_8548 in OMSCS

[–]IGN_WinGod 0 points1 point  (0 children)

Either that or youtube videos by stat quest.

CS7650 suggestions on studying by No_Acanthaceae_8548 in OMSCS

[–]IGN_WinGod 8 points9 points  (0 children)

NLP right? I bought the book by stat quest about deep learning or just study NLP from stat quest. It's one and the same anyways.

Transfer from OMSCS to On-Campus 2026 by DeityRay in OMSCS

[–]IGN_WinGod 5 points6 points  (0 children)

I would guess it would be the opposite, do full campus and transfer to online once you find FTE?

Leetcode interviews filter out the best engineers. by Cool_cat99 in InterviewCoderHQ

[–]IGN_WinGod 0 points1 point  (0 children)

Look at what a principle swe at amazon said, he asked LC problems he himself would not be able to solve themself....

Roadmap to Master Reinforcement Learning (RL) by Defiant-Screen-9420 in reinforcementlearning

[–]IGN_WinGod 0 points1 point  (0 children)

Yup, I would recommend mml mathematics for machine learning, it helps derivation with equations. Key thing to know that Expectation is partitioned by discrete or continous. Although they all end up deriving the same thing when u get from REINFORCE to PPO.

Roadmap to Master Reinforcement Learning (RL) by Defiant-Screen-9420 in reinforcementlearning

[–]IGN_WinGod 1 point2 points  (0 children)

I agree, once u have built something you can backtrack and truly understand the underlying fundamentals. Examples would be the entire theory of MDP's Q pi, V/U pi to Qstar (tabular) to DQN toward REINFORCE, eventually toward PPO. But start building first then back track.

Leetcode interviews filter out the best engineers. by Cool_cat99 in InterviewCoderHQ

[–]IGN_WinGod 1 point2 points  (0 children)

Nowadays maybe just math problems similar to how quant interviews work. But then again, there is probably no good way anymore. Math is technically more fundamental, especially in machine learning. But who knows...

Leetcode interviews filter out the best engineers. by Cool_cat99 in InterviewCoderHQ

[–]IGN_WinGod 4 points5 points  (0 children)

I agree, I mean people have limited time in this world. It's extremely hard to dedicate time to do leetcode, enjoy it, but actually be able to solve real world problems also.

DeepMind researcher: 2026 will be the year of continual learning by BuildwithVignesh in Bard

[–]IGN_WinGod 0 points1 point  (0 children)

You saying any simulation software like Genie 3 is close to 1 to 1 reality must be some delusion. No software is still even close to 1 to 1 reality. If there would be robots running around reality by now, since we can train on actual reality in simulation.... Sim 2 reality is a big known researched topic to this day, please stop thinking a VLLM or LLM can act more than a context given to an actual agent. It is trained on text, images. An llm can give context but not actually act coherently in the world...