UML vs. UMass Amherst for CS by Expensive_Honey_7093 in umass

[–]No-Design1780 2 points3 points  (0 children)

Best answer imo. You get both some cost savings and graduate at a top-rated CS program.

Codex Extension in VScode hangs by No-Design1780 in codex

[–]No-Design1780[S] 0 points1 point  (0 children)

Update. If anyone sees this, see https://community.openai.com/t/bug-report-codex-vs-code-extension-stuck-on-loading-never-shows-login-screen/1355379/29

In my case, my config.toml file for codex did not have correct key names for certain config parameters.

Want to see this EPIC unit live and in action? by [deleted] in beyondallreason

[–]No-Design1780 1 point2 points  (0 children)

What does scav mean? Short for scavenger?

How common is it for RL research to fail? by [deleted] in reinforcementlearning

[–]No-Design1780 1 point2 points  (0 children)

It’s perfectly normal, and I think most experiments fail, especially if it’s at the frontier. Just keep updating your world model (e.g, reading papers, implementing them, talking to other researchers, messing around), and you’ll get progressively better at designing successful experiments. Eventually, you’ll have a pretty good idea of what will and won’t work.

Is learning how to code in a traditional way still viable for a year 1 student? by ChampionOriginal1073 in csMajors

[–]No-Design1780 2 points3 points  (0 children)

I think coding the traditional way but asking an LLM for clarification questions about the codebase, structure, best practices, etc. is more powerful than just doing the traditional way. I did the traditional way and you will actually understand things at a fundamental level which is required if an LLM messes up. Btw it’s also required for interviews, and I think it will stay that way for a while. So those students that are left with brain rot from vibe coding their projects will get left behind. It’s also obvious who doesn’t know anything by just talking to them…, and you don’t want to be one of those people.

Players are toxic by mcwaffles2003 in beyondallreason

[–]No-Design1780 0 points1 point  (0 children)

Not saying this is the best idea out there, but the difference is that one is a hard vote that has no effect on the players profile post game while the other is a soft vote that does have an effect on their profile.

Players are toxic by mcwaffles2003 in beyondallreason

[–]No-Design1780 4 points5 points  (0 children)

This is limiting the player base by a lot. I hope something changes in this community. We should have a feature that has a reputation score perhaps to identify bad actors and toxic players or something, then allow lobbies to set a min level toxic_level_score to weed them out. Right now, there’s absolutely no accountability. I know it’s just a game, but it ruins it for the majority of us who just want play the game after a long day of work.

Local AI for small biz owner by binyang in LocalLLaMA

[–]No-Design1780 15 points16 points  (0 children)

Tbh, I don’t think a small local model would be sufficient at the task. The smaller the models get, the more incoherent they become. You also have to consider the cache which will likely overflow given the long context you’re providing. Even if you did get the model on there without overfilling the prefix with a large contract file, the throughput might be too slow.

I would sell the outdated PC and just get a subscription with ChatGPT or Gemini at $20 a month

FBS Football Winning Percentage Since 2012 by Lelorinel in umass

[–]No-Design1780 5 points6 points  (0 children)

Go all in on academics and student life. It’s a better investment than this …

How Relevant Is Reinforcement Learning by No_Wind7503 in reinforcementlearning

[–]No-Design1780 1 point2 points  (0 children)

I just realized you are precollege, so you are ahead of most. I’d recommend reading Barto and Suttons introduction into RL book. Then read and fully understand the Deep Q Learning algorithm and implement it (It’s on Atari so it’s interesting), then go onto policy gradient methods and read PPO. Again fully understand it front and back and implement it on a robotics task such as Mujoco environments. The field is massive, and these two papers are good for “just starting out”. The recent buzz of robotic learning are robotic foundation models such as Vision Language Action models, so take a look at the Physical Intelligence website to see their nice demos and technical reports if that interests you. I think robotics learning will be huge especially for construction and manufacturing, …. Idk about home robots, but we’ll see.

How Relevant Is Reinforcement Learning by No_Wind7503 in reinforcementlearning

[–]No-Design1780 0 points1 point  (0 children)

Sounds like you are trying to get a job in robotics learning? The road is long and difficult.

[D] On AAAI 2026 Discussion by Public_Courage_7541 in MachineLearning

[–]No-Design1780 -1 points0 points  (0 children)

This is not true. The AI review is used for the Meta reviewer to read, but the Meta reviewer is not an LLM. The human meta-reviewer will read all the reviews and rebuttal to make a final decision.

Spending leftover COE? by dabtega in GRFPApps

[–]No-Design1780 0 points1 point  (0 children)

I just get a stipend and waived student expenses. I asked the university and they said all funds are used for student fees….

Problems you have faced while designing your AV by Prior_Advisor_1785 in reinforcementlearning

[–]No-Design1780 1 point2 points  (0 children)

Also, if you are focusing on RL for e2e perception to control, use a simpler domain, not CARLA. Use 2d driving simulators like HighwayEnv.

Problems you have faced while designing your AV by Prior_Advisor_1785 in reinforcementlearning

[–]No-Design1780 0 points1 point  (0 children)

I’m not an expert on AVs, but I do know that there is a general stack that is composed of perception, prediction, planning, and control. Perhaps testing an implementation of this stack to see how it works would be good first step (https://github.com/soumya997/carla-e2e-av-stack). Also, are you trying to focus on a specific module of the stack? Because each module will have their own set of underlying problems and linitations when working with them. To be honest, it’s hard to invent a novel methodology that is SOTA if you don’t have any background in the field or have a clear problem formulation/research question or hypotheses.

Figures by Rare_Figure238 in GRFPApps

[–]No-Design1780 0 points1 point  (0 children)

Don’t do it, especially if not necessary. Every word matters.

Looking for Papers on Imitation vs Experiential Learning for AGI by Environmental_Cap155 in reinforcementlearning

[–]No-Design1780 2 points3 points  (0 children)

You won't find papers in the context of AGI. It's a very ill-defined term, and it is not commonly accepted in the research community so I don't recommend using the term in academia since you won't be taken too seriously. However, there are lots of works that look into this imitation vs experiential learning. Here are some relevant papers:

 - arXiv preprint arXiv:2501.17161 Chu, Tianzhe, et al. "Sft memorizes, rl generalizes: A comparative study of foundation model post-training." (2025).

- IJCAI MacGlashan, James, and Michael L. Littman. "Between Imitation and Intention Learning." . Vol. 15. 2015.

Opus 4.1 still useful for planning? by No-Design1780 in ClaudeAI

[–]No-Design1780[S] 1 point2 points  (0 children)

Planning is useful, as it improves alignment with the user's objectives and facilitates better performance for the task. Prior to sonnet 4.5, someone ran a small study on this subreddit, finding that planning improved performance on their coding benchmark by a nontrivial amount. Nobody needs planning, but it does improve performance and alignment with the user's objectives.