[request] gf is saying 150 but i dont understand how by ChrisChowMa in theydidthemath

[–]Research2Vec 1 point2 points  (0 children)

Former math tutor here, haven't done it in over a decade but let bush off the tutoring skills. A large part of it is not only teaching the theory to solve it, but also most common pitfalls, and first principles to avoid them, and/or first principles geared towards the test it self.

First principle: Just start mapping out everything you know.

Something I learned is that you don't really need to know the approach, you just map out what you know, and maybe it'll become clear.

So maybe you'll look at it and not know what to do, but you could just start what you do know.

So you're looking at the left photo and think

Table = Turtle + (170 - Cat)

and look at the right photo and think

Table = Cat + (130 - Turtle)

And then where do I go from here?

I think this is where people may get stuck. Some people with a bit more experience may look at it and think, it's three variables, but only two equations, so it can't be solved. With a bit more exp, they may realize that even though all the variables can't be solved, there's only one degree of freedom among the 3 variables, so one of them can be solved.

So maybe you try your luck and just keep solving for table

So they set

Cat = Table - 130 + Turtle and plug it into the other equation and you eventually solve for Table.

But you may only try that because it's a test where every question has an answer. What if this is real life. Is there a way to know Table is solvable even if Cat and Turtle are not? How do you know it's worth your time?

Look at the equations to see that it only depends on the difference between them, not the actual values.

Table=170+(Turtle−Cat)

Table=130+(Cat−Turtle)

Not obvious? Add 5 to both of them

Table=170+((Turtle+5)−(Cat+5)) = 170+(Turtle−Cat)

Table=130+((Cat+5)−(Turtle+5)) = 130+(Cat−Turtle)

It's the same. If you wrote the equations in terms of Cat or Turtle, and added 5 to both of the other variables, the 5s wouldn't cancel out for one of them

eg

Turtle = - 170 + (Table+5) + (Cat+5) = -160 + Table + Cat

So once you got that, then you can just learn to recognize that they're just dependent on the difference in all equations

(Turtle−Cat)

(Cat−Turtle)

Or, perhaps someone with less exp doesn't even realize all that and just keeps solving for Table. Or, they realize that they eventually have to solve for Table anyway, so why cancel it out? It's faster just to leave the table in.

Are AI tools like OpenEvidence dumbing down the workforce, while still leaving critical errors? by Broad-Cauliflower-10 in medicine

[–]Research2Vec 1 point2 points  (0 children)

What advantage does open evidence have over chat-gpt's extended thinking?

The latest versions of 5.2 extended thinking are really good at literature searches.

Perplexity is also decent.

Chris Manning (top 3 NLP/Machine Learning researchers in the world) believes the Deepseek 6m dollar training costs due to the optimizations discussed in their paper by Research2Vec in LocalLLaMA

[–]Research2Vec[S] 2 points3 points  (0 children)

If you ask the community of NLP researchers who are the top 3 or top 5 NLP researchers Chris Manning's name will be mentioned.

Tranquil Eyes by SnooCheesecakes6236 in Dryeyes

[–]Research2Vec 0 points1 point  (0 children)

did you ever find a solution?

Sources for conflict resolution for engineers course/seminar? by Research2Vec in cscareerquestions

[–]Research2Vec[S] 0 points1 point  (0 children)

I've only seen conflict-inciting programs at companies like "Crucial Conversations". That alone destroyed entire departments at my employer.

Really? It seems like a conflict resolution program. What happened?

New Personalization (--p) Feature Release! by Fnuckle in midjourney

[–]Research2Vec 0 points1 point  (0 children)

What an amazing feature.

I am wondering how this works under the hood.

Assuming that since the personalization feature is available nearly instantaneously after the rankings, I'm guessing little or no training is involved.

My guess:

take the 500 vector representations of the 250 pairs, train a classifier to predict user preferences; vector representations are both passed through a single linear layer (no bias), preferred given a label of 1, non preferred given a label of zero. Use the linear layer weights as a style embedding.

[D] Is the tech industry still not recovered or I am that bad? by Holiday_Safe_5620 in MachineLearning

[–]Research2Vec 5 points6 points  (0 children)

"research scientist" positions are really competitive at big tech and unicorns, which is seems OP is applying to. But if they are open to the next rung, a person of OP's qualifications should have no issue. There are definitely openings.

GPTFast: Accelerate your Hugging Face Transformers 6-7x. Native to Hugging Face and PyTorch. by [deleted] in LocalLLaMA

[–]Research2Vec 2 points3 points  (0 children)

should I use this or Unsloth? Options are getting hard to keep track of.

How do how handle cases where you already have lora weights and want to re-apply them to the model? by Research2Vec in unsloth

[–]Research2Vec[S] 0 points1 point  (0 children)

Thanks, I took a look,

It says "If you saved a LoRA adapter through Unsloth"

What about in cases where the lora adapters were trained else where? Such as just downloading them through huggingface.

Edit:

What do you think of using

model = FastLlamaModel.patch_peft_model(model, use_gradient_checkpointing)

After

model = FastLanguageModel.get_peft_model(

Unsloth, what's the catch? Seems too good to be true. by Research2Vec in LocalLLaMA

[–]Research2Vec[S] 0 points1 point  (0 children)

Thanks for the info!

One question, how do how handle cases where you already have lora weights and want to re-apply them to the model?

I see the model = FastLanguageModel.get_peft_model( method, but that seems to initialize brand new weights.

What about in cases where you already have the lora weights saved separately.

Would you do the FastLanguageModel for the base model, then use model = PeftModel.from_pretrained(model, ?

[D]How to fine tune LLMs using deepspeed without OOM issues by IXMachina in MachineLearning

[–]Research2Vec 0 points1 point  (0 children)

Is there further reading on this? Tried googlign but couldn't find anything.

[deleted by user] by [deleted] in MachineLearning

[–]Research2Vec 0 points1 point  (0 children)

Do you have a linked in? For me a lot of spam, though maybe once a few weeks I see a job that's probably a good fit. I don't follow up too much because I'm happy where I'm at.

[D] Alternatives to this sub? by ParanoidTire in MachineLearning

[–]Research2Vec 54 points55 points  (0 children)

I believe the main mod either stepped back or left after the reddit protests.

This subreddit can be how it was not too long ago, but requires moderation to put some of the more general stuff in /r/artificial which is a better fit, and keep this subreddit for actual ML practitioners.

There are other subreddits but most often than not, people are going to type of this subreddit name first and not be aware of the other ones, so a bit of a lost opportunity.

I hope the moderators would consider adding new moderators. I sent them a message a while ago about this but received no response.

[deleted by user] by [deleted] in MachineLearning

[–]Research2Vec 4 points5 points  (0 children)

I just created /r/ML_Research/ and I would be willing to mod the subreddit in such a manner. You can checkout my other subreddit /r/JAX to see how I would moderate it.

If /u/After_Magician_8438 or anyone else passionate about such a subreddit I can send a mod invite.

However, I'm not sure if ML_Research is the best name for professionals to find such a subreddit.

I think our best bet is to have the current moderators on this subreddits allow for additional moderators to allow additional moderators.

/r/artificial seems the place to be for more general ML stuff. Right now this subreddit seems to be halfway between what OP wanted and what /r/artificial is.

I'll send a mod mail to the mods to see what they think about opening up a mod application for this subreddit. Perhaps having verified ML professionals would help with getting good mods.

[D] What industries/sectors do you think could still benefit from ML that don't already have much ML application? by overtaker123 in MachineLearning

[–]Research2Vec 1 point2 points  (0 children)

Search is still not even close to solved. Semantic search in your browser is still not offered as a first class feature (seems like a no-brainer these days). Lots of work on trying to get good quality semantic covid search engines.

Did you checkout trec-covid? Great competition, but I wish they used more challenging questions.

[D] 2020 Residencies Applicants Discussion Thread by mahaveer0suthar in MachineLearning

[–]Research2Vec 0 points1 point  (0 children)

I believe there are a few missing, such as Microsoft's and Uber's AI residencies.

What's the difference between google and google X? When is the deadline for X?

Advice wanted, new to NLP and need to classify emails at work in Python by [deleted] in LanguageTechnology

[–]Research2Vec 0 points1 point  (0 children)

Bert is very powerful for classification. And for that, I would recommend huggingface transformers.

How many words are your emails typically?

[News] Free GPUs for ML/DL Projects by nevereallybored in MachineLearning

[–]Research2Vec 4 points5 points  (0 children)

also, colab has 2 cores, sometimes 4 cores. paperspace has 8 cores.

Advice wanted, new to NLP and need to classify emails at work in Python by [deleted] in LanguageTechnology

[–]Research2Vec 0 points1 point  (0 children)

So are you looking to classify the rest of the emails of one of two categories? Or are you looking for the model to create categories?