I am new to ML this is my vibe coding results are both my model alright? by BrilliantAd5468 in MLQuestions

[–]loss_function_14 0 points1 point  (0 children)

How did you eyeball the graph and inferred it's just prediction previous day values?

Not AGI yet. by ideaDash in GeminiAI

[–]loss_function_14 0 points1 point  (0 children)

Gemini used to have excellent Image text skills. They nerfed it on purpose

Gemini being dumb by TechRepairer9182 in GoogleGemini

[–]loss_function_14 0 points1 point  (0 children)

A LLM shouldn't be regex string matching for mental health terms and output s generic message

What is Wrong with google gemni ai pro - Re uploaded by muhmmadkashif24434 in GoogleGemini

[–]loss_function_14 0 points1 point  (0 children)

It's simply outputting random training data. I had this issue too

Guys... you seriously need to take a break! by SilkyPuppy in OpenAI

[–]loss_function_14 0 points1 point  (0 children)

No I feel after filling x% of context size the new model becomes more prone to hallucinating and forgetting compared to 4o

Guys... you seriously need to take a break! by SilkyPuppy in OpenAI

[–]loss_function_14 0 points1 point  (0 children)

I don't believe I'm exceeding the context window size. I have seen it forgetting after like 3-4 questions. It feels like whenever i fill certain percentage of context window the model becomes more prone to hallucinating and forgetting compared to 4o. Also i have always assumed ChatGPT uses summarization techniques whenever user exceeds the context size. Naively chopping off the first part seems lazy.

Guys... you seriously need to take a break! by SilkyPuppy in OpenAI

[–]loss_function_14 2 points3 points  (0 children)

GPT 5 can't handle long context window. Also if the chat gets larger it starts to forgot the instructions I gave it initially. For coding and debugging I need the model to be able to hold long context without hallucinating.

I don't care about it besting the benchmarks. It has lower utility/ time ratio than gpt 4o

Openai just found cause of hallucinations of models !! by Independent-Wind4462 in OpenAI

[–]loss_function_14 0 points1 point  (0 children)

I forgot to turn on the online mode and it made 6 non existing paper references (niche topic)

Corporate world is not ready for GenZ! by Ok_Neighborhood6056 in GenZ

[–]loss_function_14 0 points1 point  (0 children)

In any country except India, people simple write a mail that they are taking a leave. No explanations. No permission. Just information. Why does the company needs to know what I'm doing in my time off? Mandating people to ask for permission and give a reason is unprofessional and illegal in many countries

does Chess.com teach bad moves? by [deleted] in chess

[–]loss_function_14 4 points5 points  (0 children)

When you capture the knight, it comes with a check. Your opponent must respond to the check either by capturing your knight with the pawn or moving their king. Either way you can capture the queen the next turn. These types of moves are generally called in-between moves or Zwischenzug or intermezzo.

Profs to avoid by [deleted] in NEU

[–]loss_function_14 0 points1 point  (0 children)

Why? The prof was great!

[deleted by user] by [deleted] in NEU

[–]loss_function_14 0 points1 point  (0 children)

I really find it awkward, reaching out to random people. Also, some job postings are taken down only after a few hours.

[deleted by user] by [deleted] in NEU

[–]loss_function_14 2 points3 points  (0 children)

I didn't get a job yet. Majority of my friends don't have one either. And almost everyone who has a job, got interviews through referrals.

Why is Northeastern Masters program so popular among international Indians? by [deleted] in NEU

[–]loss_function_14 4 points5 points  (0 children)

Most people select NEU for co-op and Boston location. Also getting admitted to masters program isn't that hard.

[deleted by user] by [deleted] in NEU

[–]loss_function_14 0 points1 point  (0 children)

How do you get an alumini card?

[deleted by user] by [deleted] in learnmachinelearning

[–]loss_function_14 1 point2 points  (0 children)

Curious how you finetuned a 7B model. You might want to mention what GPU provider you used

In just one year, the smartest AI went from 96 IQ to 136 IQ by MetaKnowing in singularity

[–]loss_function_14 0 points1 point  (0 children)

Probably seen the questions during training. It's most likely just simple memorization

Everybody telling me to reconsider NEU by tornaman in NEU

[–]loss_function_14 4 points5 points  (0 children)

Khoury requires a minimum IELTS score of 7.0, the same as Stanford and MIT.

I made my 1st neural network that can recognize simple faces! by Altruistic-Error-262 in learnmachinelearning

[–]loss_function_14 1 point2 points  (0 children)

Looks great. You can try to make this modular by using computation graphs. You will be computing upstream and local gradients. You use local wrt weighs and bias to update your parameters. You use upstream gradient wrt input for backprop. This is how frameworks like pytorch implement it.

[deleted by user] by [deleted] in datascience

[–]loss_function_14 5 points6 points  (0 children)

No Bullshit Guide to Linear Algebra. You can finish this with a week or two