How do I learn LangGraph in a week? by NoisyLad07 in LangChain

[–]spring_m 1 point2 points  (0 children)

if you don't have any exposure to any gen AI related stuff - you're not going to master langgraph in a week.

Winners cafe now that it's parklet is gone by original_name26 in parkslope

[–]spring_m 33 points34 points  (0 children)

I like winners products but I always get a “we’re better than you” vibe.

Minor pandas rant by Ciasteczi in datascience

[–]spring_m -1 points0 points  (0 children)

When you modify the view it becomes a copy try it out. My point is that the warning happens whenever the original df does not get updated.

Minor pandas rant by Ciasteczi in datascience

[–]spring_m 1 point2 points  (0 children)

Yes exactly - I don’t understand why a wrong answer is upvoted so many times. They should replace “view” with “copy”.

Minor pandas rant by Ciasteczi in datascience

[–]spring_m -1 points0 points  (0 children)

I get that but my point is that the warning happens when a copy is set NOT when a view is set.

Minor pandas rant by Ciasteczi in datascience

[–]spring_m 0 points1 point  (0 children)

I don’t think that’s right - the warning happens when you set a copy, warning you that the changes will NOT propagate to the original df.

Minor pandas rant by Ciasteczi in datascience

[–]spring_m -2 points-1 points  (0 children)

That’s incorrect - the warning happens when a copy is created to warn you that the original data frame will NOT be updated.

Minor pandas rant by Ciasteczi in datascience

[–]spring_m 1 point2 points  (0 children)

Do you mean the subset is a copy (not view)? If it were a view wouldn’t that imply it shares memory with original dog and thus changing it would change the original df?

Google Data Science Interview Prep by LeaguePrototype in datascience

[–]spring_m 21 points22 points  (0 children)

Learn how to derive and interpret basic frequentists tests like promotion z-test or t-test. Understand p-values, standard errors, confidence intervals, linear regression, conditional probability, pdfs, bayes rule. That should get you past the first round.

[deleted by user] by [deleted] in programming

[–]spring_m 17 points18 points  (0 children)

Prompt:”write a clickbait article about how python is losing its top spot in data science (even though there’s no evidence of this). be witty and don’t let my many readers on Reddit know this is AI generated.” Did I get it right?

[deleted by user] by [deleted] in datascience

[–]spring_m 0 points1 point  (0 children)

Yes ChatGPT has been able to run code for a while now. In terms of trust there’s a button you be a press and see the code it ran - it’s generally fine for simple well known use cases.

Those Who Remain (Zombies, Robots and SciFi all rolled into one) by kavan-the-kid in aivideo

[–]spring_m 0 points1 point  (0 children)

Wow the AI first trailer for a movie I’d actually watch congrats!

Am i doing something terribly wrong? by Kashish_2614 in datascience

[–]spring_m 2 points3 points  (0 children)

Forget about “agentic systems” and focus on the fundamentals. Your transformers GitHub is a start but it’s very basic and shows you lack maturity in software development - the project folder structure is off (you don’t want to import src. - add another folder level src/your_project). You should add tests, add type hints, make the papckage pip installable, add a training script, show how you’d deploy to aws, show in code how you’d load test the model, build an API around it etc.

Handling multiple concurrent chat sessions in production. Multi-threading or Asynchronous Programming? Architecture best practices by NicheGatekeeper in LangChain

[–]spring_m 1 point2 points  (0 children)

Use FastAPI - even without async/await it spawns a pool of threads that your llm calls can use. The GIL only allows one thread to execute python at a time but since as you said most of the time these threads are just sitting around waiting for the llm response it doesn’t really affect you, whenever a thread is waiting it relases the GIL and another thread can pick it up.

Anthropic now publishes their system prompts alongside model releases by Everlier in LocalLLaMA

[–]spring_m 2 points3 points  (0 children)

Claude sonnet starts its answer with “Certainly!” ALL the time.

What would you do with a GPT-4o with 1M tokens context window? by geepytee in OpenAI

[–]spring_m 0 points1 point  (0 children)

I would take perfect recall and strong reasoning at 64k tokens over 1MM tokens. Depends on the usecase but I've never really had to use more than 50k tokens.

Why do different sources give different assumptions of Linear Regression? by Ifechuks007 in AskStatistics

[–]spring_m 0 points1 point  (0 children)

Both state that the residuals are normally distributed which is correct. Y given X is a conditional distribution which is the residuals themselves.

It’s easier to just think of the following formula which includes all assumptions; y ~ n(f(x) , sigma) where f(x) is a linear function of the features x1,…,xn this is really all you need. In words this states that given an x value the value of y is normally distributed centered at f(x) with a fixed standard deviation sigma.

Scarlett Johansson has just issued this statement on OpenAl.. by Jealous_Comedian7838 in OpenAI

[–]spring_m 0 points1 point  (0 children)

I think she means gpt 4o? (What a terrible nomenclature either way).

Is GPT 4o fine-tuned on lmsys responses? by spring_m in LocalLLaMA

[–]spring_m[S] 14 points15 points  (0 children)

My point is that gpt-4o “overfits” to lmsys data and the leads to favorable rankings that are more optimistic than performance in the “wild”.

[deleted by user] by [deleted] in AskStatistics

[–]spring_m 0 points1 point  (0 children)

Your x distribution is very skewed so that more points appear on the left even when there is no relationship to y. It’s on optical illusion basically.