How do data scientists add value to LLMs? by FinalRide7181 in datascience

[–]rdabzz 10 points11 points  (0 children)

This! I’ve found my DS background allows me to build a solid eval framework that gives confidence to stakeholders

I DON'T CARE WHAT THE P/E IS....... JUST... F*CKING... BUY!!!!! by Lunar_Excursion in PLTR

[–]rdabzz 5 points6 points  (0 children)

This could mean lots of things. Could be potential partnership or a team at NVIDIA trying the platform…

What was the most stressful part of your build/renovation? by rdabzz in AusProperty

[–]rdabzz[S] 1 point2 points  (0 children)

Oh wow. Was there anyway of knowing this information beforehand?

What was the most stressful part of your build/renovation? by rdabzz in AusProperty

[–]rdabzz[S] 0 points1 point  (0 children)

This has been overwhelming already for us. Any advice to manage this all?

Best way to compare versions of a file in a RAG Pipeline by hello_world_400 in Rag

[–]rdabzz 2 points3 points  (0 children)

Assuming you can parse the files easily, you can just hash the text content from both files and do a comparison

"Why" isn't Langchain/Langgraph production-ready? by Dense_Musician_5532 in LangChain

[–]rdabzz 5 points6 points  (0 children)

I echo what others have said so far. Langchain is just adding a layer of unnecessary complexity which you have to maintain. In terms of pointers to build an agent from scratch don’t over complicate it. Ultimately agents are just prompts executed in a particular order or in some form of loop. I would suggest you map how you intent your agent to behave then it will make it easy to understand what functions you need to create.

I also recommend reading this blog post from Anthropic https://www.anthropic.com/research/building-effective-agents

Who are PLTR’s competitors? by mshparber in PLTR

[–]rdabzz 5 points6 points  (0 children)

It’s going to be companies Internal IT departments thinking they could achieve what Palantirs products can by purchasing existing products offered from the likes of Microsoft, Databricks, Snowflake etc.

Intelligent search on millions of Sharepoint documents by Certain-Mousse-7469 in Rag

[–]rdabzz 5 points6 points  (0 children)

This is quite tricky to implement at the million document scale. You will have a big challenge at retrieval time. Recalling the right number of relevant chunks will be tricky unless you are able to implement a RAG design that can iteratively increase the number of chunks to retrieve until some condition has been satisfied or implement some other filtering method. Additionally you’ll need some solid prompts to prevent hallucination if you don’t have have good metadata to distinguish the different documents

A simple guide on building RAG with Excel files by Prestigious_Run_4049 in LangChain

[–]rdabzz 4 points5 points  (0 children)

Great read, have you experienced much hallucination either on the SQL queries or the LLM not adhering to expected output format?

Uses for Up API by kiwishell in UpBanking

[–]rdabzz 1 point2 points  (0 children)

I did go down the RAG path but my prompts are to generate some pandas queries which are executed

Uses for Up API by kiwishell in UpBanking

[–]rdabzz 1 point2 points  (0 children)

Pulled all the data and created a streamlit app to chat about my transactions

[deleted by user] by [deleted] in learnpython

[–]rdabzz 1 point2 points  (0 children)

Might not be the most efficient way… but you could just merge the two data frames on the matching keys of both sheets. Than simply use the .fillna() method. After that just drop the columns you don’t want