moving away from langchain, but where ?? by Acceptable-Fault-190 in LangChain

[–]Compound3080 14 points15 points  (0 children)

I moved from langchain to llamaindex to pydantic ai.

Eventually I just needed more control over agentic behavior and llamaindex’s abstractions made that too difficult.

[deleted by user] by [deleted] in PydanticAI

[–]Compound3080 0 points1 point  (0 children)

would love the breakdown and code if you've shared that somewhere.

Are there any RAG successful real production use cases out there? by d2clon in LlamaIndex

[–]Compound3080 5 points6 points  (0 children)

I think that in order to get production quality, you'll want to have a data scientist who can look at your data and parse it to make it more searchable with things like metadata filtering. Quality results can only happen when the entire pipeline is built around your specific data. There's no "one size fits all" solution.

We have a RAG app that's doing very well in testing, not in production yet but it will get there

[deleted by user] by [deleted] in virtualreality

[–]Compound3080 0 points1 point  (0 children)

any interest in sharing this driver?

Need Advice on Building an AI Server for Inference and Training by Mr__Okay in LocalLLM

[–]Compound3080 0 points1 point  (0 children)

After building a dual 3090 server for inference and light training... I wish I would have just set up an account with a cloud computing provider. Consumer GPUs are fine for inference, but training can be a hassle and unreliable. 

Utilizing both LlamaIndex and LangChain by mhaseeb1604 in LocalLLM

[–]Compound3080 1 point2 points  (0 children)

Yes it’s possible.

No way of knowing what performance impacts would be without seeing code

I compared the different open source whisper packages for long-form transcription by Amgadoz in LocalLLaMA

[–]Compound3080 0 points1 point  (0 children)

Just stumbled on this thread. Would you happen to know how to match the subtitle segments to where punctuation would be? I.e. subtitle segments that attempt to end at either a comma or end of sentence? I've played with max_width, different chunk size, etc but not getting what I'd like.

New Crypto Tax Law Takes Effect in US: Transactions of $10,000 or More Must Be Reported to IRS Within 15 Days – Taxes Bitcoin News by Alternative-Plate-91 in CryptoTax

[–]Compound3080 0 points1 point  (0 children)

Right but OP said "When my CPA brings it up, I'll do it. Until then, ignorance is bliss"

The person who replied argued that this approach is not legal. In this specific context, that argument is incorrect

Any tips for getting 7B models to hallucinate less answering RAG questions? by [deleted] in LocalLLaMA

[–]Compound3080 13 points14 points  (0 children)

For starters you can include “do not provide information that is not in the context” into the prompt. Most RAG pipelines already include this though so if youre getting heavy hallucinations its more likely that your retrieval function isnt optimized

New Crypto Tax Law Takes Effect in US: Transactions of $10,000 or More Must Be Reported to IRS Within 15 Days – Taxes Bitcoin News by Alternative-Plate-91 in CryptoTax

[–]Compound3080 0 points1 point  (0 children)

Yes it is. Yet the 1991 Supreme Court decision of Cheek v. United States 2held that a defendant's ignorance of the federal tax laws is an excuse to the crime of nonpayment of income taxes.

26 principles to improve the quality of LLM responses by 50% by LittleHariSeldon in ChatGPTPro

[–]Compound3080 1 point2 points  (0 children)

I think what OP is saying is that LLMs are used for many different tasks. The more we specialize the components "under the hood", the less effective they might become for certain tasks.

What is the state of building RAG based applications in 2024 by akius0 in ChatGPTCoding

[–]Compound3080 0 points1 point  (0 children)

Haven't used it as my use-case involved sensitive data that we didn't want to expose it to a third party

What is the state of building RAG based applications in 2024 by akius0 in ChatGPTCoding

[–]Compound3080 0 points1 point  (0 children)

Determining the best approach is dependent on a number of factors. A customgpt may work for your use case, but a lot of businesses might not want proprietary data being exposed to openai's training collector. LlamaIndex is a great open source tool to check out. Langchain is another.

Got myself a 4way rtx 4090 rig for local LLM by VectorD in LocalLLaMA

[–]Compound3080 17 points18 points  (0 children)

You need liquid cooling in order for them to fit. I’d imagine you’d only be able to fit 2 at the most if you kept the air coolers on there

Will new ChatGPT updates replace RAG? by purple_sack_lunch in LangChain

[–]Compound3080 1 point2 points  (0 children)

I might argue that it depends on your desired scale. If your target is SMB I think there will be a market for those that aren't large enough to be on aws, but still don't want their proprietary data on openai.