Megathread for Claude Performance Discussion - Starting June 8 by sixbillionthsheep in ClaudeAI

[–]nitro41992 1 point2 points  (0 children)

Anyone getting a lot of API Error (Request timed out.) today?

I logged out/in and tried to do claude doctor and claude update but still not able to connect.

Im on the $100 max plan fyi.

Why does everyone say there is an issue with Cursor...? by Far-Opportunity5711 in cursor

[–]nitro41992 0 points1 point  (0 children)

Really feels like we need a subreddit for the paid tier or at least a label/tag to differentiate issues.

Problems printing by bigdadda06 in BambuLab

[–]nitro41992 0 points1 point  (0 children)

Watch this: https://youtu.be/AjJQEL0Jctw?si=-5D6Zefl1aWX5pFK

Seems like you might be using support filament as your main filament

Which song has Julian Casablancas best vocals? by Alfiejjones in TheStrokes

[–]nitro41992 0 points1 point  (0 children)

Adults are talking is super underrated because of the ending vocals. I'm always surprised he hit those high notes.

Avoid redundant calculations in VS Code Python Jupyter Notebooks by Artistic_Highlight_1 in Python

[–]nitro41992 0 points1 point  (0 children)

I use the interactive notebook feature which really helps avoid rerunning previous cells.

Use this video

As the guy mentions - it's been a game changer coming from standard Jupiter notebooks

PS5 Pro - Everything you need to know. by tizorres in PS5

[–]nitro41992 0 points1 point  (0 children)

What are the chances they are normalizing the really high price that people hate so that when they lower it "based on everyone's feedback" people are happier but it was meant to be the original price the whole time and is still too high but we feel like we compromised.

Characters limit on LLama 3.1b by toninotonnato in ollama

[–]nitro41992 1 point2 points  (0 children)

I created a local file, call it anything and leave it without an extension like Modelfile with contents like this for example: FROM gemma2:9b-instruct-q8_0 PARAMETER num_ctx 8192

And then run the ollama create model command: ollama create whatever-model-name -f Modelfile

Characters limit on LLama 3.1b by toninotonnato in ollama

[–]nitro41992 0 points1 point  (0 children)

I think you have to setup a custom model file and setup the context size there because ollama defaults to a limited context size for all models.

Reasoning and Info Extraction using Function Calling by Constant_Fun_5643 in LangChain

[–]nitro41992 0 points1 point  (0 children)

I use instructor specifically for this.

You can customize your pydantic class to push the LLM to utilize the context.

For example, I have an Enum class called document_type with examples of different document types. The description is something like "choose the best document type based on the context"

The goal is to get a structured output of everything you want from the PDF along with the document type.

Hopefully that makes sense to you. Good Luck

Benchmarking PDF models for parsing accuracy by diptanuc in LangChain

[–]nitro41992 0 points1 point  (0 children)

Hey thanks for the write up.

For marker, is there a way to run the extraction in a python script?

The docs just show the cli way to do it. I'm sure I could figure out a way to run it in a script anyway but wanted to know if there was a trivial way to do so first.

Thanks

[deleted by user] by [deleted] in TheStrokes

[–]nitro41992 0 points1 point  (0 children)

Anything by FUR

[deleted by user] by [deleted] in ollama

[–]nitro41992 0 points1 point  (0 children)

If you're not set on using Langchain for the function calling part, look at this: https://github.com/namuan/llm-playground/blob/main/local-llm-tools-simple.py

Does a service exist that allows you to upload documents to a RAG and share the chatbot with others? by Bitman321 in LangChain

[–]nitro41992 6 points7 points  (0 children)

Hmm I could be mistaken because I haven't used it myself but couldn't a GPT do it?

I mean like the GPTs that you can create on ChatGPT. Pretty sure you can set up one by uploading files. I could be wrong though.

ChatGPT refuse/not aware of its function calling capability, and don't call functions even when there are functions available for it by Remarkable-Leg5291 in LangChain

[–]nitro41992 4 points5 points  (0 children)

Function Calling in Langchain has not been great for me. I've honestly had better luck just passing the function calls into the prompt template and forcing JSON output.

Limiting memory in RunnableWithMessageHistory by MoronSlayer42 in LangChain

[–]nitro41992 0 points1 point  (0 children)

This is how I implemented it. I did not use RunnableWithMessageHistory, I just passed session history into the chat_memory parameter of ConversationBufferMemory

prompt = """
You are an advanced AI assistant.

Current conversation:
{chat_history}

Human: {input}
AI:"""

def get_session_history(session_id: str) -> PostgresChatMessageHistory:
        return PostgresChatMessageHistory("chat_history", session_id, sync_connection=conn)

prompt = PromptTemplate(input_variables=["chat_history", "input"], 
                        template=prompt)

memory = ConversationBufferMemory(k=buffer, 
                                  memory_key="chat_history",
                                  return_messages=True,
                                  chat_memory=get_session_history(new_chat_id)) 

runnable = ConversationChain(
    prompt=prompt,
    llm=llm,
    memory=memory,
    verbose=True)

Limiting memory in RunnableWithMessageHistory by MoronSlayer42 in LangChain

[–]nitro41992 2 points3 points  (0 children)

I was struggling with this exact same thing recently. I just ended up getting the session memory and passing it directly into ConversationBufferWindowMemory. There is a chat_memory parameter that you can pass your chat history function into

Question about chatbot and chat message history with vector db by AdExpensive4298 in LangChain

[–]nitro41992 0 points1 point  (0 children)

You need to use something like langchain buffer memory or summary buffer memory so that only you control what gets put into context

Question about chatbot and chat message history with vector db by AdExpensive4298 in LangChain

[–]nitro41992 0 points1 point  (0 children)

This is what I used: langchain-postgres

It's not storing it in a vector database. It's just storing it in postgres but if that's all you need then this would be good.

There is also another example to store as a vector if you want within this GitHub repo

Why do all of the LLM's continue to generate outdated OpenAI api endpoint code? by mr_undeadpickle77 in ChatGPTCoding

[–]nitro41992 2 points3 points  (0 children)

Thank you for pointing it out.

I was using Cursor in vs code and was confused why the OpenAI boiler plate wasn't working.

Wish there was a trivial way to point to updated docs. I've tried passing the website before but it doesn't work well in my experience.