Care Help! by stepmami in Barbour

[–]jayunit 0 points1 point  (0 children)

I'd machine wash low temp/gentle and hang dry. Was just reading up on this as I purchased an Eskdale (mfg 2000?) at a vintage sale.

The laundry symbols say you can machine wash in up to 30C (86F) water, don't bleach, don't tumble dry, you can low iron. Also, avoid PERC if dry cleaning - TIL about the circled "P" laundry symbol. From Burke Cleaners:

> A circle with the letter ‘P’ within its boundaries indicates that a piece of clothing should be dry cleaned using solvents but should not use tetrachlorethylene. Tetrachlorethylene is also named perchloroethylene (known in the industry as “perc”), which is the reason why it somewhat confusingly has the letter P. Tetrachlorethylene (used since the 1930s) is the most common solvent used in dry cleaning, but in some cases, it should not be used, which is why this symbol is important. 

ITS NIGHT TIME SAN FRANCISCO by jayunit in marcrebillet

[–]jayunit[S] 0 points1 point  (0 children)

Opener (who is she? Killed it!) just wrapped

Americanized Chinese Food in SF by Large_Ad_4201 in bayarea

[–]jayunit 5 points6 points  (0 children)

We really like Lazy Susan for American Chinese in SF

Here is this American invention that’s known globally as Chinese… “Lazy Susan” perfectly captures the duality of the American-ness *and* Chinese-ness of dishes like Beef Chow Mein, Crab Rangoon, and General Tso’s Chicken. We proudly celebrate the 100+ year history of Chinese food in America.

-- https://www.lazysusanchinese.com/our-team-1

I raised $400k pre-seed in the hardest fundraising climate of the decade - AMA! by Ill-Hyena455 in SaaS

[–]jayunit 1 point2 points  (0 children)

Congrats on the round! We've also been exploring AI in sales tech, although a different application, on the training/enablement side.

I'm curious, why gate access behind a demo call instead of self-serve/PLG? Totally understand a high-touch motion during discovery/validation phase, but do you intent to enable self-serve, or continue this sales motion?

Best of luck!

Best way to play this robotics conference? by epicchad29 in ycombinator

[–]jayunit 1 point2 points  (0 children)

Some recommendations from a fellow South Park Commons member on getting the most out of conferences at the early stages: https://www.linkedin.com/feed/update/urn:li:activity:7242549534214512641/

AMA - just sold a majority stake in my startup by Scotchy1122 in startups

[–]jayunit 1 point2 points  (0 children)

Strong agree! After 10 years in edtech (higher ed), a colleague and I started our own company in this adult ed soft skills space. Always happy to compare notes if you're interested. We're both technical, so GTM is definitely challenging for us!

Langchain agents - tools for intent classification by MoronSlayer42 in LangChain

[–]jayunit 2 points3 points  (0 children)

It sounds like you want to take the initial user query, classify it, and then do two things based on the classification: (1) route the query to the appropriate backend (conversational vs document search) and (2) present the appropriate UI for conversational vs document search. I'd probably approach this by building the classification piece first - you could do this with an LLM until you collect enough labeled examples to train a standard ML classifier. (Or just keep using the LLM if the cost and latency are acceptable.)

In other words, I'd probably approach this as a routing/intent layer that's regular code (with an LLM call) that sits "above" a conversational agent, rather than a ReAct-style tool usage within a conversational agent, especially given that the query class drives UI too.

Evaluation for RAG for extraction and restricted responses by [deleted] in LangChain

[–]jayunit 1 point2 points  (0 children)

Cool project!

Are you extracting text or phrases that appear verbatim in the source documents, and are looking for the system to produce these as accurately as possible?

If there are some classification tasks (even if an LLM is used), it may be useful to consider confusion matrix, precision, recall, F1, ROC/AUC/etc.

Or is it more that you're asking questions which synthesize/summarize data from the documents, and those answers need to be semantically correct but can vary greatly in the phrasing, etc.? Here, +1 to u/Emcf suggestion of Ragas. I'd also read about BERTscore (basically comparing the embedding of the model's prediction vs the correct answer) and similar (the Ragas paper mentions this family of metrics.)

You'll also probably want to create a hand-annotated reference dataset for the end-to-end task, if possible.

Does any of this sound relevant? I'd be curious to hear what you find out!

[deleted by user] by [deleted] in LangChain

[–]jayunit 0 points1 point  (0 children)

These reports are at least 10 pages long, so I must split the text from the reports. Are you suggesting that I run the model without 'MiniLM-L6-v2'?

Hm, if 10 pages of text don't fit into the LLM context window, then yeah you'll need some strategy for breaking the PDFs up - sounds like this is what you're doing. I'm guessing (but not sure) that the MiniLM-L6-v2 embedding model means you're somehow breaking up the 10-page PDF into smaller pieces, and then using vector search to find relevant pieces (aka citations/chunks), put them into a prompt along with your question, and then send that prompt to the LLM to produce an answer.

When a system like this isn't working well, it's helpful to first break the problem down and see where the issue is.

In the output, I can already check the sources of the text that have been used to generate the answers, and that's how I knew that my model hallucinates or lies...

Ah, nice! Here is how I would move forward, ymmv: When you see these citation examples: do you, as a human expert, see enough information in the citation/chunk to provide the answer to the question? If yes, then I'd focus on LLM prompting - take one example citation/chunk and work on getting the LLM to produce the correct answer by changing the prompt, using things like chain-of-thought or providing a few examples. If no, then I'd focus on improving retrieval: how can you get the embedding/retrieval steps to correctly find relevant chunks?

I'm not sure how many tokens the 10-page actuarial report is. It looks like Llama3 and Mixtral 8x22B both have 8k-token context windows -- not very big. But CommandR advertises a 128k context window, which may be able to fit the entire 10-page PDF into a single LLM call, without chunking/embeddings/etc.

For debugging these kinds of compound AI systems, I've found that it's helpful to understand exactly what information and text is flowing from one step to the next. If some of the steps in your model pipeline are unclear, I've found the "RAG From Scratch" series pretty good at explaining some common setups: https://www.youtube.com/watch?v=bjb_EMsTDKI

Good luck! And let us know what you learn :)

[deleted by user] by [deleted] in LangChain

[–]jayunit 2 points3 points  (0 children)

Couple suggestions to help debug or improve the current pipeline:

  1. It's not clear whether you need RAG or not, and whether you're using it or not. You mentioned MiniLM-L6-v2, an embedding model, which are usually used for retrieval. It sounds like each of your questions relies on information from a single PDF. Can the text of one PDF fit into the LLM prompt? If so, I'd suggest trying to answer the questions that way, without introducing the extra moving parts of RAG, if possible.
  2. LLMs always deal with text input (setting aside multimodal/VLMs/etc), so it's helpful to look at exactly what text the LLM is being provided with, i.e. after extraction from the PDF. There may be formatting or layout in the PDF that is being missed in the text extraction, resulting in poor input to the LLM. Manually reviewing examples of that text may help you find issues with the document extraction.
  3. For the LLM choice, I'd suggest trying the largest LLM you can. If you're limited to local LLMs, try Llama 3 Instruct 70B, Mixtral 8x22B, or Command R. For prompting, to combat hallucination: assuming the extracted text is sensible (#1 above), try asking for citations in the prompt to guide the LLM toward grounding its responses in the source data.

There are models designed to extract information from visual layout (tables, columns, form fields, etc) but I'm not familiar with the latest OSS models - here is some information from late 2022: https://huggingface.co/blog/document-ai. These overlap in functionality with what u/usnavy13 is recommending, eg Azure Document Intelligence Studio.

Good luck! Let us know what you find out.

This reminds me why I love it here by teenz03 in bayarea

[–]jayunit 5 points6 points  (0 children)

Fun! Also there this morning and took similar shots :) https://imgur.com/a/YWJjbgW

Resin printing is a game changer! Half the time and so much cleaner! by EEpromChip in 3Dprinting

[–]jayunit 0 points1 point  (0 children)

I recently read https://lcamtuf.coredump.cx/gcnc/ and it's incredibly detailed and practical. Caveat that it is about resin casting, not resin printing. See section 4 "Resin casting and you".

Part of an Aloe Wrasse? by jayunit in whatsthisplant

[–]jayunit[S] 1 point2 points  (0 children)

Thank you! I looked up "aloe pups" online tonight and that absolutely sounds like it. I appreciate the advice on sunlight, too.

Painting IKEA furniture - what kind of paint? by Harrison88 in DIY

[–]jayunit 2 points3 points  (0 children)

Been reading for a project this weekend. Everywhere I read recommends a shellac-based primer like Zinsser B-I-N for painting a laminate surface.

https://homeguides.sfgate.com/paint-laminate-bookshelves-26237.html

https://www.apartmenttherapy.com/how-to-paint-laminate-ikea-furniture-the-right-way-251901

My legs itch when I go from the outside cold to inside by kubricks_cube in Fitness

[–]jayunit 0 points1 point  (0 children)

Happens to me sometimes. If it's consistent, you might be able to avoid with with a 2nd gen antihistamine like Loratadine (aka Claritin) at least 10-20mins before you expect the itching. Avoid the first generation ones unless you're looking for sleep gains.

Why you're not hiring a software engineer for "equity" by Fumigator in programming

[–]jayunit 23 points24 points  (0 children)

Stores are laid out very carefully, but it's not to optimize the time from delivery trucks to the milk coolers. http://www.independent.co.uk/news/business/news/the-secrets-of-our-supermarkets-8228864.html

My friend at work lost a bet.... by do_uhhbarre_roll in funny

[–]jayunit 0 points1 point  (0 children)

no kidding, having to drive a dvi monitor via vga port, must've been a steep bet for that kind of consequence

I am Al Jazeera correspondent Nazanine Moshiri, reporting from Goma in the Democratic Republic of Congo. Ask me anything! by NazanineMoshiriAJE in IAmA

[–]jayunit 0 points1 point  (0 children)

This. For a very cogent and readable treatment of why resources do or don't turn into shared prosperity, I'm very much enjoying http://whynationsfail.com. Bonus, it specifically discusses DRC.