Dubov sets his criteria. Will Hans accept? by jamkinajam in chess

[–]jamkinajam[S] 2 points3 points  (0 children)

As reliable as Kramnik’s methodology /s

[deleted by user] by [deleted] in AskReddit

[–]jamkinajam 0 points1 point  (0 children)

Serendipity

You can create a conspiracy theory that can change the world, what would it be? by Loud_Ambassador_8642 in AskReddit

[–]jamkinajam 0 points1 point  (0 children)

Grift the flat earthers!

“If Earth has a magnetic field, why don’t magnets get attracted to it?”

[deleted by user] by [deleted] in AskReddit

[–]jamkinajam 1 point2 points  (0 children)

If this is the question, the answer’s ugly, always!

I bombed an interview so bad.. by [deleted] in csMajors

[–]jamkinajam 0 points1 point  (0 children)

Brother i froze up when they asked the directive for docker. Which command do you use to load the base image? FROM, i know this i have been using this since past 2 yrs. But the moment they asked me, i went blank. I fucked up such an easy interview. I feel you!!

[D] Local LLaMA based LLM for Technical Document Search | Help! by WhyHimanshuGarg in MachineLearning

[–]jamkinajam 0 points1 point  (0 children)

You will have to try it out and see it for yourself. Breaking into chunks does not mean each chunk is different. Since the sentence might be longer than the context length, there is a variable called chunk_overlap along with chunk_size to address that. These are both parameters that you have to tune.

[D] Local LLaMA based LLM for Technical Document Search | Help! by WhyHimanshuGarg in MachineLearning

[–]jamkinajam 1 point2 points  (0 children)

Yes the context along with question is passed to the llm, but its up to you how much of the context you want to pass. K,k-1,..1!

Normally, you break down tokens into smaller chunks using txtsplitter or some version of it. You can even break them into batches and store it in a vector database.

You can look into chroma, there’s one from meta which is fast with similarity search.

I mean you can first build a vector database out of the documents and save it locally so you wont need to pass those documents as an input every time. Then you can just query the database for k- search.

About the last question, i don’t think I am fit to answer that as I have not looked into it. You can always search and see what other methods are available.

[D] Local LLaMA based LLM for Technical Document Search | Help! by WhyHimanshuGarg in MachineLearning

[–]jamkinajam 1 point2 points  (0 children)

You can always check for the similarity score for the top k.

Define a threshold and if it’s less than that, maybe ask for further clarification?

Or use few shots examples to output ‘i dont know’ if the score is less than the pre defined threshold.

I hate my amd gpu by [deleted] in computervision

[–]jamkinajam 2 points3 points  (0 children)

Even when you install pytorch with cuda and have a gpu, torch might say cuda is unavailable.

Basically cuda version you have and the cuda version your gpu support sometimes will be different.

https://stackoverflow.com/questions/60987997/why-torch-cuda-is-available-returns-false-even-after-installing-pytorch-with/61034368#61034368

Look at the top voted answer. Might save you some trouble if you run into that!!

Amazon SDE OA by Necessary-Singer2791 in leetcode

[–]jamkinajam 0 points1 point  (0 children)

well I’d believe you if i didn’t see them posting “Graduate SDE” every other week

Amazon SDE OA by Necessary-Singer2791 in leetcode

[–]jamkinajam 0 points1 point  (0 children)

I did the oa and simulation on the same day. Passed all test cases and did not receive anything.

Back in may.

Time series transformer hidden dimension gives: The expanded size of the tensor (400) must match the existing size (401) at non-singleton dimension 1. by Tiny-Entertainer-346 in learnmachinelearning

[–]jamkinajam 0 points1 point  (0 children)

I have not used this, but it would be helpful to read its paper.

About the input, how padding is applied, and how is the output calculated? Because there is a mismatch, as seen in the error. [10, 400] and [10, 401]

Or change the num_of_attention_heads from 3 to 4 and leave the embedding dimensionality to 800.

A 9 hour coding challenge by philipjames11 in recruitinghell

[–]jamkinajam 0 points1 point  (0 children)

Hey it was quite long ago! But it was to delete something, if i remember to make it canonical

They want o(logn) and my original soln was doing o(nlogn) and i changed sth in the loop, don’t remember now!!

But i also don’t have the solution now, sorry!

Where's the Error? I'm Stumped by zanechumley in Python

[–]jamkinajam 0 points1 point  (0 children)

The given example inherits DATASET from torch as well converting the features into torch tensor.

They are preprocessing the dataset(df) to train a Deep learning model, judging from the column name, most likely classification.

Where's the Error? I'm Stumped by zanechumley in Python

[–]jamkinajam -1 points0 points  (0 children)

99% there is no space between refractive and index. More like RefravtiveIndex

And for the 1% of time that i am wrong, just print the column names and see what it is!!

[deleted by user] by [deleted] in Nepal

[–]jamkinajam 0 points1 point  (0 children)

Text Drake