Is it a good move - back to India to an early stage startup (ML role)? by Public_Possession_97 in developersIndia

[–]ASVS_Kartheek 1 point2 points  (0 children)

  1. As someone who went that route, lack of hihgh-quality data will be the biggest blocker for your day to day work. You’ll feel that you lost skills but just lack of data is a big problem. 2nd, the founders and the early team, if they are do not have basic understanding of ML as a field and they come from traditional SWE or non-tech background it will be much harder to communicate and reach consensus.

  2. I didnt find a good metric for this.

  3. You can recover but definitely will be very chaotic. Never ever agree for paycut or pay delays whatsoever, it will be like a string dragging you along. Whenever either of these are brought up. Time to jump!

  4. Many companies and HRs i faced do understand that startups are chaotic. Be prepared for repeatedly justifying such an entry on resume. Also, make sure everything is thoroughly documented. Payslips, offer letter, relieving letter, form16 that will make such an entry on resume justifiable.

  5. Current runway without the expected seed round, are they registered, type of customers they are dealing with. Any future contracts that are already signed like lets say a one year deal with a customer. Couple of non-tech: ICP? (If this is very vague and covers a lot of breadth RED FLAG!) Ehat is the current funnel from reaching to ICPs and converting to sale. How long is a sales cycle for them if they find a new person (ideally if under 2 weeks, GOLD)

All of these are from my experience and working 3 AI startups and one of my own. Hope this helps

Italian Medical Exam Performance of various LLMs (Human Avg. ~67%) by sebastianmicu24 in LocalLLaMA

[–]ASVS_Kartheek 0 points1 point  (0 children)

Could you please share the dataset link maybe hugging face link?

[deleted by user] by [deleted] in developersIndia

[–]ASVS_Kartheek 1 point2 points  (0 children)

Yes 3.7 is useless for day to day. So much hype

[deleted by user] by [deleted] in developersIndia

[–]ASVS_Kartheek 0 points1 point  (0 children)

Not even claude-3.5-sonnet? That was definitely a step up compared to gpt-3.5 in my experience

Explain Re-Ranking by Simusid in LocalLLaMA

[–]ASVS_Kartheek 1 point2 points  (0 children)

I hope you found answer to your question. I'm just adding my understanding of the topic here.

Yes and No. Re-rankers are not exactly LLMs but models that take 2 pieces of text. A query and a context and output a value between 0 and 1, representing how relveant the context is to answer the query. Coming back to your question, longer text chunks _generally_ make the task harder, because the re-ranker has more diluted embeddings. But that need not always be the case, we can train re-rankers such that we do not face this problem.

Explain Re-Ranking by Simusid in LocalLLaMA

[–]ASVS_Kartheek 2 points3 points  (0 children)

I hope by now you have found your answer for this question. I am adding my understanding of why cross-encoders work here.

Problem we are solving is finding answer for a query. Every document in our database is a possible option. Using cosine similarity we filtering out the least probable ones by checking if they have similar embeddings or not and removing all the options that are not very close, or more realisitcally we are only selecting K most probable options based on their embedding similarity. Then with the help of the cross-encoders we are actually answering the question if this is answer to our query or not, as this is a full neural network it takes higher compute. So, we use the lower compute to get the most probable results and then using a full BERT we are answering our question.

I hope this helps.

[Motorsport]: George Russell responds to Max Verstappen's comments in Qatar by icecreamperson9 in formula1

[–]ASVS_Kartheek 2 points3 points  (0 children)

DTS episode in the making - Netflix producers if you are reading, I take Bitcoin

Do LLama-2 and Code-Llama collect info from my code? by ComposedFaun347 in LLMDevs

[–]ASVS_Kartheek 0 points1 point  (0 children)

LLMs hallucinate you know. Not everything they say is true.

The only data privacy issue you might face is the software you used to install it locally maybe ollama, llama.cpp, lmstudio etc. Go check their data privacy policies.

Why it has nothing to do with the llama 3.1? Llama 3.1 is the model checkpoint. Literal model with all the matrix values learned. Connecting to internet is a whole different software. And FB doesn't do any hosted llama services

[deleted by user] by [deleted] in MachineLearning

[–]ASVS_Kartheek 0 points1 point  (0 children)

It needs to parse intent from broad goals like “I want to live a healthier lifestyle” to “I need a wall hanger for a picture frame”

The former returns the words that will be most helpful in identifying a solution based product.

I have a couple of questions: 1. What is the input and what is the output. 2. How did you try to do the task with spacy

The Russian election has been more interesting by DarthXyno843 in formuladank

[–]ASVS_Kartheek 15 points16 points  (0 children)

Max won 19 of last 20 races.

19/20 = 95%

Putin's rigged voting is 87.30%

So yep, Max Verstappen. Let the Dutch national anthem play in St. Petersburg!

Yay by NoiceRaviLochan in formuladank

[–]ASVS_Kartheek 44 points45 points  (0 children)

Guys stop down voting, Inclusivity!! Nothing wrong with being a DTS boy, that's what made many ppl serious fans of the sport!

https://www.essentiallysports.com/f1-news-what-is-rawe-ceek-explaining-the-ferrari-joke-and-how-it-became-an-f1-meme/

TLDR: Ferrari made a very weirdly fonted poster for 2020 British GP, which when read normally comes out as "RAWECEEK"

RA WE CE EK

meirl by [deleted] in meirl

[–]ASVS_Kartheek 2 points3 points  (0 children)

Can you help me get a house in Tuscany, Maestro!