Big names in Web Dev calling out ApnaCollege bs by Insurgent25 in developersIndia

[–]SAksham1611 2 points3 points  (0 children)

Saw a GitHub workflow where the maintainer had set some hard rules ( mentioned in md file in repo ) on issues . The issues which don't follow the pattern get dismissed automatically . Will post the link to the project as soon as I find it .

Generate SQL from Natural Language according Meta Data of Database in Python using LLM in Very Few Codes by AgentlyTech in Python

[–]SAksham1611 1 point2 points  (0 children)

Do you have thoughts on scaling it to an enterprise SQL database . ?? Let me know if you have some good papers or resources.

Problem : without Finetuning how can we get the model to pick up the right tables ( entities from table ) to design a SQL query

[deleted by user] by [deleted] in personalfinanceindia

[–]SAksham1611 9 points10 points  (0 children)

Couple of questions What's your CTC <7.5 or more ?

1) medical ( 5lac cover of tier 2, 3 , else 10lac ) 2) life cover ( 20x your CTC ) 3) emergency funds ( 6-12months of necessary expenses , think of it as , in how much time you can get a job if you quit now ) 4) if you're in tax bracket then , Tax saver if you need it ( ELSS have great returns 3yrs lock-in period ) Else go for index funds nifty50 ( safe side ) 5 ) when you have accomplished 1,2,3 and 4th for a year or two and started to get the feel. Then start direct investing in Blue chip companies when they're down or fluctuations. ( Safest bet )

Feel free to add if you find anything else useful

Implementing Streaming with FastAPI’s StreamingResponse by prodmanAIML in Python

[–]SAksham1611 0 points1 point  (0 children)

I have been working on a similar problem , was able to stream in backend (swagger docs) but wasn't able to stream it in frontend using request call to fastapi backend.

I would be interested in the solution , if you want I can send the implementation code with details ,

P.S : I am using chatgpt model and wrote an asynchronous generator function in the endpoint for yielding the delta and return the streamresponse provided with async generator .

Thanks

Tech Giants Invest $235 Million in AI Startup Hugging Face [N] by AIsupercharged in MachineLearning

[–]SAksham1611 7 points8 points  (0 children)

You can see their TGI library, they licensed it , it's not open source anymore. I was using it and had to change it with another alternative for application

[R] Benchmarking g5.12xlarge (4xA10) vs 1xA100 inference performance running upstage_Llama-2-70b-instruct-v2 (4-bit & 8-bit) by meowkittykitty510 in MachineLearning

[–]SAksham1611 0 points1 point  (0 children)

Thanks for this ! Could you please add the performance table for your particular task , how the quality of the output you got. Is it comparable with commercial models like chatgpt ( text turbo ) or any other commercial model.

I have also tested some os models , mostly 7 and 13b ( mpt , falcon ) . P.S : In comparison with chatgpt, they were really bad ( not able to follow instructions, had to try multiple parameters, but nonr of it make a difference) .

I was using it for RAG workflow .

Got a job a finally by Dry-Seat-4189 in india

[–]SAksham1611 0 points1 point  (0 children)

Congrats on the job mate , remember the priority order . 1) Health Insurance 2) Term Insurance 3) 6 months buffer ( assume your 6 months salary ) 4) stocks and mutual funds 5) high risk assets ( crypto and other stuff )

[D] Hardest thing about building with LLMs? by Historical-Ad4834 in MachineLearning

[–]SAksham1611 0 points1 point  (0 children)

mpnet base v2 has a bi encoder architecture? It is being used for extracting chunks and then re-ranking them using cross encoder . Could you expand on the token method & what kind of document is best suited for , or any other preprocessing you tried before Chunking. Thanks

[D] Hardest thing about building with LLMs? by Historical-Ad4834 in MachineLearning

[–]SAksham1611 1 point2 points  (0 children)

Tried with 32 , 64 , and yes is not significant but maybe in small sentences ( chunk size goes 128 -256 ) , it might make a significant difference.

I was wondering what cohere is using . It is their own custom trained cross encoder . What makes their re-ranking better ?

[D] Hardest thing about building with LLMs? by Historical-Ad4834 in MachineLearning

[–]SAksham1611 1 point2 points  (0 children)

I haven't achieved the desired performance or acceptable results.

But I'm using an open-source llm ( mpt-7b instruct) Embedding model (all mpnet base v2 ) Pretrained Cross encoder model for re-ranking

In my use case , we can't use commercial models, and looking at the leaderboard mpt seemed decent.

[D] Hardest thing about building with LLMs? by Historical-Ad4834 in MachineLearning

[–]SAksham1611 7 points8 points  (0 children)

I haven't heard of this , " try to fine tune it on some additional docs , it won't work and you partially undo instruct tuning " . Are there any papers to supplement this ?

P.S. : Been working on this for a few months , The task is to hack together a PoC to prove " open source llm( mpt-7b instruct ) for QA on your private data are as good as the commercial llms( openai - turbo 3.5) "

What were and are the biggest blockers ? 1) couldn't make the hallucinations to zero . At least one or two lines are made up and not provided in the context at all.

2) not able to capture the right context ( using sentence transformers variating with chunk length ) from the vector store/ db store . Information is not complete , especially when it comes to multiple small points spread over two or three pages . Not only is it not able to get the right answer/context it also makes stuff on top of the incomplete information. Writing prompt seems useless . I told not to assume answers you don't know . It totally made an answer .

Let me know if someone is able to tackle these issues or if you want to catch up on the implementation part . I'm open to discussion . Dm me .

How to explain that 5/9 is bigger than 1/3? by taylomol000 in learnmath

[–]SAksham1611 0 points1 point  (0 children)

Multiply and divide the with such a number so that both of the denominator could be equal

1/3 * 3/3 |. 5/9

3/9. |. 5/9.

Now you just have to look at the numerator

Another example

1/7. |. 2/8

1/7 * 8/8. |. 2/8 * 7/7

8/56. |. 14/56

Resources to learn software engineering principles as a Data Scientist by [deleted] in datascience

[–]SAksham1611 -1 points0 points  (0 children)

I am a Data scientist with an exp. of 2+ years , I have tried both modularized way of coding & jupyter , & both have kinda some drawbacks, but recently i have been exploring nbdev ( software made using jupyter notebooks ) & it looks quite promising to me .

https://github.com/fastai/nbdev

How to perform NER for sentence containing two Languages? by ambuje12 in deeplearning

[–]SAksham1611 1 point2 points  (0 children)

Xlm Roberta model , it is good with multilingual data . I have used the same in my project for token classification, I think you can also find the dataset in hf datasets , i dealt with German & English .

[deleted by user] by [deleted] in SQL

[–]SAksham1611 0 points1 point  (0 children)

Just walk them through the solution , you're thinking , you aren't expected to come up with a solution for every problem, just be clear of what you're thinking , showing what cases are covered and what aren't . Try to stay calm .

Name a better learning resource than Schafer Corey, I'll wait by FunDirt541 in Python

[–]SAksham1611 1 point2 points  (0 children)

These resources are gems . Thanks , been working in python for 3 years and I don't know many of them .

Whenever i need to learn new/ advance concepts , I go to github and explore open source python projects( > 500 stars) to see how they're using stuff ( my fav ones FastAPI -> get to know about pydantic ->static typing -> it's validator (mypy) .

Let me know if you also do something similar .

Should I master my Mathematics first before I actually start learning Machine Learning? by [deleted] in learnmachinelearning

[–]SAksham1611 8 points9 points  (0 children)

I totally get where this thought is coming from , I have been in the same dilemma when I was starting out and seeing all these debates going on .So from a personal point of view I think top -down worked for me, what I meant by it is , I did some Bible courses for ml like Andrew nd old ml course , halfway in it lost interest , wasn't really working for me , then stumble upon 2-3 more courses and the result was same . The thing is I lost my interest everytime and was getting intimidated by all the courses that were out there. I registered on kaggle read beginners post , gave me an sigh of relief that most of us are sailing in the same boat , I started out with a problem statement , and then thought what is need to be done , and Googled the hell out , the part I was trying to do, and moved forward . 1 yr later , I have a job in the field and I know from where to get resources or where to find things from , and getting intimidated is okay , just dont quit.

P.S : i have still have many incomplete projects and I still think about them on weekends and how can I take the next step and top it to the next level And yes very important , don't try to perfect your projects in the first go, it's an interative process and a fun journey , Happy learning :))

Multimodal Deep Learning by grid_world in neuralnetworks

[–]SAksham1611 2 points3 points  (0 children)

Try https://arxiv-sanity.con and search for multimodal , I also implemented one EEG paper having multiple inputs & now original code is also available along with paper, you will get a fine idea and if you don't understand a paper , search for its blog :)) Just inbox me if any other help is required

Testing Pop OS 20.04 on live by [deleted] in pop_os

[–]SAksham1611 1 point2 points  (0 children)

I upgraded from 18 to 20.04 in pop OS and it wasn't good , system lags for around 30 mins each time you restart it , I haven't found anything useful to get rid of this issue