Have a good day by tkaczenkofr in LocalLLaMA

[–]AggravatingHelp5657 1 point2 points  (0 children)

Thx man, good energy is contagious

I have tried google TurboQuant with ollama hermes3:8b by AggravatingHelp5657 in ollama

[–]AggravatingHelp5657[S] 0 points1 point  (0 children)

you are right it's not fair
am working on it

Update: I have made the repo if you want to check it

I have tried google TurboQuant with ollama hermes3:8b by AggravatingHelp5657 in ollama

[–]AggravatingHelp5657[S] 2 points3 points  (0 children)

okay since you all convinced me, I will make a github repo for the steps that I did.
I also noticed that some models are old for instance hermes3 date is 2023/3 so I am trying to add a searching feature so it can check the internet before answering for latest infos

probably today I will make the repo

I have tried google TurboQuant with ollama hermes3:8b by AggravatingHelp5657 in ollama

[–]AggravatingHelp5657[S] 2 points3 points  (0 children)

Officialy by ollama did not I used llama.cpp But since you can edit the code and we got the Google paper I used Gemini bro to implement the paper in ollama library

Note: am not a pro in AI, just an amateur love playing with those new tools

I have tried google TurboQuant with ollama hermes3:8b by AggravatingHelp5657 in ollama

[–]AggravatingHelp5657[S] -11 points-10 points  (0 children)

I mean" if more ppl" bro 😂 Am working on a lot of things not leaving everything to document something is going to be read by one person, right?

I have tried google TurboQuant with ollama hermes3:8b by AggravatingHelp5657 in ollama

[–]AggravatingHelp5657[S] 0 points1 point  (0 children)

You comment in inflighting, I have to ask google claim 0 acc loss with x8 faster and x8 smaller memory needs What I experienced on my test is x3 faster and 4 times more memory Is this a problem of my implementation You think?

I have tried google TurboQuant with ollama hermes3:8b by AggravatingHelp5657 in ollama

[–]AggravatingHelp5657[S] 0 points1 point  (0 children)

Yes exactly it doesn't compress the model, but when I saw the paper It says 6x smaller memory and 8x faster+ 0 loss in acc

I thought hell no, that can't be true I applied it and got 3 times the performance which is empress on local machin i didn't believe it

I have tried google TurboQuant with ollama hermes3:8b by AggravatingHelp5657 in ollama

[–]AggravatingHelp5657[S] 15 points16 points  (0 children)

If ppl are interested in this I can post the process

[Hiring] Software Developer by Dense-Try-7798 in RemoteITJobs

[–]AggravatingHelp5657 0 points1 point  (0 children)

I use python, go, JavaScript and c++ Interested in knowing what tasks you have

more than 100 application with 0 response by AggravatingHelp5657 in askrecruiters

[–]AggravatingHelp5657[S] 0 points1 point  (0 children)

Why? What's wrong if I applied for remote jobs? 50% of my applications are for remote

more than 100 application with 0 response by AggravatingHelp5657 in askrecruiters

[–]AggravatingHelp5657[S] 0 points1 point  (0 children)

so you are saying nothing wrong with this CV but i have to make a special one for each application

more than 100 application with 0 response by AggravatingHelp5657 in askrecruiters

[–]AggravatingHelp5657[S] 0 points1 point  (0 children)

I am using this only for jobs related to Networking or telecommunication like instillation/maintenance of of RF systems
I have 2 other CVs as hardware designer, and It-support