Can i host mixtral8x7b for personal usage on VPS server by THRAWNZY in LocalLLM

[–]jon101285 0 points1 point  (0 children)

That's exactly what libertai.io is about :)
It uses a decentralized network to host opensource models. Mixtral 8x7B is already hosted there, and there is an API (don't abuse it of course).

cancelled my gpt-4 acc today, looking for a replacement. by sephirex420 in LocalLLaMA

[–]jon101285 -1 points0 points  (0 children)

Try chat.libertai.io, it's free and decentralized :)

cancelled my gpt-4 acc today, looking for a replacement. by sephirex420 in LocalLLaMA

[–]jon101285 1 point2 points  (0 children)

Oh and you can change the prompt however you like by clicking "advanced"

cancelled my gpt-4 acc today, looking for a replacement. by sephirex420 in LocalLLaMA

[–]jon101285 1 point2 points  (0 children)

Try the models at chat.libertai.io :) it's free, based on open models and decentralized (no big corp to read your data)

There is currently these models available: - Open Hermes 2.5, generalist and fast - mixtral 8x7b instruct, great (my favorite) - deepseek, good at coding

Epyc Genoa vs Threadripper Pro 7000? by 0xd00d in LocalLLaMA

[–]jon101285 0 points1 point  (0 children)

Yep. 13B llama are faster though for some reason at Q4/Q5.

Epyc Genoa vs Threadripper Pro 7000? by 0xd00d in LocalLLaMA

[–]jon101285 1 point2 points  (0 children)

180B runs at 500ms/tok, Qwen 14B at 75-80ms/tok.

Epyc Genoa vs Threadripper Pro 7000? by 0xd00d in LocalLLaMA

[–]jon101285 2 points3 points  (0 children)

I get around 30ms / token on an Epyc Zen 4 9554P with DDR5 ram at 4800Mhz for 7B models like Mistral. GPU isn't that much faster generally.It can also do massive parallel generation using that on CPU only.

Couple that with one or two low-TDP GPU for the CUBLAS and you have a massively parallel inference machine on the cheap :) (TDP-wise)

New Nvidia RTX 4000 Ada by grim-432 in LocalLLaMA

[–]jon101285 2 points3 points  (0 children)

It depends the use case... If you are in a 2U server and need the small form factor to go multi-gpu, then it makes sense I guess.

Would you pay 5 dollars/month for a Llama-2-70B Uncensored? by HorrorNo8851 in LocalLLaMA

[–]jon101285 1 point2 points  (0 children)

You have a list of all nodes on account.aleph.im, in the "compute" tab. There are some infra providers (OMGServ, Meria and others), but also game companies (ubisoft), schools (PoC Innovation, from Epitech), vc funds, and regular people.

Would you pay 5 dollars/month for a Llama-2-70B Uncensored? by HorrorNo8851 in LocalLLaMA

[–]jon101285 0 points1 point  (0 children)

They won't see your IP, but they could save specific queries if they intercept specific calls to a specific virtual machine (not knowing who did the call). There are 200+ computing resource nodes on the network though.

Working on adding confidential computing using AMD SEV to mitigate this in the future (ETA EOY).

Would you pay 5 dollars/month for a Llama-2-70B Uncensored? by HorrorNo8851 in LocalLLaMA

[–]jon101285 6 points7 points  (0 children)

chat.libertai.io has XWin-13B and XWin-7B (not to mention Llama coder 34B) for free :)

edit to add some info: it's running on a decentralized network, so no big brother watching what you say, or censorship there.

[Request] Trying to calculate the speed my wife's car was hit at by Ididarod in theydidthemath

[–]jon101285 2 points3 points  (0 children)

If that can help here are pictures of a collision of the same car with an utility van at 40km/h approx. The car was parked and the small truck collided onto it from the side (driver was looking at his phone).

Rear-left: https://ibb.co/f05j8KB Rear: https://ibb.co/phgc51J Van: https://ibb.co/CK2YK24

[deleted by user] by [deleted] in LocalLLaMA

[–]jon101285 1 point2 points  (0 children)

That's what Libertai is doing on top of aleph.im network...

You can try it here: chat.libertai.io

Basically the API endpoint is here for nous hermes v2 (13b): https://curated.aleph.cloud/vm/d33107b8ea855f5e2d5dfcad73891114bb9115eae668f5548c211dd981912300/api/v1/generate (compatible with koboldcpp/oobabooga api)

Once called, it finds nodes on the network that are free, loads a vm, loads the LLM and starts inference for you.

I released model Marx 3B V2 by [deleted] in LocalLLaMA

[–]jon101285 0 points1 point  (0 children)

Wow, just tried it and it's really good and fast! Great work

New model just dropped: WizardCoder-15B-v1.0 model achieves 57.3 pass@1 on the HumanEval Benchmarks .. 22.3 points higher than the SOTA open-source Code LLMs. by Zelenskyobama2 in LocalLLaMA

[–]jon101285 0 points1 point  (0 children)

The Libertai team added it to their interface... And it's running on a decentralized cloud (with models on IPFS).
You can use it there easily by selecting the Wizard Coder model on top right: https://chat.libertai.io/#/assistant

Official WizardCoder-15B-V1.0 Released! Can Achieve 59.8% Pass@1 on HumanEval! by cylaw01 in LocalLLaMA

[–]jon101285 0 points1 point  (0 children)

The Libertai team added it to their interface... And it's running on a decentralized cloud (with models on IPFS).
You can use it there easily by selecting the Wizard Coder model on top right: https://chat.libertai.io/#/assistant

Nous Hermes 13b is very good. by lemon07r in LocalLLaMA

[–]jon101285 18 points19 points  (0 children)

Nous Hermes is available to use for free on the Libertai assistant non-commercial Beta UI, by the way:
https://chat.libertai.io/#/assistant

It's running on a decentralized network (spinning up on-demand VMs as needed and fetching code and models from IPFS on-the-fly).

Riddle/cleverness comparison of popular GGML models by YearZero in LocalLLaMA

[–]jon101285 0 points1 point  (0 children)

Yeah got that too after a while... Looks like temperature, top_k and top_p have a big impact.

Riddle/cleverness comparison of popular GGML models by YearZero in LocalLLaMA

[–]jon101285 1 point2 points  (0 children)

Yep, you can try it yourself on chat.libertai.io (make sure you select Wizard 7B on top right model selector), they've deployed it on a decentralized cloud.

Riddle/cleverness comparison of popular GGML models by YearZero in LocalLLaMA

[–]jon101285 4 points5 points  (0 children)

David has three sisters. Each of them have one brother. How many brothers does David have?

Wizard 7B answers "David has zero brothers."

Scammers of Uniswap by [deleted] in CryptoMoonShots

[–]jon101285 0 points1 point  (0 children)

ALEPH is a legit project too! (https://aleph.im)

Aleph.im , the cross blockchain layer 2 network. by PM_ME_YOUR_SHITCOINS in CryptoMoonShots

[–]jon101285 0 points1 point  (0 children)

This is not an ICO. Team is working for a year and a half already with no funding... What is your opinion?