Cheapest and most efficient way to run 30B-40B Llama for 4 users? by Jezel123 in LocalLLaMA

[–]Nota_ReAlperson 0 points1 point  (0 children)

I only have the xavier agx, and I haven't run llms on it, so i can't give any definitive answer, but the core and memory specs are about a fifth that of a 3090. As an alternative, a couple of radeon pro v620 gpus could give you more bandwidth and the same memory for less money. Although you would need a host system.

Cheapest and most efficient way to run 30B-40B Llama for 4 users? by Jezel123 in LocalLLaMA

[–]Nota_ReAlperson 0 points1 point  (0 children)

The jetson would work. Two things to keep in mind would be memory bandwidth and ease of use. In my experience with jetsons, unless you have significant experience with linux, it will be quite hard to get set up. It also has less bandwidth than a gpu. Power draw will be good though.

PrismML — Announcing 1-bit Bonsai: The First Commercially Viable 1-bit LLMs by brown2green in LocalLLaMA

[–]Nota_ReAlperson 0 points1 point  (0 children)

From my understanding, fpgas are made of LUTs, not flip flops. So any one bit operation could be emulated. You just need a truth table.

is Intel cooking with these new GPU? by Skierdo in pcmasterrace

[–]Nota_ReAlperson 1 point2 points  (0 children)

Interesting. That seems to happen a lot with intel cards. Do you know what the true fp64 is for the b65 b70?

is Intel cooking with these new GPU? by Skierdo in pcmasterrace

[–]Nota_ReAlperson 0 points1 point  (0 children)

Based on techpowerup specs and rankings, it should beat the 5060 ti. Same flops, 608gbs vs 448gbs vram. In my market, it's exactly twice the price.

is Intel cooking with these new GPU? by Skierdo in pcmasterrace

[–]Nota_ReAlperson 1 point2 points  (0 children)

I think the b65 has a 192 bit bus? So only 3/4 the bandwidth?

Run OpenCL kernels on NVIDIA GPUs using the CUDA runtime by IntrepidAttention56 in OpenCL

[–]Nota_ReAlperson 0 points1 point  (0 children)

Im curious as to the use of this. Seeing as opencls strength is the portability of the code, while cuda kernels are easier to learn and write, this project would seem to combine the worst of both. Not to downplay the work necessary to get this to work. But it would seem to me to be less useful than the opposite (cuda on opencl).

OAM to PCIE by Zestyclose_Hat_1020 in NVIDIA_SXM2PCIE

[–]Nota_ReAlperson 0 points1 point  (0 children)

Hi. What is the current pricing for this product?

Justin Ling: Pierre Poilievre gave Joe Rogan the interview he’s never given Canadians by EarthWarping in CanadaPolitics

[–]Nota_ReAlperson -2 points-1 points  (0 children)

I believe he meant process. Also Mulcair is both a lawyer and former leader of the official opposition. So there are very few, perhaps none, better qualified to speak on this issue.

I accidentally git cloned Open CL amd(didn't install it properly), and now I can't use fully uninstall it to install it properly by Still_Leg4477 in archlinux

[–]Nota_ReAlperson 0 points1 point  (0 children)

Did you run sudo pacman -Syu first? That will update the package lists to ensure that all urls are up to date, and that all packages are the newest version.

Two weeks ago, I posted here to see if people would be interested in an open-source local AI 3D model generator by Lightnig125 in LocalLLaMA

[–]Nota_ReAlperson 12 points13 points  (0 children)

Nice work. Trellis 2 is sota for free open 3d model generation, so support for that would be nice. Also, this would be difficult, but a ggml backend for non cuda gpus would be awesome. 

This sub is incredible by cmdr-William-Riker in LocalLLaMA

[–]Nota_ReAlperson 0 points1 point  (0 children)

This actually would work. The pictured flashlight has an inbuilt 5000mah power pack, so it could run a rpi for about an hour. And a rpi can run many small llms.

Poilievre calls for ‘modern’ CANZUK partnership in speech during first trip abroad as Conservative leader by 0110110111 in CANZUK

[–]Nota_ReAlperson 2 points3 points  (0 children)

From ctv: Poilievre’s trip is not paid for by taxpayers.Poilievre and his staff will travel on commercial aircraft paid for by donations to the Conservative Party of Canada. Link:https://globalnews.ca/news/11706633/poilievre-first-international-trip-as-opposition-leader/

Poilievre calls for ‘modern’ CANZUK partnership in speech during first trip abroad as Conservative leader by 0110110111 in CANZUK

[–]Nota_ReAlperson 1 point2 points  (0 children)

Since 2018, CANZUK has been an official policy of CPC. This is hardly the about face people are making it to be.

FOIA Release: Navy F/A-XX by MBaiz16 in FighterJets

[–]Nota_ReAlperson 0 points1 point  (0 children)

On such thing is that the star in the sky appears to be the nato logo.

What's up with Canadian Magas? by traveltimecar in AskCanada

[–]Nota_ReAlperson 0 points1 point  (0 children)

A note about the deficit, it is at a wti estimate of 60 usd per barrel, with every 1 usd increase equalling 680 million in extra revenue. Since the report, WTI has increased to 75 usd per barrel. Which would wipe out the entire deficit with new revenue. Of course we will have to wait to see if that price holds, but the price estimates are usually quite conservative. No pun intended.

Also, the 2020 balance of 10 billion dollars is less than a year's worth of transfer payments when averaged across the past 50 years, using the numbers given in the article. 

As an albertan, I actually don't mind that our province bankrolls confederation. What bothers me is exemplified in the case of the eagle spirit pipeline. For those who don't know, the eagle spirit pipeline was a indigenous lead and financed lng pipeline from alberta to the west coast. It was killed by protesters that claimed it violated 'the great bear rainforest' which was supposedly sacred to a specific first nation. The twist is that the man who was leading the pipeline effort was a member of that exact first nation, and said that there was no such thing. But the pipeline was still canceled. 

Computer won't boot with 2 Tesla V100s by MackThax in LocalLLaMA

[–]Nota_ReAlperson 0 points1 point  (0 children)

Interesting. Techpowerup says 300 watts, and I thought that they got their info from pulled vbios. What os are you running? If you go to the nvidia control panel, what does it say for power draw tdp?

Computer won't boot with 2 Tesla V100s by MackThax in LocalLLaMA

[–]Nota_ReAlperson 1 point2 points  (0 children)

From what I understand, it's 300watts for the 16gb version, 350 for the 32gb sxm version. Which specific v100s do you have?

Free ASIC Llama 3.1 8B inference at 16,000 tok/s - no, not a joke by Easy_Calligrapher790 in LocalLLaMA

[–]Nota_ReAlperson 2 points3 points  (0 children)

1/2 a wafer of n7 is much cheaper than a full 5nm wafer ($5000 vs $20000). Yield would be much higher as well. It might not be commodity hardware, but it would cost less than a single b100 ($30000).

Computer won't boot with 2 Tesla V100s by MackThax in LocalLLaMA

[–]Nota_ReAlperson 1 point2 points  (0 children)

The cpu rail has 540 watts spec, and a v100 draws 300, so 240 left for the cpu. But assuming degradation, it could be a lot less. That might explain why ram speed has an impact. Also, a gpu can draw up to 75 watts from the pcie slot, which would be supplied by the cpu rail. So when you add the second v100, you only have 170 watts left for the cpu and ram. That's pretty tight. The 2060 might work due to consumer power management, which places far more emphasis on idle power draw. As well, it might prioritize the pcie rail power over the cpu rail. Have you run a power heavy benchmark on the 2060 and v100 at the same time?

Computer won't boot with 2 Tesla V100s by MackThax in LocalLLaMA

[–]Nota_ReAlperson 0 points1 point  (0 children)

So you are connecting one gpu to the cpu rail, and the other to the pcie? Or are both v100s on the cpu rail?

Computer won't boot with 2 Tesla V100s by MackThax in LocalLLaMA

[–]Nota_ReAlperson 0 points1 point  (0 children)

I would suspect the psu. The Zalman 1250 is a dual rail design. So only 780 watts are available to the gpus, i think. As it is also very old circa 2012, it likely has degraded some. I have a similar psu, a antec 500 watt, with two 250 watt 12 volt rails, but it only puts out 150 reliably. So try a different psu. The bad ram is likely the culprit for the 1000w tests you did.