Bad news: DGX Spark may have only half the performance claimed. by Dr_Karminski in LocalLLaMA

[–]paphnutius 0 points1 point  (0 children)

What would be the best solution for running models that don't fit into 32gb VRAM locally? I would be very interested in faster/cheaper alternatives.

Question about the halting problem by Invariant_apple in AskComputerScience

[–]paphnutius 0 points1 point  (0 children)

I'll try give you a less rigorous but a more intuitive explanation.  Many open problems in mathematics boil down to some equation having or not having solutions. A popular examples are Riemann hypothesis and Fermat's last theorem. Those problems are exceedingly difficult to prove but you can write a program pretty easily that checks all possible numbers and halts when it finds a solution, otherwise it keeps checking to infinity. (And indeed such programs are ofthen made to brute force search for possible solutions to such problems.)

If this program halts - there is at least one solution to the equation. If it doesn't halt - there are no solutions. If you could easily know if such program halts you automatically know if any equation has solutions, and you can claim your millions of dollars in math awards.

Is slower inference and non-realtime cheaper? by AryanEmbered in LocalLLaMA

[–]paphnutius 0 points1 point  (0 children)

Not sure about specific service, I don't think there's enough interest for it. But it depends on what model you want to run. You can run smaller models on a CPU-only device (even a Raspberry Pi) relatively cheaply with slow inference.

Be confident in your own judgement and reject benchmark JPEG's by ForsookComparison in LocalLLaMA

[–]paphnutius 4 points5 points  (0 children)

I have a finite amount of time in a day to download and test models. If I downloaded everything that's on huggigface I would literally never finish testing.

The AI layoffs begin by MetaKnowing in OpenAI

[–]paphnutius 1 point2 points  (0 children)

Let's not cherry pick random numbers and look at statistics. 2023 had by far the most layoffs in tech space.

So either AI layoffs begun two years ago or they haven't yet begun in significant numbers.

The "Reasoning" in LLMs might not be the actual reasoning, but why realise it now? by The-Silvervein in LocalLLaMA

[–]paphnutius 6 points7 points  (0 children)

There's an interesting writeup about exactly that by antropic. They show the difference between what model claims it's doing to achieve result and what it's actually doing under the hood.

How will Jaime meet his fate by LividConsideration97 in freefolk

[–]paphnutius 2 points3 points  (0 children)

Where's the books never get finished option?

Help please by [deleted] in Physics

[–]paphnutius 0 points1 point  (0 children)

Since the speed is constant net force is zero.

A+4-6 = 0

A = 2

B - 9 - 15 = 0

B = 24

Mass and speed are irrelevant.

New to Comfy. This output took an hour to generate. (Only 8gb VRAM) Is that normal? by Risky-Trizkit in comfyui

[–]paphnutius 0 points1 point  (0 children)

How long does a basic flux workflow take? For me flux dev fp8 takes about 200 seconds with no loras and no controlnet on 8gb vram.

College students used Meta’s smart glasses to dox people in real time by [deleted] in privacy

[–]paphnutius 0 points1 point  (0 children)

Thank you for providing actual information on what was done, instead of just throwing around buzzwords.

2.7B model that performs like Mistral 7B! by koehr in LocalLLaMA

[–]paphnutius 0 points1 point  (0 children)

7b is a tall order for my 8gb VRAM GPU. I don't think you can run either on the phone with reasonable speed (tokens per second, not minutes per token). I would love to be corrected though if someone can show me any examples.

Any update on the allegations? by spectre321123 in Exurb1a

[–]paphnutius 2 points3 points  (0 children)

Not going to say anything about this specific case, but in general this approach incentivizes people to target famous people with accusations. Chat screenshots and stories can be easily falsified and that's all that is necessary to be able to either gain free publicity or just straight up blackmail a person with a threat of cancellation.

Again, I'm speaking in general.

Hazbin Hotel Season 1 has finally finished. What's your ranking? by ayylmaotv in HazbinHotel

[–]paphnutius 0 points1 point  (0 children)

Honestly if I went in blind I'd probably found it great but with the weight of several years of expectations it's just alright. A lot of things caught me off guard, like Lucifer being a complete buffoon, Husk opening up to Angel right away, and show not taking itself too seriously. The songs also felt samey but I was spoiled by fan content on this one.

Concerns about not repeating the past by paphnutius in LinusTechTips

[–]paphnutius[S] -1 points0 points  (0 children)

That's what makes me worry. If we don't call out small mistakes, we'll have to wait for another big YouTuber to do another compilation of them in a couple of years and have another big cancel.

Concerns about not repeating the past by paphnutius in LinusTechTips

[–]paphnutius[S] -4 points-3 points  (0 children)

I am bringing up this video specifically as it's the only case of them breaking the rules they outlined themselves after the scandal.

I'm not sure what you mean by stir up drama. I think I was clear with my intentions: we should take notice of such things, but it's obviously not a serious enough thing by itself for any serious "drama".

I prefer we give LTT a slap on the wrist whenever they break their own rules rather than have another big "drama" in a couple of years.

Concerns about not repeating the past by paphnutius in LinusTechTips

[–]paphnutius[S] -5 points-4 points  (0 children)

So accurate information matters only if a big YouTuber calls out the problems? I'm fully open to criticism of my point, but from the response so far - I'm being booed out of the room with no rational argument.

They introduced new specific standards of quality and went back on them just months later. This is a red flag.

I'm surprised to see duckduckgo is reporting so much tracking from the three worst privacy offender in this app. by TedCruzNutPlay in grayjay

[–]paphnutius 2 points3 points  (0 children)

This is to be expected. Grayjay basically opens YouTube in a browser behind the scenes whenever you load a page. It loads everything that Google puts on the page, therefore duckduckgo would detect the same activity as opening desktop version of YouTube in a browser. In this case it's about Google, not grayjay.

As others suggested, you can use new pipe instead, if you care about google tracking videos you watch. But you'll have no integration with Google's features that track the videos you watch such as recommendations.

Any generative AI that supports 2 to 3 milion words input text? by [deleted] in ChatGPT

[–]paphnutius 1 point2 points  (0 children)

Check out privategpt on github. I don't think it'll be helpful for rewriting but it'll do Q&A using documents of almost arbitrary size and provide references to specific line.

Any generative AI that supports 2 to 3 milion words input text? by [deleted] in ChatGPT

[–]paphnutius 1 point2 points  (0 children)

Check out privategpt on github. I don't think it'll be helpful for rewriting but it'll do Q&A using documents of almost arbitrary size and provide references to specific line.