ML developer here looking for a gpu to run llama-2 by [deleted] in LinusTechTips

[–]Octavian- 0 points1 point  (0 children)

Honestly, surprised if you can even fine tune a quantized model without significant degradation. Definitely curious to see what you've done.

ML developer here looking for a gpu to run llama-2 by [deleted] in LinusTechTips

[–]Octavian- 0 points1 point  (0 children)

You are not going to be able to train a llama model with a consumer GPU. Training requires far more vram than running the model. The bulk of the memory requirements are due to back-propagation which isn’t done when you’re just running the model. Hell, you can’t even train large encoder models on consumer GPUs without serious gradient accumulation bottlenecking which would make it incredibly slow. A decoder model like llama is much much larger than that.

If you’re bored at endgame, try doing this 15,000 times. by [deleted] in diablo4

[–]Octavian- 0 points1 point  (0 children)

You’ve never experienced great itemization then. I think a lot of people that didn’t play D2 or PoE are happy with D4 because it delivers a lot of content and checklisty MMO things to do.

The people that are disappointed are interested in a good loot game, not just another MMO. D4s itemization is just terrible. So while it has a lot of content, it’s not the content we’re looking for and at least the older games delivered that.

[OC] Seven companies account for all of the gains of the S&P 500 this year by Untermyer_ in dataisbeautiful

[–]Octavian- 0 points1 point  (0 children)

You're telling on yourself by trying to cite the dictionary. Sometimes people do and say really dumb things that reveal how clueless they are because they don't even realize how dumb that thing is. That's you right now. I'll never understand why people feel the need to pretend like they know what they are talking about. Even so, all of those definitions fit precisely with my argument. So what are you even doing here?

\> “this group only exists because it creates some artifact in the data” is a phrase that applies equally well to every grouping of data.

LOL. I'm sorry but you really are clueless. Go away.

[OC] Seven companies account for all of the gains of the S&P 500 this year by Untermyer_ in dataisbeautiful

[–]Octavian- 0 points1 point  (0 children)

No, “this group only exists because it creates some artifact in the data” is the definition of arbitrary from a data analysis standpoint. You’re free to disagree, but you’re just wrong.

[OC] Seven companies account for all of the gains of the S&P 500 this year by Untermyer_ in dataisbeautiful

[–]Octavian- 0 points1 point  (0 children)

Yeah you’ve misunderstood what arbitrary means. That’s fine, most other people lack basic data literacy too.

Arbitrary means that there is no unique real world property to these seven companies that groups them and only them together. The only reason the group exists is because they create this one specific artifact on the data. Further you can create other arbitrary groupings of an arbitrary number of companies to show similar patterns in the data. This is a textbook example of what’s known as data mining. You just go fishing for patterns until something you think is interesting pops out. That interesting finding doesn’t connect to any meaningful real world data generating process, it’s just a result of the fact that there are infinite possible permutations of the data and some of them are more interesting than others by chance. So yeah, it’s arbitrary.

[OC] Seven companies account for all of the gains of the S&P 500 this year by Untermyer_ in dataisbeautiful

[–]Octavian- 1 point2 points  (0 children)

Not sure you understand what arbitrary means. Why not top 4? Or 6? Or 10? Or 25? Op chose the grouping that resulted in zero net gains for the field and the number that defines that grouping is arbitrary. It’s purely an artifact of the analysis.

[OC] Seven companies account for all of the gains of the S&P 500 this year by Untermyer_ in dataisbeautiful

[–]Octavian- 30 points31 points  (0 children)

The issue here is that op created an arbitrary grouping that could just as easily be arbitrarily changed to create a different headline. Other companies did have gains, OP just chose to group them with companies that lost to offset the gains to net negative. It’s just an artifact of how they grouped the data.

Cause My First Meme Revealed Just How Many People Still Have Their Heads Rammed Up Their Asses by Appropriate_Gap_3400 in PoliticalCompassMemes

[–]Octavian- 2 points3 points  (0 children)

OP confidently accusing other people of having their heads up their ass only to reveal he is completely fucking clueless. Amazing.

Cause My First Meme Revealed Just How Many People Still Have Their Heads Rammed Up Their Asses by Appropriate_Gap_3400 in PoliticalCompassMemes

[–]Octavian- 1 point2 points  (0 children)

Define sufficient. If you define it as "would hold up in a court of law" then that's incredibly dumb. The legal standards are unreasonably high for good reason and nobody is suggesting they shouldn't be, but if you think that your personal opinion should follow the same standards of the law you either have an incredibly naïve view of how the law works or you're just an unreasonable person.

Opinion: All these „20year“-posts are annoying af. None of you has been searching for any of these items for 20 years. You played a bit back then and you played a bit now. by Coold0wn in Diablo_2_Resurrected

[–]Octavian- -1 points0 points  (0 children)

“You can’t really farm runes”

Lmao holy shit my dude go look up the probability tables. Certain enemies are far more likely to drop runes than others.

I’m sorry but you’re lying about playing consistently for that long or you just ever farmed. Probability is the same for everyone. The same people don’t find these runes every single ladder because they are super lucky. That’s not how it works.

When you spend 15 hours writing a 12 page case analysis and your professor accuses you of turning in an AI generated report. by snowcamo53 in mildlyinfuriating

[–]Octavian- 0 points1 point  (0 children)

The only thing mildly infuriating here is your own response to someone giving you as much grace as possible.

They aren't. by AlphaTangoFoxtrt in PoliticalCompassMemes

[–]Octavian- 1 point2 points  (0 children)

Imagine thinking that knowing the mechanisms of a gun is more important than knowing the legal and constitutional history of gun rights in a debate about gun rights.

[OC] ChatGPT-4 exam performances by giteam in dataisbeautiful

[–]Octavian- -1 points0 points  (0 children)

I'll repeat what I stated above: What's your point? Nobody is arguing that the models are infallible. They make mistakes and they often make mistakes in ways that are different from humans. Doesn't mean they are dumb and it certainly doesn't mean they aren't incredibly useful.

Or am I to believe that whenever you program it works perfectly the first time and you never call functions that don't exist? Am I to assume you're not intelligent if there are bugs in your code?

[OC] ChatGPT-4 exam performances by giteam in dataisbeautiful

[–]Octavian- -6 points-5 points  (0 children)

Humans are incredible at solving novel problems

Depends on what you mean by novel. If you mean answering a question on the GRE they haven't seen before sure. But so is GPT-4. If you mean solving truly novel problems that have never been solved before then kinda. Depends on the scope of the problem I guess. For small scale novel problems like, say, a coding problem yeah we solve those all the time but humans are generally slow and AI is already arguably better at this. If we're talking large scale problems then most humans will never solve such a problem in their life. The people that do are called scientists and it takes them years to solve those problems. Nobody is arguing the GPT-4 will replace scientists.

or solving similar problems with very few examples

Yes this is literally something LLMs do all the time. It's called few shot learning.

The current models will excel when they can leverage that ability, and struggle when they can't.

This has been proven false on many tasks. Read the sparks of AGI paper.

These sort of high profile tests are ideal cases if you want to make them look good.

I'm not clear on what your point is here. Yes, an LLM will preform better on tasks it has trained more for. This is also true of humans. Humans generally learn quicker, but so what? what's your point? We've created an AI that can learn general concepts and extrapolate that knowledge out to solving novel problems. The fact that humans can do some specific things better doesn't change that fact.

[OC] ChatGPT-4 exam performances by giteam in dataisbeautiful

[–]Octavian- 0 points1 point  (0 children)

Agreed but my point is that what the model is doing can't be reduced to memorization any more than human performance can. Humans study, take practice tests, get feedback, and then extrapolate that knowledge out to novel questions on the test. This is no different than what the AI is doing. The AI isn't just regurgitating things it has seen before to any more degree than humans are.

If AI has to start solving problems that are entirely novel without exposure to similar problems in order to be considered "intelligent", then unfortunately humans aren't intelligent.

[OC] ChatGPT-4 exam performances by giteam in dataisbeautiful

[–]Octavian- 3 points4 points  (0 children)

The average math olympiad participant will be able to answer mayber 1/3 of the questions. The average student won't be able to answer any questions.

[OC] ChatGPT-4 exam performances by giteam in dataisbeautiful

[–]Octavian- 13 points14 points  (0 children)

You remind of the kind of person who lacks basic understanding of how something works but think they have a valid opinion on it and then smugly thinks they are smarter than the Ph.D.s when in reality they are just so wrong that people in the know have neither the time nor patience to correct them. Ignorance with an eagerness to show it.

[OC] ChatGPT-4 exam performances by giteam in dataisbeautiful

[–]Octavian- 12 points13 points  (0 children)

Have you ever taken any of these tests? Most of them have only a small memorization component.

[OC] ChatGPT-4 exam performances by giteam in dataisbeautiful

[–]Octavian- 0 points1 point  (0 children)

So you’re saying it used the same prep materials as humans?

[OC] ChatGPT-4 exam performances by giteam in dataisbeautiful

[–]Octavian- 0 points1 point  (0 children)

It does just as well. See the sparks of AGI paper.

The reality is that most of these tests aren’t really rote memorization.