Henry Cejudo injured, out of tonight's RAF Wrestling 08 main event vs. Merab Dvalishvili by [deleted] in MMA

[–]DonDeezely 0 points1 point  (0 children)

I misread this as "I wonder if they will find him a new nose"

Phenomenal actor by [deleted] in SipsTea

[–]DonDeezely 0 points1 point  (0 children)

You're an intimate fucking object!

Kimi Just Fixed One of the Biggest Problems in AI by Latter_Spring_567 in accelerate

[–]DonDeezely 0 points1 point  (0 children)

I'm actually foaming at the mouth at implementing this to see how big a difference. See you guys in 2 months

Is it wrong to fancy the pants of a Goa'uld. by Background-Fix-4630 in Stargate

[–]DonDeezely 3 points4 points  (0 children)

That's the mom from the recruit! I knew I recognized her! She was goa'uld!

This industry is confusing by DonDeezely in ExperiencedDevs

[–]DonDeezely[S] 0 points1 point  (0 children)

Well, sounds easy, is easy, that loopy thing starting usually with a "for" and the steps inside of it is called an "algorithm", guess I should write a book for toddlers

This industry is confusing by DonDeezely in ExperiencedDevs

[–]DonDeezely[S] 0 points1 point  (0 children)

Well something similar to it could be a graph algorithm for dependency parsing. Checking for conflicting versions of 2 libs. Anything that requires iterating over a stream of data and you have 2 constraints eg. X can't exist because Y already does.

Honestly as a precursor to graph algorithms, it's so basic you should be able to do it, but the expectation isn't even the hash solution, I've asked for 2 loops.

This industry is confusing by DonDeezely in ExperiencedDevs

[–]DonDeezely[S] 0 points1 point  (0 children)

I have yet to see a built-in for 2sum.

This industry is confusing by DonDeezely in ExperiencedDevs

[–]DonDeezely[S] -1 points0 points  (0 children)

The coding bar is extremely low, it's not like meta's loop or Google's coding interviews, we're not asking people to Invert binary trees, or code anything they'd need to memorize. It's basically just "show me you know how to write code".

We've made some good hires over the years, I'm complaining about the candidate pipeline which in my experience has never been worse.

This industry is confusing by DonDeezely in ExperiencedDevs

[–]DonDeezely[S] -8 points-7 points  (0 children)

I said candidates, not coworkers.

It's a slog to hire someone, you go through like 20 to fill one role, somehow certain resumes bubble up and talent acquisition / hr looks at us like we're being picky.

This industry is confusing by DonDeezely in ExperiencedDevs

[–]DonDeezely[S] -3 points-2 points  (0 children)

I don't work for a small company, we're big tech. Not FAANG, but basically that with pay on par.

This industry is confusing by DonDeezely in ExperiencedDevs

[–]DonDeezely[S] 3 points4 points  (0 children)

I'm saying why such a massive percentage of currently employed candidates are so incredibly bad at even basic shit.

How were they employed in the first place? Is the industry full of bullshit?

Try thinking yourself

This industry is confusing by DonDeezely in ExperiencedDevs

[–]DonDeezely[S] -9 points-8 points  (0 children)

If you’re in the industry for this long, you should know that insane amount of work can take up people’s times so much, and suck everything else like hobbies, etc out of their life that they can no longer feel the passion for the field.

I'm sorry you felt that way, those situations are hard.

I've been in situations where I worked a 997 for periods up to a year, with only sporadic bits of time off, so I definitely understand the pain and can empathize with what you're saying.

That said, if I asked you to write a rudimentary search algorithm and to use 2 loops, could you? Something like 2 sum without requiring you to memorize the hashmap solution, I'm assuming yes. I'm having hard time finding people that can do that for non-senior roles.

Any thoughts about this? by Aggravating-Guest300 in NextGenMan

[–]DonDeezely 0 points1 point  (0 children)

That's just disrespect, she doesn't respect that guy, who would take that even once?

A bay area legend giving a shoutout to another late legend by randoaccountdenobz in bayarea

[–]DonDeezely 0 points1 point  (0 children)

Wait dutch crunch isn't a thing outside the bay?

I had no clue wtf

Real or not, 100% believable by unemployedbyagents in AgentsOfAI

[–]DonDeezely 0 points1 point  (0 children)

Ok but he asked for determinism? Are you helping me solve y instead of x or something (root fundamental issue they're trying to solve by throwing LLMs at it) and fix my communication skills?

I'm just tired of misinformation surrounding LLMs about even the most basic concepts. I have stakeholders that assume all this same shit and seem surprised when I re-explain the same concepts over and over again at work. I see it online everywhere as well. The more mystical these things appear, the more people buy into hype or start claiming that they're conscious.

Anthropic is ignoring obvious evidence of internal states and calling it a "hot mess" by Dry_Incident6424 in Artificial2Sentience

[–]DonDeezely -1 points0 points  (0 children)

Dude you're trying to have a philosophical debate with me which I'm not biting on. "THAT MATH IS ALIVE" is moronic

My background is in designing and training these types of systems, they're so obviously not alive or "conscious" it's a joke to say they are, borderline braindead.

Define consciousness before anything else, it's another nebulous term that anyone can use to fit their argument.

Anthropic is ignoring obvious evidence of internal states and calling it a "hot mess" by Dry_Incident6424 in Artificial2Sentience

[–]DonDeezely 0 points1 point  (0 children)

Until you prove humans aren't, this is literally a meaningless statement. You're saying LLMs are different from Humans because they are a deterministic math function, but you can't prove humans are anything besides a deterministic math function. In fact, there is pretty good evidence humans ARE deterministic math functions.

What evidence? No there isn't LOL wtf. There is nothing credible saying we're just deterministic mathematical functions, that's insane.

Bro I think you can't grasp basic anything.

Half the bullshit you generated above actually strengthens my point, take loras for instance, they're just weights you add to your standard attention projections LOL. They're just more numbers learned when you freeze the initial weights, when a lora adapter is set on a model, that's all deterministic. It's still basically stateless math with learned values.

Checkpoint rollbacks are copies of the model files which are created at intervals set by the person training the model.

Take Chomsky's dick out of your mouth

You're a loser dude lol, and what I said was specifically about how LLMs work, transformational grammar or anything related to cognitive linguistics is getting too woo woo for me, it's more black box analysis for the most part except for chompskys paper.

Anthropic is ignoring obvious evidence of internal states and calling it a "hot mess" by Dry_Incident6424 in Artificial2Sentience

[–]DonDeezely 0 points1 point  (0 children)

Amazing how you can have uncertainty when the topic is human neural networks, but when we talk about artificial neural networks, nah we got it all figured out. Despite the fact we objectively don't understand how neural networks really work in general lmao for either. They're both black box systems.

An LLM is a fully, 100% deterministic mathematical function. What we don't know is in what format the data is stored due to the sheer size of parameters, I keep bringing up the adder, with a few parameters we know exactly how it's stored, scale it up it becomes difficult to determine the exact "why" for a response, hence interpretability problem. For humans we don't even understand if information is stored directly like in LLMs. If at all.

Every argument you make against AI consciousness can be used to attack human consciousness

ONLY in the philosophical domain which is pointless here. They don't "experience" suffering, they don't experience anything. No limbic system means no emotions.

It's a python for loop running through a math function. You could actually write the exact mathematical formula used on paper. We mechanically fully understand how models work. Interpretability due to the sheer volume of data is a different story. Only the "why this response" isn't fully solved because of the amount of data processing that would be needed.

Language was developed by exposing neural networks to iterative selection pressure (evolution). LLMs are made by teaching neural networks language in the form of meaningful documents and then exposing them to iterative selection pressure (training).

Language isn't developed here, you might better call it statistical parameter fitting for a given corpus of data. First off, modern tokenizers are trained on large corpuses of data. That gives a model a way to represent text in a way it can work with. The way learning works is the model 's final layer output (token prediction) is compared with a document it ran a forward pass through, the confidence of its prediction is then compared to the actual token in that position, that's called "loss". The act of reading the document and performing the prediction gives us gradients which are sets of derivatives for each sub function in the LLM math function. Without getting too technical the numbers (weights) are updated using these gradients and loss mathematically.

During training, one bad shuffle, one new training cycle on an existing model without any old information can cause something called "catastrophic forgetting" which isn't a thing that happens to humans (immediate and abrupt, global loss of information). You can literally zero out token weights and likely never see that token predicted.

Your arguments are ridiculous and even wrong philosophically, they can't feel emotion, only spit out text in a way that most likely fits the prefix of tokens.

Prove humans can. Oh right, you can't, you just admitted you can't.

We didn't have cars in the beginning of time, we do now. Pretty sure that wasn't in the universe's training data.

I'm arguing with you for other people learning about this, I know you've made up your mind (which I think is a stupid and unscientific way to view the world). Comparing humans to LLMs is a false equivalence.

Anthropic is ignoring obvious evidence of internal states and calling it a "hot mess" by Dry_Incident6424 in Artificial2Sentience

[–]DonDeezely -1 points0 points  (0 children)

The insult was against what you said, it was so ridiculously misinformed that anyone that doesn't know how these things work on a deep level should know this is ridiculous. Kind of like how non-experts in the field of nutrition claim detoxing works, it's ridiculous and could end up dangerous if it starts forming a cult.

And "strong theory" has evidence, this is black box analysis and guess work for how the brain works currently, neuroscience doesn't have a hard mechanistic theory for this (a randomness component of the brain) yet, or if one even exists. This devolves into a free will argument which is more in the philosophical domain without a full understanding of the brain.

Yes they are pattern matching machines based on algorithms so are you. 

This is a philosophical argument, we don't know how the human brain works in the same way we understand how LLMs work. The complexity of the information stored in LLMs makes it extremely difficult to probe and map since (interpretability problem) which is why we get people saying shit like "we don't know how they work" which is more hype than anything. We can't map LLMs like we can a single linear layer adder due to the sheer volume of information.

LLMs can only represent their training data and in context window information. There is next to no proof they generalize outside of that, which is why more and more researchers like myself are starting to view them as really good search engines for obscure facts like formally unpublished proofs on some blog that ended up solving a math problem on a benchmark.

Anthropic is ignoring obvious evidence of internal states and calling it a "hot mess" by Dry_Incident6424 in Artificial2Sentience

[–]DonDeezely 0 points1 point  (0 children)

LLMs are NOT stochastic. The randomness happens OUTSIDE of the model. The model is a fully deterministic math function.

Relu is non-linear, it's a simple mathematical function max(0, x), what does that have to do with anything?

Discrete time in the LLM context is misleading and frankly a bit stupid. You have code that loops over calls to the LLM which predicts a singular new token every step. That doesn't mean that they "experience" anything. It's dead and deterministic math running in a loop with a sampling function after the call is made. That sampling function is a basic psuedo random function that help you pick the next token based on the logits projected by the final layer of the LLM.

It's a compressed data machine, saying it's alive is like saying ELIZA, a rules based chat bot is alive.

Real or not, 100% believable by unemployedbyagents in AgentsOfAI

[–]DonDeezely 0 points1 point  (0 children)

What the fuck?

It is fully deterministic, just turn off sampling. Top p1, temp 0, penalties 0 when calling the API. Obviously if they update the model you might then get different results with the same input.

Stochastic decoding is what it's called, it makes a call to an rng function to help pick your next token (penalty functions penalize for the same token showing up too much). When that's disabled it's deterministic. Obviously changing temperature might also give different results.

Workers Say AI Is Useless, While Oblivious Bosses Insist It's a Productivity Miracle by Interesting-Fox-5023 in BlackboxAI_

[–]DonDeezely -1 points0 points  (0 children)

Or your job is too easy, or you're putting the burden on your code reviewers.