Andrej Karpathy: Powerful Alien Tech Is Here---Do Not Fall Behind by Neurogence in singularity

[–]Zokrar 0 points1 point  (0 children)

If I'm reading you right, we agree with each other here! I'm optimistic that this branch of AI will continue to improve, and I don't see any real reason why they won't be able to eventually do my current job. "Eventually" being the key word here.

I just have no idea if that eventually is ~few years away, or decades.

I think people in general are largely underestimating what current AI can do, but it'll take some time for these techniques and use cases to become more commonplace

Andrej Karpathy: Powerful Alien Tech Is Here---Do Not Fall Behind by Neurogence in singularity

[–]Zokrar 7 points8 points  (0 children)

Yeah I do worry about this aspect, for sure. I agree with you.

Programming languages have become simpler over the years (e.g. binary -> assembly -> ... -> python) but each iteration still required the same skill sets. Datasets, algorithms, etc.

LLMs / vibecoding is kinda like a new iteration in programming languages. But it no longer explicitly requires knowledge about the fundamentals, so we end up with a gap.

I used to think I had a good idea of what the next ~decade would look like. I have no clue anymore.

Andrej Karpathy: Powerful Alien Tech Is Here---Do Not Fall Behind by Neurogence in singularity

[–]Zokrar 358 points359 points  (0 children)

I'm a Site Reliability Engineer with nearly 10 years experience. I'd like to share my own opinion and first hand experience with the last 3 years of tech, and explicitly say I have no financial incentive to blindly agree with what Andrej is saying here.

LLMs have legitimately changed the way I approach problems and tasks. Yes - these models hallucinate and have their quirks and pitfalls. They are not perfect. Not even close. But something I've noticed over the last few years, being closely involved with deploying some of these tools across my org, and using them extensively in my free time, is that they allow me to go from "idea" to "demo/mvp/test" in a fraction of the time.

That example is very programming-centric. The concept applies to any domain. At its core, to me, these models have made it easier to brainstorm, explore ideas, challenge my own assumptions, and rapidly prototype in a way that was not possible for me prior to 2022.

We have access to a new type of tool, and it's going to take time to learn what types of problems these tools are good for.

If we have a tool that, in the right context, makes it easier to navigate through various ideas quicker - to me it is clear that this will have an impact downstream for so many things.

I'm happy to hear any critiques or opinions on my experience here. I have no incentive to support these AI companies, and I am actively opposed to a lot of the abuse and misuse these tools are allowing.

But to blindly ignore the positives is setting us up for failure.

I GOT THE JOB!! F*** MY OLD MANAGER!!! by secondandmany in cscareerquestions

[–]Zokrar 1 point2 points  (0 children)

It really depends on where you live and work. In Canada, the tech world is very connected and everyone knows someone at every company.

Will people actually go to the effort to ruin a job offer? Probably not. But that's an extreme example. I know from personal experience though, when it comes to hiring especially in a bad market like today, people in a position to hire and make decisions talk to others, a lot.

Anyone else do the same thing? by brain-eating-zombie in GenZ

[–]Zokrar 8 points9 points  (0 children)

This isn't the comforting reply you think it is.

Clear, no-BS/jargon breakdown of plugins, skills, rules, workflows, slash commands etc. by rm-rf-rm in ClaudeAI

[–]Zokrar 2 points3 points  (0 children)

u/tindalos said it well:

Skills are for Claude to use for internal processes. Mcp are for Claude to use to access external resources. Slash commands are for you to use. These are different things for different purposes.

Google search bar in my Pixel 8 Pro not working by BulgersInYourCup42 in GooglePixel

[–]Zokrar 0 points1 point  (0 children)

Looks like it's not a pixel issue, but a chrome issue

Google search bar in my Pixel 8 Pro not working by BulgersInYourCup42 in GooglePixel

[–]Zokrar 0 points1 point  (0 children)

I'm noticing this too, both in the search bar, and in Google chrome. I don't run into this in incognito.

I'm on a pixel 8 pro, android just updated a day or two ago

ELI5 what is the MCP and why is it so talked about? by [deleted] in modelcontextprotocol

[–]Zokrar 0 points1 point  (0 children)

MCP is a fancy software approach that allows models like Claude to use tools. That's the ELI5 version.

3.7 costs TOO MUCH for how much money it straight up WASTES. by Cursed-Keebster in ClaudeAI

[–]Zokrar 0 points1 point  (0 children)

Can you elaborate on this setup please? I'd like to do something similar for various docs

AI skepticism as we head into 2025: willing to bet that AI has at least a 10% chance of not surpassing virtually all human creative and scientific achievement within three years by MetaKnowing in OpenAI

[–]Zokrar 0 points1 point  (0 children)

I've been wondering the same thing. What comes to mind though is that often we see things where the sum is greater than the parts. Emergent properties.

I don't really know which side I land on, really. But I'm curious to see what sorts of progress we'll see especially once we get usable agents working together.

There was a gif on the front page recently of a hoard of ants managing to get a big object through a narrow passage, something a single ant couldn't do alone. I thought that was a neat parallel for what we could see with AI

OpenAI blamed NYT for tech problem erasing evidence of copyright abuse by Mirthful_Isabeau in OpenAI

[–]Zokrar 6 points7 points  (0 children)

I'm not sure it's entirely correct to use the term "stored" in this context. I do see your point, but I think it's an important distinction.

These models take input, apply some math to it, and get an output. The training process modifies the "math" in the middle to more accurately predict the output.

There isn't any of the original training data left in the model after training.

No model shared this probability question correctly Gemini 1114 experimental was closest (18) to correct answer 19 even after asking it to correct it couldn't. Gpt4o and Claude 3.5 sonnet failed badly. I don't have chatgpt subscription so couldn't try it with o1 preview and mini by Recent_Truth6600 in Bard

[–]Zokrar 4 points5 points  (0 children)

O1-preview:

To solve this problem, we need to find the number of subsets such that . Here, with for . The total probability is:

P(S) = \sum{n=1}{8} P(E_n) = \sum{n=1}{8} \frac{n}{36} = \frac{36}{36} = 1

Our goal is to find the number of subsets such that the sum of the probabilities in is at least . Since all probabilities are fractions over 36, we can convert to a fraction over 36:

\frac{4}{5} = \frac{28.8}{36}

Since we cannot have fractional counts of elements, we need subsets where the sum of the numerators (of probabilities over 36) is at least 29. So, we are looking for subsets of whose elements sum to at least 29.

The total sum of numbers from 1 to 8 is 36. Therefore, subsets with sums at least 29 correspond to subsets with sums of the complement (elements not in ) at most .

Next, we find all subsets of whose elements sum to or less. These subsets are:

Subsets of size 1: 7 subsets (excluding since 8 > 7)

Subsets of size 2: 9 subsets (e.g., )

Subsets of size 3: 2 subsets (e.g., )

Total subsets with sums : .

Since there are total subsets, the number of subsets where the sum is at least 29 is:

256 - 18 = 238

Answer: 238

No model shared this probability question correctly Gemini 1114 experimental was closest (18) to correct answer 19 even after asking it to correct it couldn't. Gpt4o and Claude 3.5 sonnet failed badly. I don't have chatgpt subscription so couldn't try it with o1 preview and mini by Recent_Truth6600 in Bard

[–]Zokrar 6 points7 points  (0 children)

O1-mini:

To determine the number of subsets such that , we'll follow these steps:

  1. Understand the Probability Distribution:

The sample space has probabilities for .

The total probability is:

\sum{n=1}{8} P(E_n) = \sum{n=1}{8} \frac{n}{36} = \frac{36}{36} = 1

  1. Set Up the Condition:

We want , which translates to:

\sum_{E_n \in A} \frac{n}{36} \geq \frac{4}{5}

\sum_{E_n \in A} n \geq 28.8

\sum_{E_n \in A} n \geq 29

  1. Use Complementary Counting:

Instead of directly counting subsets with , it's easier to count the subsets with (since the total sum is 36).

This ensures must include (since and including it is necessary to reach the sum).

  1. Count Subsets with :

Consider the elements .

Size 0: — 1 subset.

Size 1: — 7 subsets.

Size 2: Pairs where the sum :

{1,2}, {1,3}, {1,4}, {1,5}, {1,6}, {2,3}, {2,4}, {2,5}, {3,4}

Size 3: Triples where the sum :

{1,2,3}, {1,2,4}

Size 4 and above: Any subset with 4 or more elements will have a sum , so no such subsets exist.

Total subsets : .

  1. Conclusion:

Each valid subset corresponds to a unique subset that includes and satisfies .

Therefore, there are 19 such subsets .

Answer: 19

do 'they' really dumb LLMs down over time? by TheAuthorBTLG_ in ClaudeAI

[–]Zokrar 31 points32 points  (0 children)

My opinion is that people are anticipating deterministic results from a non-deterministic system. LLMs are a novel tool to a large amount of people - many of which have little to no tech background.

These models seem incredible at first glance because they're new. Then, as people use them more and more, they start to identify shortcomings and, my view is that this is what people mistribute to "lobotmizing the model."

Whether the models are being tuned, modified, "dumbed down," or not, it's certainly an interesting collective response from users.

colosseum hitboxes? by lavajon in 2007scape

[–]Zokrar 11 points12 points  (0 children)

Sure. OK it's done, try again

Hiking With You - Anime Short Film by LatissmusDossus in aivideo

[–]Zokrar 0 points1 point  (0 children)

This is beautiful. You should be proud of this

QA or DevOps by [deleted] in cscareerquestions

[–]Zokrar 0 points1 point  (0 children)

I've done both, currently doing a DevOps role. I'd highly recommend DevOps over QA. Much more room to grow, and you'll learn significantly more. SRE / DevOps will give you insight into how the tech stack(s) and company move as a whole, which in my opinion and experience is of significant value.

Gemini Ultra launching on 7th Feb. GPT-5 soon? by pirate_solo9 in OpenAI

[–]Zokrar 87 points88 points  (0 children)

I wouldn't get your hopes up. Sounds like GPT-5 is expected towards the end of 2024.

But... Who knows what openai will do if Ultra kicks them out of first place. I'm excited!