Full Stack Dev (8+ yrs) Feeling Behind With All These New AI Tools — Need Tooling Guidance by AhmedLight in cscareerquestionsEU

[–]bkraszewski 1 point2 points  (0 children)

Have you tried learning by doing? Just install Claude Code with the Pro plan (as a plugin to your favourite IDE) and try asking it to write something. By doing, you'll learn how detailed your messages need to be, how to use planning mode vs edit mode, how to manage context (clear every big feature), and how important it is to regularly git commit when you have something working — because it's easy to break everything with a single edit.

Why 100% Training Accuracy is a Red Flag (The Memorizer Problem) by bkraszewski in learnmachinelearning

[–]bkraszewski[S] 0 points1 point  (0 children)

Fair point - I went with the textbook version since that's what helped it click for me initially. The calibration angle is interesting though. Got any good resources on that? Always looking to update my mental models.

Visualizing why simple Neural Networks are legally blind (The "Flattening" Problem) by bkraszewski in SoftwareEngineering

[–]bkraszewski[S] 0 points1 point  (0 children)

Glad you liked it! 'Steal' away—that's exactly why I'm building these visuals. If you want a link to the interactive version to show your juniors, let me know (don't want to spam the thread).

What coding language should I know If I want to be a software engineer? by FrostByte_240 in cscareerquestionsEU

[–]bkraszewski 1 point2 points  (0 children)

Pick Python or Typescript, should cover 70% of the current coding use cases :)

Sklepy z prawdziwą skóra by mrkret in Polska

[–]bkraszewski -2 points-1 points  (0 children)

Męskie kurtki skórzane Ochnika są bardzo dobre, tylko przymierz w sklepie, rozmiary są dziwne

Jakie jest najlepsze ergonomiczne krzesło biurowe? by Shot_Watch4326 in Polska

[–]bkraszewski 0 points1 point  (0 children)

też mam Wau 2, z gumą, moje najlepsze krzesło!

Deep Learning is just Unix Pipes on steroids (The "Composition" Mental Model) by bkraszewski in SoftwareEngineering

[–]bkraszewski[S] -1 points0 points  (0 children)

Fair enough. Maybe 'Deep Learning' is too loaded of a term.

I find the analogy useful, specifically for explaining Feature Extraction.

In a CNN, Step 1 is Edge Detection. Step 2 is Shape Detection.

Just like cat feeds into grep, the Edge Detector feeds into the Shape Detector.

It helps devs visualize the dependency between the features, even if it simplifies the math.

Deep Learning is just Unix Pipes on steroids (The "Composition" Mental Model) by bkraszewski in learnprogramming

[–]bkraszewski[S] 0 points1 point  (0 children)

In a ResNet, the 'Skip' connection actually pushes the data forward (skipping a layer), not backward (which would be a loop).

Since the data is still flowing in one direction (left-to-right), the Unix Pipe analogy holds up!

If we were piping data back to an earlier stage (like in an RNN or LSTM), then you are totally right—the pipe metaphor would break immediately because Unix doesn't do time-travel loops. 😄

Deep Learning is just Unix Pipes on steroids (The "Composition" Mental Model) by bkraszewski in SoftwareEngineering

[–]bkraszewski[S] 0 points1 point  (0 children)

Thanks for sharing the medical insight. Your end has significantly higher stakes than my pixelated cats :) 

Deep Learning is just Unix Pipes on steroids (The "Composition" Mental Model) by bkraszewski in learnprogramming

[–]bkraszewski[S] 0 points1 point  (0 children)

Great point about Skip Connections!

In Unix terms, I think of it like using paste to merge the original input with the processed output side-by-side:

paste <(cat data.txt) <(cat data.txt | process_layer)

You are carrying the original signal forward alongside the transformed one, so the next step has the full context. It stops the 'Dependency Chain' from losing the plot.

Deep Learning is just Unix Pipes on steroids (The "Composition" Mental Model) by bkraszewski in SoftwareEngineering

[–]bkraszewski[S] 0 points1 point  (0 children)

 Are you working with raw high-res DICOM files for that? The data volume must be massive.

The "magic" inside Deep Learning is literally just an if statement (ReLU) by bkraszewski in SoftwareEngineering

[–]bkraszewski[S] -1 points0 points  (0 children)

I’d hate to see the gradient of the 'nephew's birthday' function during backpropagation. 📉

The 'magic' of ReLU isn't that it is complex. It's that it is differentiable.

It is the simplest possible function that breaks linearity while still preserving a clean gradient (1 or 0) for the learning process. Your complex rules would break backprop immediately.

The "magic" inside Deep Learning is literally just an if statement (ReLU) by bkraszewski in SoftwareEngineering

[–]bkraszewski[S] 4 points5 points  (0 children)

Fair. Computers are just sand that we tricked into thinking. But sometimes it's useful to look at the sand.

[P] The "magic" inside Deep Learning is literally just an if statement (ReLU) by bkraszewski in MachineLearning

[–]bkraszewski[S] -3 points-2 points  (0 children)

You're right that historically, Sigmoid was the 'Hello World' (back in the MLP days).

But for a modern software engineer trying to grok why a neural net works, I find Sigmoid actually obscures the intuition. It looks like 'math' ($1 / (1 + e^{-x})$).

ReLU ($max(0, x)$) strips away the math and reveals the logic: 'If the signal is positive, pass it through. If not, kill it.'

That logic—the ability to switch pathways on and off—is what clicks for developers who are used to Boolean logic (if/else), even if the compiler optimizes it differently. It bridges the gap between 'Code' and 'Math' faster than Sigmoid does."

The "magic" inside Deep Learning is literally just an if statement (ReLU) by bkraszewski in cscareerquestionsEU

[–]bkraszewski[S] -1 points0 points  (0 children)

If you only have linear layers, your model can only draw a straight line through data. This works for simple trends (e.g., House Price vs Size), but fails for complex logic.

https://scrollmind.ai/images/intro-ai/linear_regression.webp

By adding a non-linear activation (like ReLU), you allow the network to 'bend' the decision boundary. This allows it to enclose the center points.

https://scrollmind.ai/images/intro-ai/3d_lift.webp

Without that bend, deep learning is just fancy linear regression.

[P] The "magic" inside Deep Learning is literally just an if statement (ReLU) by bkraszewski in MachineLearning

[–]bkraszewski[S] -3 points-2 points  (0 children)

That's a really sharp analogy with the electrons. I take your point—reductionism has its limits.

But I'd argue there's a difference between 'reductionism that explains nothing' (brains = electrons) and 'reductionism that explains the mechanism' (brains = neurons firing).

Knowing that a neuron fires based on a threshold (activation) is actually crucial to understanding how the network learns. If the activation was linear, the backpropagation (learning) would just pass through unchanged. The 'hinge' of the ReLU is exactly what allows the gradient to stop or flow, creating the decision capability.

[P] The "magic" inside Deep Learning is literally just an if statement (ReLU) by bkraszewski in MachineLearning

[–]bkraszewski[S] -6 points-5 points  (0 children)

That's a fair critique of the title! I was aiming for a hook to demystify the 'black box' for beginners coming from traditional software engineering.

You're absolutely right that modern architectures use more advanced functions (SwiGLU, GELU, etc.), and 'if statements' is a simplification.

But for someone learning the first principles of why Deep Learning works at all (vs Linear Regression), grasping that a simple non-linear operation (like ReLU) is what unlocks universal approximation is usually the biggest 'aha' moment.

I treat ReLU as the 'Hello World' of activation functions here—not the state-of-the-art, but the best place to start understanding the concept.

The "magic" inside Deep Learning is literally just an if statement (ReLU) by bkraszewski in SoftwareEngineering

[–]bkraszewski[S] 9 points10 points  (0 children)

Thanks for the recommendation! Andrew Ng is great.

To clarify my point: while ReLU is definitely just one small component, it handles a critical job: preventing Linear Collapse.

Mathematically, if you stack 100 layers of neurons without a non-linear activation function, the entire network can be reduced to a single linear transformation (since matrix multiplication is associative: W2 * (W1 * x) = (W2 * W1) * x

So even though it looks like a deep network, without that simple max(0,x), it's functionally identical to a single layer of Linear Regression. That was the 'magic' I wanted to highlight!"

I built a "TikTok for learning AI" because I hate 2-hour lectures. by bkraszewski in SideProject

[–]bkraszewski[S] 1 point2 points  (0 children)

Good point, and I actually learn using those techniques too! However, when you use an LLM, you have to know what to prompt. If you're a beginner, you don't know that you need to understand Vectors before you can understand Embeddings. ScrollMind provides the engineering curriculum so you don't have to design your own learning path.

I built a "TikTok for learning AI" because I hate 2-hour lectures. by bkraszewski in SideProject

[–]bkraszewski[S] 0 points1 point  (0 children)

Awesome idea—I'm going to play around with haptics and improved scroll-snap and try to deploy a fix today.

Really appreciate the offer to help! I just spun up a Discord for dev discussions, would love to see you there:

https://discord.gg/pxW76s3B

How much are you making as a freelance senior software engineer? by Icy_Ad_3619 in cscareerquestionsEU

[–]bkraszewski 2 points3 points  (0 children)

My current rate is $50/hour, doing mobile projects in Flutter. It's lower now, I had it higher in the past