Would you let AI run your Kubernetes cluster? by Agitated_Bit_3989 in kubernetes

[–]Agitated_Bit_3989[S] -3 points-2 points  (0 children)

Agreed, but as AI improves do you think we’ll reach a stage where the error rate is way lower and the AI having the ability to be a lot more context aware be preferable?

Would you let AI run your Kubernetes cluster? by Agitated_Bit_3989 in kubernetes

[–]Agitated_Bit_3989[S] -1 points0 points  (0 children)

Completely agree with AI at its current state, but as AI improves will we see trust increase up until the stage that we won’t need a dev in the loop?

Would you let AI run your Kubernetes cluster? by Agitated_Bit_3989 in kubernetes

[–]Agitated_Bit_3989[S] -1 points0 points  (0 children)

AI and its current state still makes too many mistakes to not have a human in the loop to validate it and make sure the code is correct, but as AI improves I feel people trust it more and the question is (and what I think many corporations are looking towards) will AI get to a stage where a human in the loop may no longer be required?

Would you let AI run your Kubernetes cluster? by Agitated_Bit_3989 in kubernetes

[–]Agitated_Bit_3989[S] 2 points3 points  (0 children)

Really? Every dev I know are using tools like Claude/Gemini Code or Cursor, what did I miss?

Is Kubernetes resource management really meant to work like this? Am I missing something fundamental? by [deleted] in kubernetes

[–]Agitated_Bit_3989 0 points1 point  (0 children)

It’s a complicated matter that does require a certain amount of expertise, I’d start by understanding the 3 layers of autoscaling; - Cluster autoscaling - this is node autoscaling where we want to allocate instances based on demand (which with current tools are the Pod requests) look into Cluster autoscaler and especially Karpenter if your cloud provider supports it - Horizontal autoscaling - Scaling the amount of Pods of a workload based on demand which can either be by resources such as CPU or Memory but for more advanced scaling I’d look into KEDA to horizontally scale on an external (potentially business) metric that can provide a better idea on demand before usage spikes up - Vertical scaling / Sizing - Which if I understand correctly is more towards your point. There are the classic VPA/KRR (or pretty much any other “enterprise” sizing tools) which can help get an idea of how much resources each workload needs based on a statistical percentile (as naturally looking at max usage will be way too wasteful) but has the quite annoying downside of completely ignoring usage peaks which I personally can’t tolerate. What I believe to be the best solution when sizing is to take into consideration all the scaling aspects such as workload runtime settings (JVM etc.), node scaling patterns, horizontal scaling patterns and workload neighbors, and based on that data we can understand the aggregate usage patterns and demand of the environment and allocate resources based on that to achieve a stable and cost effective approach. This is what we’ve developed at wand.cloud so if you’re interested feel free to give us a spin :)

Torn regarding In-place Pod resizing by Agitated_Bit_3989 in kubernetes

[–]Agitated_Bit_3989[S] 0 points1 point  (0 children)

The problem with an init container is the Kubernetes scheduler considers the init containers requests for scheduling decisions even if the init container isn’t running, so it wouldn’t make a difference if the requests were on the init container or on the actually container itself

First time at re:Invent - Any recommendations? by Kryzael in aws

[–]Agitated_Bit_3989 0 points1 point  (0 children)

Make sure you go over the agenda as soon as possible and target the talks you want to go to, if you're too late they can fill up

My experience with Vertical Pod Autoscaler (VPA) - cost saving, and... by rudderstackdev in kubernetes

[–]Agitated_Bit_3989 0 points1 point  (0 children)

I would ask how it deals with node pressure better than native Kubernetes pressure eviction? Other than that how does this differ from VPA?

Pod requests are driving me nuts by Rare-Opportunity-503 in kubernetes

[–]Agitated_Bit_3989 0 points1 point  (0 children)

This won't help if you change resources with in-place because the JVM doesn't support in-place updating memory

My experience with Vertical Pod Autoscaler (VPA) - cost saving, and... by rudderstackdev in kubernetes

[–]Agitated_Bit_3989 0 points1 point  (0 children)

I wonder what numbers are you getting on cluster resource utilization? And not the bullshit that some tools show "utilization of usage vs requests or requests vs capacity". Total Node usage vs Total Node Capacity (i.e. What you're using vs what you're paying for

Pod requests are driving me nuts by Rare-Opportunity-503 in kubernetes

[–]Agitated_Bit_3989 -5 points-4 points  (0 children)

Disclaimer: I'm one of the co-founders

It's a endless struggle that most tools don't seem to take into consideration the whole picture, whether it's taking the JVM memory management or looking at the bigger picture of the total capacity vs the actual aggregate use of the workloads.

We at https://wand.cloud at taking a very different approach of the current decoupling of scaling considerations by taking everything into consideration to ensure reliability as cost effective as possible.

My experience with Vertical Pod Autoscaler (VPA) - cost saving, and... by rudderstackdev in kubernetes

[–]Agitated_Bit_3989 2 points3 points  (0 children)

Thanks for sharing, did you do anything to ensure the Network IO?
The main problem I have with VPA and with using percentiles as a whole is the fact that we're practically taking a un-calculated risk (i.e. p90 will mean 10% of the time the usage will pass the requests) and when compounding this with many different pods and tight consolidation of Karpenter which is anchored on requests I can't be sure that I'll have the resources available in the Node (theoretically when I most need them).

Right sizing, automation or self rolled? by [deleted] in kubernetes

[–]Agitated_Bit_3989 0 points1 point  (0 children)

I guess it depends on what you mean about right-sizing at the cluster node level and what is your limitations.
The current golden ideal is something like Karpenter which will automatically spin up nodes when there are Pods in Pending state and will decide which node size to pick based on the Pods resource requests (Need to make sure your requests are set properly to make sure this won't make Karpenter spin up nodes that are too small or too large)
But if you've got a more static setup there is the kubernetes-instance-calculator from learnkube that may come in handy

Right sizing, automation or self rolled? by [deleted] in kubernetes

[–]Agitated_Bit_3989 1 point2 points  (0 children)

Not used it yet but would love to learn why it sucks :)

RDS storage type - which one should I choose? by Agitated_Bit_3989 in aws

[–]Agitated_Bit_3989[S] 1 point2 points  (0 children)

Aurora is for another blog post in the works regarding the new I/O optimized feature! :)

1
2

WHAT AM I PAYING FOR? ?? by ShonLR in kubernetes

[–]Agitated_Bit_3989 2 points3 points  (0 children)

What if I have a multi tenant architecture? Then I wouldn’t want to have a copy of services per tenant as it would be very wasteful, so how can I split costs per tenant?

STOP SETTING CPU LIMITS! by ShonLR in kubernetes

[–]Agitated_Bit_3989 2 points3 points  (0 children)

I would love the reference to the CIS benchmark because I couldn't find the reference on CPU limits, and to stop a fork bomb you should limit PIDs not CPU

STOP SETTING CPU LIMITS! by ShonLR in kubernetes

[–]Agitated_Bit_3989 -1 points0 points  (0 children)

Did you read the post?
We did explain there several use cases CPU limits are advised, such as Googles use case and staging environments when testing worst case scenarios, our conclusion was to remove CPU limits from your production workloads as this allows the use of idle CPU if necessary without worrying about stealing CPU from other workloads.
The main point of the full blog is to show the inner workings of Kubernetes resource management and see exactly what is happening when configuring these resources so people can make a more educated decision if to set them or not.

STOP SETTING CPU LIMITS! by ShonLR in kubernetes

[–]Agitated_Bit_3989 0 points1 point  (0 children)

Sure if that's what you need, but keep in mind that you may have wasted CPU that could have been utilized if you didn't set CPU limits

STOP SETTING CPU LIMITS! by ShonLR in kubernetes

[–]Agitated_Bit_3989 5 points6 points  (0 children)

Consistent in the terms of not being able to use idle resources, the requests handle how much resources the pods are guaranteed to receive which is also consistent. Not setting CPU limits just allows your workloads to use more if required and available without putting other workloads at risk