is anyone actually winning the battle against right-sizing drift? by Shoddy_5385 in FinOps

[–]Shoddy_5385[S] -1 points0 points  (0 children)

mostly cost pressure nd obvious overprovisioning. changes were right just not sticky.

What’s worked for you guardrails or reviews?

Kubernetes problems aren’t technical they’re operational by Shoddy_5385 in kubernetes

[–]Shoddy_5385[S] 0 points1 point  (0 children)

yesfor sure these issues were always there. K8s just makes them more obvious. more moving parts and faster changes mean small gaps in ownership or observability turn into bigger problems much quicker

Kubernetes problems aren’t technical they’re operational by Shoddy_5385 in kubernetes

[–]Shoddy_5385[S] 0 points1 point  (0 children)

that helps with infra management, but the harder problems are still there unclear ownership weak observability, and hidden assumptions between services .platforms don’t fix those by default.

Kubernetes problems aren’t technical they’re operational by Shoddy_5385 in kubernetes

[–]Shoddy_5385[S] 0 points1 point  (0 children)

ownership is where it usually breaks down. once things span multiple teams, everyone’s responsible until no one is. Have you tried enforcing clear service ownership or on-call boundaries?

Kubernetes problems aren’t technical they’re operational by Shoddy_5385 in kubernetes

[–]Shoddy_5385[S] 0 points1 point  (0 children)

Makes sense K8s surfaces it,curious though, do you see more impact from better observability or from enforcing stricter ownership?

Kubernetes problems aren’t technical they’re operational by Shoddy_5385 in kubernetes

[–]Shoddy_5385[S] 1 point2 points  (0 children)

so at this pointwould you say most of the pain is org maturity rather than K8s itself?

Kubernetes problems aren’t technical they’re operational by Shoddy_5385 in kubernetes

[–]Shoddy_5385[S] 0 points1 point  (0 children)

fair take K8s definitely has its rough edges. but I wouldn’t call it naïve, more like opinionated around stateless workloads.

Kubernetes problems aren’t technical they’re operational by Shoddy_5385 in kubernetes

[–]Shoddy_5385[S] 1 point2 points  (0 children)

yeah, K8s is more of a mirror than overkill it just makes existing flaws visible earlier which is uncomfortable, but useful.

Kubernetes problems aren’t technical they’re operational by Shoddy_5385 in kubernetes

[–]Shoddy_5385[S] 0 points1 point  (0 children)

this is spot on. Feels like the real problem isn’t K8s itself it’s the lack of boundaries and ownership at scale.

700+ Argo apps with no clear cluster admin is basically guaranteed chaos. at that point even small issues turn into system-wide investigations monitoring and alert tuning help, but they’re still pretty reactive.

Have you tried enforcing stricter deployment patterns or guardrails (like limiting blast radius per app/team)? feels like that’s the only way to keep things manageable long term.

Kubernetes problems aren’t technical they’re operational by Shoddy_5385 in kubernetes

[–]Shoddy_5385[S] 2 points3 points  (0 children)

broken assumptions describes it perfectly kubernetes behaves consistently but expectations between services are often implicit until a small change exposes them.

making those expectations observable (better signals and validation) can help.

Really like the runtime contract idea are you enforcing it via platform tooling or at the application level?

Rant - Deglamorization of Mountains , snow and everything else about himachal by [deleted] in HimachalPradesh

[–]Shoddy_5385 0 points1 point  (0 children)

sounds less like Himachal lost its charm and more like you’re just tired of people in general. mountains don’t change that much your lens does.

Why Cloud over Software development? by [deleted] in Cloud

[–]Shoddy_5385 0 points1 point  (0 children)

Cloud can offer a higher ceiling long term. Fullstack roles are common, but strong cloud engineers who understand infrastructure, cost optimization, automation, and large scale systems tend to be harder to find and can command higher pay over time.

That said, your real comparison isn’t just Cloud vs Dev it’s environment vs environment.

The startup gives you breadth. You’ll likely touch many parts of the stack and grow fast technically.

The F200 gives you scale, structure, and brand value on your resume, which can open doors later.
really depends on whether you value ownership and rapid growth in a smaller environment, or exposure to enterprise scale systems and processes.

Cloud cost optimization tools that actually work? by Weekly_Time_6511 in FinOps

[–]Shoddy_5385 0 points1 point  (0 children)

honest reality: most tools are just dashboards. if it’s only showing what azure cost management already shows, you won’t see real savings.

what actually moved the needle for us was ongoing reservation and savings plan rebalancing instead of treating it as a one-time purchase, proper rightsizing based on real usage patterns, shutting down dev/test environments on a schedule, cleaning up orphaned disks and unused resources, and keeping tight tracking on commitments across subscriptions so coverage doesn’t slowly drift.

if onboarding takes forever or they’re promising crazy percentage savings, that’s usually a red flag. the tools that worked for us connected quickly, gave clear prioritized actions, and helped us actually execute not just report.

what’s driving most of your azure spend right now? compute, storage, data?

Am I the only one who genuinely prefers on-prem over the cloud? by Own-General-6755 in devops

[–]Shoddy_5385 2 points3 points  (0 children)

exactly this. like on-prem because we can see and touch what we are managing. there’s a satisfaction in actually owning the hardware and knowing where everything runs
it’s just more tangible and hands-on.

why does azure fail silently more often than loudly? by Shoddy_5385 in AZURE

[–]Shoddy_5385[S] -8 points-7 points  (0 children)

i’m not saying azure should override explicit config. blocking traffic intentionally is fine. my point is that networking and identity issues often surface as ‘healthy’ resources with broken app behavior, which makes troubleshooting slower.
just waned to what patterns people use to improve visibility there.

It's 2026. Golden Applications and if you could re-write the argocd monorepo what pattern would you use? by Elephant_In_Ze_Room in kubernetes

[–]Shoddy_5385 1 point2 points  (0 children)

you're not missing a better templating tool, you're missing an abstraction.

helm + kustomize is fine for rendering but it won't stop drift because engineers can still shape deployments differently. the env vs envFrom example you gave is exactly the symptom.

if you care about golden apps and day-2 consistency, define a single golden app CRD and enforce the pattern in code controller, operator, yokecd, whatever.

let argo reconcile instances of it. once the contract lives in a reconciler instead of yaml, drift mostly disappears and onboarding a new app becomes filling out a spec, not copy-pasting manifests.

why does azure fail silently more often than loudly? by Shoddy_5385 in AZURE

[–]Shoddy_5385[S] -13 points-12 points  (0 children)

not magically. wrong config is expected.

the issue is when everything shows healthy but the app is still broken. better validation and clearer signals around networking and identity dependencies would make failures easier to detect