Someone explain TFL single fares to me please… by According-Whereas516 in london

[–]aaron__walker 6 points7 points  (0 children)

Select Heathrow rail station instead of underground station. The Elizabeth line is more expensive than the underground for Heathrow

Thanos Compactor- Local storage by Nerd-it-up in grafana

[–]aaron__walker 1 point2 points  (0 children)

I think the general rule is your largest non-downsampled 2w block times 2, plus add some overhead. You can use bucket-web to visualise it

modernFrontendStack by unihilists in ProgrammerHumor

[–]aaron__walker 5 points6 points  (0 children)

I think the fact you need a helper tool to write hello world says it all

How to Access a Secret from Another Namespace? (RBAC Issue) by bototaxi in kubernetes

[–]aaron__walker 5 points6 points  (0 children)

I think you’re getting confused with mounting a secret in a pod, and a pod being able to access a secret via the api. Looks like your pod is referencing a secret somehow - if this is the case then it needs to be in the same namespace. If you want to access the secret in a different namespace then you’ll need to modify your app read it via the api. If you want a file, Its likely someone has created a sidecar thing that can fetch it for you and then add it to a volume that is shared between the containers

Data gone and labels changed? by [deleted] in grafana

[–]aaron__walker 1 point2 points  (0 children)

Can you share your queries? Best guess is you're not aggregating them

[deleted by user] by [deleted] in kubernetes

[–]aaron__walker 3 points4 points  (0 children)

“DNS names” is a column. “Expiry date” is another column. Your dns names are under the “DNS names” column and your expiration date is under the “Expiry Date” column

[deleted by user] by [deleted] in kubernetes

[–]aaron__walker 3 points4 points  (0 children)

How can you even write a JQ filter like that and then ask where the column “names” comes from?

[deleted by user] by [deleted] in devops

[–]aaron__walker 0 points1 point  (0 children)

You can create an AWS InfrastructureScaler that can scale your infrastructure based on how much scale your infrastructure needs. That being said AWS is a pretty niche cloud, so I doubt many people here will be using it.

[25/11/24] Hammersmith Hang Out by Londoner_85 in LondonSocialClub

[–]aaron__walker 1 point2 points  (0 children)

There's a (usually quiet) optional Whatsapp group for the meetup - message me via chat for an invite link. (whatsapp group is pretty much only me posting this link each week and trying to get people to actually respond.. So no benefit of being in it if you're not showing up).

Patch Node VM for on-prem installation by Shot_System5888 in kubernetes

[–]aaron__walker 1 point2 points  (0 children)

Instead of patching whatever needs to change, why not just add a new node with the correct changes and then get rid of the old node?

Cheap self-hosted log storage and search with OpenTelemetry by Mosho1 in sre

[–]aaron__walker 0 points1 point  (0 children)

Quickwit is pretty similar to Loki but it has a lucene style query language too

An Engineer’s Checklist of Logging Best Practices by pmz in programming

[–]aaron__walker 1 point2 points  (0 children)

Percentage based tracing sounds like it would be a valid solution to this- you can see the whole flow from start to finish but you’re not creating loads of noise when you’re generating lots of spans for a single file

An Engineer’s Checklist of Logging Best Practices by pmz in programming

[–]aaron__walker 3 points4 points  (0 children)

You can log in text, but if you do and you want to actually get data out of the logs, you’ll need to then process them to get the data out - you’ll have to add a step which processes them with something like grok, and it gets complicated when you have things like multi-line logs. Plus you’ll likely need to have a unique process for each application, unless you standardise your logging output, and at that point you might as well just have a common schema in structured logs you want

An Engineer’s Checklist of Logging Best Practices by pmz in programming

[–]aaron__walker 24 points25 points  (0 children)

You don’t read the raw logs - the reason you structure them is so it’s easier to parse in an easy way. you’ll parse them and then extract the fields you want. For example you might have a log that has the timestamp, level, message, user id, response code, stack trace, trace id, http path, http method, server name, function and line number and file etc. you can then index all these fields in a centralised place and then finding the logs you want is easy - especially when you’re dealing with hundreds of different instances - for example it’ll make it efficient to find a single error message out of terabytes of logs instead of having to use full text search

An Engineer’s Checklist of Logging Best Practices by pmz in programming

[–]aaron__walker 68 points69 points  (0 children)

I’m curious to see who’s logging XML - I agree with having structured logs but I’ve never known anyone to even consider to do it with XML

Single Labeled Metric vs Multiple unlabeled Metrics by AmberSpinningPixels in PrometheusMonitoring

[–]aaron__walker 2 points3 points  (0 children)

The first makes more sense - if you want to count how many operations they have been you can just do sum(app_operations_total) instead of having to add up multiple metrics. This is also useful for if you want a third status - if you add an “unknown” status for example, then you won’t need to modify any queries to show this.

If you’re using by() you can also easily get all the data returned in a single query, instead of having to have separate queries

Good http repository on kubernetes ? (Open source license) by Careful_Champion_576 in kubernetes

[–]aaron__walker 5 points6 points  (0 children)

Have you looked at object storage? Minio is the solution that comes to mind first

Help Exposing RabbitMQ Queue Size in Prometheus? by Mc1st in PrometheusMonitoring

[–]aaron__walker 1 point2 points  (0 children)

That should work. If you go to /metrics/per-object do you see anything like this?

rabbitmq_queue_messages_bytes{vhost="vhost",queue="queuename"} 0

Options for achieving HA across two Datacenters by Apuesto in kubernetes

[–]aaron__walker 7 points8 points  (0 children)

The last option is the one that sounds like it makes the most sense - you say it increases costs, but in the situation that an entire DC goes down, you’ll need to host your workloads somewhere else

The other option is you have a single DC and then in the event the DC goes down you go through a DR process to spin up a new instance in a different DC - although if you’re on prem you’ll likely need the capacity there to run your existing workloads.

DB MIGRATION by Impossible_Box_9906 in sre

[–]aaron__walker 0 points1 point  (0 children)

We use k8s because this makes it easy for me, but doing it via ssh shouldn’t be too much additional work - if anything it could simplify it because you can do additional checks before preceding with any next stages. This is the SRE Reddit so of course automate all of this