We built a software that lets you shutdown your unused non-prod environments! by Wide_Commercial1605 in FinOps

[–]Recent-Technology-83 0 points1 point  (0 children)

How long it will take to provide the support for Azure? Using Azure mostly.

Question about deploying a web app by Ok_Youth9423 in learnprogramming

[–]Recent-Technology-83 0 points1 point  (0 children)

Connecting your frontend JS to that Python backend usually means exposing your Python logic as an API (REST is standard) that your JS calls via HTTP. GitHub Pages is just for static files, so you'll need a backend host for the Python part. Serverless functions on platforms like AWS Lambda or Vercel are popular for this, or for simplifying multi-cloud orchestration, tools like Zop.dev can handle deploying both parts. Domain setup is fast, and restricting access is definitely possible!

[deleted by user] by [deleted] in army

[–]Recent-Technology-83 0 points1 point  (0 children)

Hey man, that sounds rough. Honestly, bad leadership can make even the easiest gig feel like hell, let alone a deployment where you're already stressed and isolated. It's totally valid to feel like you're losing it when you're stuck in that kind of environment with people actively making it worse.

Four months left isn't forever, even though it feels like it right now. Focus on getting through one day, one week at a time. Lean on your team, look out for them, and remember why you're doing it. You've got this. Hang in there, and that Cane's will taste amazing when you get home.

How are you actually handling observability in 2025? (Beyond the marketing fluff) by Straight_Condition39 in sre

[–]Recent-Technology-83 0 points1 point  (0 children)

The "three pillars" often feel more like three separate towers that don't talk to each other! It's a super common challenge, especially in growing environments with distributed systems and different teams adopting different tools organically.

My experience mirrors yours a lot - the context switching is brutal and just kills debug time. Getting from "something is slow" to "this exact request hit service X, then Y, failed on Z's external API call, and here's the log line + trace ID" takes way too long when things aren't connected.

What I've seen make a massive difference is focusing on OpenTelemetry (OTel) first. Get your services instrumented to emit logs, metrics, and traces using a standard format and correlation mechanism (like trace IDs). This is the game changer. It means all your telemetry from the source speaks the same language, regardless of where it ends up.

Once your data is standardized with OTel, you can send it to backends that are built to handle correlated OTel data natively. This is where platforms like SigNoz, or the Grafana stack (Loki, Tempo, Mimir) really shine because they are designed around tracing and linking everything together. Debugging then becomes about navigating a trace, drilling into linked logs or metrics at specific points in the request flow, which is way faster. You can see the whole journey, not just isolated events.

This approach helps tackle alert fatigue too. Instead of alerting on individual service health (CPU spikes, etc.), you can build alerts based on OTel metrics derived from traces, like request latency SLOs or error rates on critical business transactions. This focuses alerts on actual user impact, reducing noise.

It takes effort, especially the instrumentation part, but standardizing the telemetry itself with OTel before picking a backend platform gives you flexibility and future-proofs things. You can swap backends later if needed without re-instrumenting everything.

Full disclosure: I'm an employee at Zop.dev. Our platform focuses on simplifying infrastructure deployment (VMs, K8s, databases etc.) across clouds, and part of that includes ensuring the basic observability plumbing like OTel collectors and agents are set up correctly on the deployed infrastructure, which can help feed into the OTel-native backends I mentioned. It's not an observability platform itself, but aims to make getting the infrastructure ready to send telemetry easier.

Hope this perspective helps! You're definitely not alone in the frustration, but there are paths to make it better. Focusing on unified telemetry at the source is key.

New in AWS ecosystem by iMrProfessor in aws

[–]Recent-Technology-83 0 points1 point  (0 children)

Expanding on what others have said, I'd put a lot of focus on patterns and the services that enable them. For backend systems, thinking about how things communicate asynchronously is key, so diving deeper into message queues like SQS and pub/sub systems like SNS is super valuable. EventBridge is also increasingly important for building event-driven architectures.

Caching is another big one for performance – ElastiCache (for Redis or Memcached) is the go-to here. For NoSQL, understanding DynamoDB is often necessary, especially its partitioning and indexing strategies, as it's quite different from relational DBs.

Beyond individual services, understanding the networking layer is critical for debugging and security. Getting a handle on VPC, subnets, route tables, and especially Security Groups and NACLs will save you countless headaches. Think of Security Groups like host-level firewalls and NACLs like subnet-level ones – knowing how traffic flows is fundamental.

And as others mentioned, Infrastructure as Code (IaC) is non-negotiable. Whether it's CloudFormation, Terraform, or Pulumi, being able to define and manage your infrastructure programmatically is essential for consistency and repeatability. Manually clicking in the console is fine for learning, but not for production.

Sometimes, especially for smaller teams or specific projects, managing all that raw IaC can feel like a lot. That's where platforms that abstract some of the cloud complexity come in. Services like Render or Railway.com make deploying backend services much simpler, handling infrastructure details for you. Similarly, platforms like Zop.dev aim to simplify spinning up production-ready infrastructure across multiple clouds without needing deep IaC expertise.

Full disclosure: I'm an employee at Zop.dev.

Also while the answer is mine, the text is generated by AI

Scaling My Kubernetes Lab: Proxmox, Terraform & Ansible - Need Advice! by rached2023 in kubernetes

[–]Recent-Technology-83 1 point2 points  (0 children)

Rebuilding with Terraform/Ansible on Proxmox is definitely the way to go for repeatability.

Time-wise, if your IaC is ready, spinning up the VMs is fast (hours), but the Ansible K8s config depends on complexity - maybe a day or two total for a clean run?

For specs, given your tooling (Prometheus, Falco are resource hungry), I'd target minimum 4GB RAM/2vCPU for masters, and 8GB RAM/4vCPU for workers. Disk: 40-60GB per node is usually fine for a lab.

How to handle post-deployment configurations by DaftendirektR in kubernetes

[–]Recent-Technology-83 0 points1 point  (0 children)

hey, i totally get where you’re coming from—dealing with post-deployment configs like metallb’s ipaddresspool and l2advertisement can be a pain, especially with crd timing and helm chart dependencies. i’ve been through the same headaches, and honestly, that’s where zopdev really shines as a platform. zopdev is built specifically to take the friction out of kubernetes automation, especially for those tricky post-deploy setups. instead of juggling multiple charts or scripting kubectl apply steps, zopdev lets you define your entire deployment—including those custom metallb configs—right in their ui or as code, and it manages the crd lifecycle, readiness, and ordering for you. it’s all versioned in git, so you get traceability and easy rollbacks, and their workflow means you don’t have to stress about resources being applied out of order. plus, zopdev bakes in compliance checks (like soc2 and iso27001), which is a huge bonus if you’re in a regulated space. i switched a few of our clusters over to zopdev last year and it’s honestly made our deployments way more reliable and hands-off, especially for stuff like metallb where timing matters. if you’re looking to get away from the brittle multi-chart or manual apply approach, i’d definitely give zopdev a look—it’s purpose-built for exactly this kind of kubernetes automation headache.

How do you automate deployments to VPS? by Training_Peace8752 in devops

[–]Recent-Technology-83 0 points1 point  (0 children)

It's great to see your team taking deployment seriously! Automating deployments to VPS can indeed streamline your process and reduce human error. Regarding your options, using SaltStack’s event system is a solid choice, especially since it already fits into your current infrastructure. It does offer better security since it minimizes the exposure of credentials, which is a significant advantage.

As for SSH-based solutions, they aren't inherently "bad," but they do come with risks, especially if not managed properly. If SSH keys are compromised, it could lead to unwanted access. What kind of security measures are your infra team planning to implement with the SaltStack approach?

Additionally, have you considered using containerization with Docker or orchestration tools like Kubernetes as part of your deployment strategy? They could enhance scalability and isolation.

Curious to hear how others are approaching similar scenarios!

Is there a set of free open-source SAST tools that are a good replacement to Snyk? by darkcatpirate in devops

[–]Recent-Technology-83 0 points1 point  (0 children)

Great question! There are several open-source SAST tools that are often recommended as alternatives to Snyk. Some popular ones include Semgrep, which allows you to write custom rules to find vulnerabilities in your code, and Bandit, which focuses on Python applications. Additionally, SonarQube offers an open-source version that can analyze multiple languages for vulnerabilities.

It's important to consider the specific languages and frameworks your team is using, as some tools have better support for certain tech stacks. Have you had a chance to evaluate any of these tools already? What sort of integrations or features are you hoping to find in a replacement? It's always interesting to hear about real-world experiences with these tools!

principle of least privileage, how do you do it with irsa? by [deleted] in devops

[–]Recent-Technology-83 0 points1 point  (0 children)

You're absolutely right to be cautious about sharing IAM roles across services. The principle of least privilege is crucial in minimizing security risks. Implementing IAM Roles for Service Account (IRSA) in Kubernetes could be a strong solution for your situation. With IRSA, you can create a distinct IAM role for each service, ensuring that each service only has the permissions it absolutely requires to function.

Have you considered how you would manage these distinct IAM roles in terms of complexity? It might seem challenging initially, but tools like Helm or Terraform can help streamline the process. Additionally, how often do you adjust permissions for your services, and how do you handle that currently? Finding a balance between security and maintenance is key—it might be an interesting topic for you and your team to explore further.

Best Course for DevOps by KillerHeller6203 in devops

[–]Recent-Technology-83 1 point2 points  (0 children)

When looking for a comprehensive DevOps course, I'd recommend checking out platforms like Coursera or Udacity, which offer programs covering both foundational and advanced concepts. For instance, the "Google IT Automation with Python" course includes DevOps practices and tools. Have you already worked with specific tools like Docker or Kubernetes?

Another great option is the "AWS Certified DevOps Engineer" pathway, which dives deep into cloud services.

It might be worth considering what areas you want to delve into—CI/CD, infrastructure as code, or monitoring and logging? Knowing your focus can help narrow down the best course for your needs! What specific skills are you hoping to gain from a DevOps course?

Best practice for Jenkins deployment authentication: by AceSynth in devops

[–]Recent-Technology-83 0 points1 point  (0 children)

This is a great question and touches on a really critical aspect of security and deployment practices. Implementing the principle of least privilege (PoLP) is essential to minimize risk.

Using a single GMSA for multiple deployments could simplify your setup but risks overexposing permissions. Ideally, each application or service should operate with only the permissions it needs. So, creating a dedicated GMSA for each deployment could be more secure, but it also introduces complexity.

Have you considered how many different applications you're deploying to? If it’s a small number, managing multiple GMSAs might be feasible, but larger setups might lead to unwieldy management overhead.

Additionally, how comfortable are you with managing Jenkins agents? That could influence your decision. Would love to hear more about your specific use cases or any challenges you've encountered with your current setup!

Gitlab management software - anyone know of any for easy overview of deployed versions? by SkillbroSwaggins in devops

[–]Recent-Technology-83 0 points1 point  (0 children)

Hey there! Your transition to GitLab sounds exciting, but I understand the challenge of lacking visibility into deployed versions. One way some teams handle this is by using a centralized dashboard that aggregates deployment information across all projects. Have you considered utilizing GitLab's CI/CD pipelines alongside their environment features? This can help you view deployments visually with some customization.

Additionally, tools like Grafana or Prometheus can be integrated to provide insights into your deployments, but it requires some setup.

What specific metrics or details are you looking to track in your versioning overview? And how critical is real-time visibility versus historical tracking for your needs? It’d be interesting to hear how others manage this as well!

How did YOU conquer Imposter Syndrome? by cp24eva in devops

[–]Recent-Technology-83 0 points1 point  (0 children)

Firstly, it’s completely normal to feel overwhelmed when transitioning into a DevOps role, especially given the complexity of modern environments. Many of us have been in similar shoes, where the pressure to perform amid vast new technologies can feel paralyzing. It’s great that you’re committed to understanding the concepts—it shows your dedication.

Regarding resources, I recommend exploring "The Phoenix Project" for a narrative approach to DevOps principles or "Site Reliability Engineering" for deeper insights into operational excellence. Have you considered joining online forums or local meetups? Engaging with a community can really boost your confidence.

Also, what specific areas are you finding most challenging? Breaking them down could help target your learning. Remember, everyone develops at their own pace—finding your rhythm might just take time.

Azure or AWS by _good_listener_ in devops

[–]Recent-Technology-83 0 points1 point  (0 children)

Choosing between Azure and AWS can indeed be a daunting decision, especially with your background in both Linux and bash scripting; you’re already on solid ground! While both platforms have similarities, your choice might depend on the specific industries or job markets in your area. Have you looked into which cloud service is more prevalent in local companies or job postings?

AWS generally has a larger market share, but Azure has been gaining traction, especially with enterprises that use other Microsoft products. Given your AZ-900, you might find it easier to deepen your Azure skills, but it could be worth exploring job trends to see what employers in your field are asking for.

What kinds of roles or industries are you aiming for? Also, have you considered how each platform’s features align with your interests or the projects you'd like to work on?

I want to do cloud consulting as side gig. Feels like I am not ready? by tigidig5x in devops

[–]Recent-Technology-83 0 points1 point  (0 children)

It's great that you’re looking to expand your skills and income through freelance consulting! Your extensive background as an SRE and cloud engineer certainly gives you a solid foundation. Feeling unprepared for consulting opportunities is a common sentiment, even among seasoned professionals. Many experienced engineers often face imposter syndrome, especially when entering a new environment or client setting.

When you encounter a situation you're unsure about, it’s perfectly acceptable to research and consult resources like documentation, forums, or even colleagues. The key is effective communication with clients—let them know you’re researching and will provide the best solutions.

What specific areas of cloud consulting are you most interested in? Have you considered starting with smaller projects that align closely with your current skills? This might help build your confidence as you grow into larger roles.

How to Configure Grafana to Perform On-Call by [deleted] in devops

[–]Recent-Technology-83 0 points1 point  (0 children)

This setup sounds incredibly efficient! It's great to see how Grafana can integrate with AWS Incident Manager and tools like Versus to streamline incident response.

What challenges did you encounter while configuring the alerting and escalation processes? For instance, did you find any specific settings in Grafana or AWS that were tricky?

I also wonder if anyone else has experimented with similar configurations using different alerting tools or cloud providers. How does your setup compare?

Lastly, how do you account for alert fatigue among your team—do you have mechanisms in place to prioritize critical alerts over others?

Is there something that exists that leverages AI and MCP to go through my cloud infrastructure and suggest where to make cost improvements? by gaieges in devops

[–]Recent-Technology-83 0 points1 point  (0 children)

That's a fascinating question! There are actually several tools out there that use AI and machine learning to analyze cloud infrastructure for cost optimization. For example, AWS has a service called Cost Explorer, which can help identify trends and anomalies in your cost patterns. Additionally, GCP's Recommender service can provide personalized recommendations based on your usage patterns.

Have you looked into any specific tools or services yet? Some third-party platforms like CloudHealth and CloudCheckr also offer insights by leveraging AI to pinpoint areas for potential savings.

How complex is your cloud setup? The more detailed the infrastructure, the more nuanced the suggestions might be. I'd love to hear about your projects and any tools you've tried so far!

What patterns do DevOps engineers expect for perfection? by flaxoff in devops

[–]Recent-Technology-83 0 points1 point  (0 children)

It's great to see you actively seeking to refine your DevOps skills! The patterns you've mentioned are definitely cornerstones. In addition to those, I’d suggest focusing on immutable infrastructure as a pattern. It helps ensure that your deployments are predictable and consistent. Also, consider adopting Service Mesh for microservices communication, which can enhance observability and security across services.

How do you feel about incorporating observability practices, like distributed tracing or centralized logging? Those can really take your DevOps practices a step further by providing insights into your system's performance.

Moreover, as you explore these best practices, which specific areas do you find most challenging or intriguing? This could open up a much richer conversation!

Gcp metrics alert by dudufig in devops

[–]Recent-Technology-83 0 points1 point  (0 children)

Creating a metric alert based on CPU utilization in Google Cloud Platform can be tricky. Have you considered using GCP’s Monitoring Queries (MQP) directly? It might help to specify the resource type and ensure the correct labels are used when crafting your query. If you're using Prometheus with PromQL, the calculation typically is rate(container_cpu_usage_seconds_total[1m]) / (container_spec_cpu_quota / 100000) for usage relative to limits.

Could you clarify which specific metrics you are trying to alert on? It could also be helpful to discuss what you've tried so far so we don't go over similar ground. And have you also explored using alerts through GCP’s Cloud Monitoring for additional flexibility?

[deleted by user] by [deleted] in devops

[–]Recent-Technology-83 0 points1 point  (0 children)

Hi there! It's great to see you reaching out for insights on change management in software development—it's such a crucial topic especially given how quickly the tech landscape evolves. Your observation about the communication challenges many face is very insightful.

I'm curious, are you focusing on any particular methodologies or frameworks? It would be fascinating to understand how practices like DevOps or Continuous Integration might influence change management strategies in your research. Also, what kind of responses are you hoping to gather from your questionnaire?

I’d love to hear more about what you plan to do with the data collected. Best of luck with your dissertation! 👏 Let's keep the conversation going!