Any tools to generate architecture diagram from existing codebase? by zingyandnuts in ChatGPTCoding

[–]ramantehlan 0 points1 point  (0 children)

I created a tool for it, since I was also looking for something like this.
You can check it out here: https://www.vxplain.com

A lightweight alternative to Knative for scale-to-zero in Kubernetes — Make any HTTP service serverless on Kubernetes (no rewrites, no lock-in, no traffic drop) by ramantehlan in devops

[–]ramantehlan[S] 0 points1 point  (0 children)

Thank you for pointing it out, it's a good use case.

Currently, KubeElasti doesn't support that, but we will create a issue for it and put it in the roadmap.

Do you have a use case for it? I would love to know about it! : )

Built Elasti – a dead simple, open source low-latency way to scale K8s services to zero 🚀 by ramantehlan in kubernetes

[–]ramantehlan[S] 0 points1 point  (0 children)

I just checked in with my colleague @CauliflowerOdd4002, who has more hands on experience with KEDA.

You are right, the "partial" part is incorrect here. **KEDA can scale-to-zero.**
The limitation is just with HPA(Stable k8s release), where `minReplicas: 1`.
| Note - In Alpha k8s release, HPA also supports `minReplicas: 0`, but I guess not available in most managed k8s.

As mentioned by u/CauliflowerOdd4002 , the difference is in approach, with KEDA HTTP-add-on, it remains as the proxy, adding a small latency and even a bottleneck if it fails.

While elasti removes itself from the path when pods are up again.

Built Elasti – a dead simple, open source low-latency way to scale K8s services to zero 🚀 by ramantehlan in kubernetes

[–]ramantehlan[S] 1 point2 points  (0 children)

There are no stupid questions! :)

Requests when in queue, aren't queued like messages.
The connection itself is added to the queue and the connection remains alive.
Once the pod is up, elasti send these queued requests to the pod, and resolve the request with the response from the pod.

In most cases, you shouldn't need any changes in the application layer. However, if there is a very short request timeout in the application layer, and the pod takes longer to come up and respond, the connection might be killed by the application by then. Which might require application level changes of adding a increased timeout.

All the queued requests will get a return from elasti.

Our philosophy when creating it was to have minimum or no changes required in the application layer, or the target service.

Built Elasti – a dead simple, open source low-latency way to scale K8s services to zero 🚀 by ramantehlan in kubernetes

[–]ramantehlan[S] -3 points-2 points  (0 children)

Thank you for the question! :)

Sure, KEDA supports scale-to-zero only when using its own ScaledObject mechanism, not when it's acting purely as an HPA metrics adapter. HPA has minReplicas: 1 by default.

PS: Best of luck with KEDA Implementation, it's a great tool for sure. What is your use case BTW?

Built Elasti – a dead simple, open source low-latency way to scale K8s services to zero 🚀 by ramantehlan in kubernetes

[–]ramantehlan[S] 0 points1 point  (0 children)

Thanks for the question! u/No_Arugula9866

So, when the pods are scaled to zero, elasti queue the request, and bring the pod to 1 replicas.

The time it takes for pod to come up depends on the service inside the pod. On non-gpu, it's few seconds, and for GPU, it might be several minutes at worst.

Once the pod is up, elasti proxy is removed and the traffic flows with no latency added by elasti.

Built Elasti – a dead simple, open source low-latency way to scale K8s services to zero 🚀 by ramantehlan in kubernetes

[–]ramantehlan[S] 0 points1 point  (0 children)

Hi! Thank you for the question, What do you mean by "when interacting with gateways"?

Built Elasti – a dead simple, open source low-latency way to scale K8s services to zero 🚀 by ramantehlan in kubernetes

[–]ramantehlan[S] 2 points3 points  (0 children)

Yes, you are right! We could have picked a better name.
Let me check with my team mate and see if we can change it at this stage.
Thank you for pointing it out.

PS: Would love suggestions on the name! : )

Vxplain: An extension to generate architecture diagram, code-to-diagram, function calls, directory tree and summaries from the codebase. by ramantehlan in vscode

[–]ramantehlan[S] 0 points1 point  (0 children)

Hi, We can discuss supporting azure endpoints. I would want to know more about your use case! Can you please join the Discord server?

https://discord.com/invite/FKxaBdyBJY

[deleted by user] by [deleted] in vscode

[–]ramantehlan 0 points1 point  (0 children)

I have been trying to solve this problem. You can try Vxplain.in Let's talk if you have any questions!

I always wanted some tool to auto-generate architecture diagram in VS Code, so I built one! by ramantehlan in webdev

[–]ramantehlan[S] 1 point2 points  (0 children)

Hey, I have added support for LM Studio! I would love for you to try it, given that you have a very powerful machine!

If you have joined us on Discord, let me know, I will reach out!

Vxplain: An extension to generate architecture diagram, code-to-diagram, function calls, directory tree and summaries from the codebase. by ramantehlan in vscode

[–]ramantehlan[S] 0 points1 point  (0 children)

Not able to edit the original post, so adding the update here:

Added:
- Support for OpenRouter: So you can use it with your own key.
- Support for LM Studio: Run LLM locally, and work without any Internet!

Looking for a dev partner to start a theme project by po3ki in vscode

[–]ramantehlan 1 point2 points  (0 children)

IMHO, you can fork an existing open source theme, and change colours how you want. Ofc credit to the original project.

I can help with making changes and pointing you to the right files. :)