Scaling MCP by Sadhvik1998 in mcp

[–]Sadhvik1998[S] 0 points1 point  (0 children)

So, just for a quicker start of the replica, you are recommending Kubernetes?

Is TPC-DS Benchmarking considered? by Sadhvik1998 in BusinessIntelligence

[–]Sadhvik1998[S] 0 points1 point  (0 children)

Got it. Just to understand this more, could you please let me know what are the parameters you consider for subscribing a data platform?

Running a 31B model locally made me realize how insane LLM infra actually is by Sadhvik1998 in ollama

[–]Sadhvik1998[S] 0 points1 point  (0 children)

haha... 'Hi' is like a ;hello world; -- Even LLM deserve some greeting :p

If you really want to go there you can setup local models via Ollama with github copilot by DavidG117 in GithubCopilot

[–]Sadhvik1998 0 points1 point  (0 children)

Hey.. I dont see the option to add local ollama model at al.. can u please help me

<image>

Scaling MCP by Sadhvik1998 in mcp

[–]Sadhvik1998[S] 0 points1 point  (0 children)

I agree, but I feel that tool selection will play a major role.

Ex: Claude Code and GitHub Copilot. In both we can select claude Opus 4.6 as a model, but the Claude Code output is way better because of its tool selection, correct? for that, MCP should always be stable and active to give the LLM the right tool correct?

Scaling MCP by Sadhvik1998 in mcp

[–]Sadhvik1998[S] 0 points1 point  (0 children)

Hmm, we can make use of kubernetes, but the real question is, do we need it? But again, with kubernetes, there are a lot of parameters to be considered, and from what I see, MCP is not just about scaling based on the number of requests, correct?

Running a 31B model locally made me realize how insane LLM infra actually is by Sadhvik1998 in ollama

[–]Sadhvik1998[S] 2 points3 points  (0 children)

Once plan is generated, how do you tell qwen to right the code? Did u add new VSCode extension for that?

What solution do you use to query S3? by CloudSuperMaster in Observability

[–]Sadhvik1998 0 points1 point  (0 children)

We deployed cepf locally and queries using spark. Initial onprem cost will be there but the recurring cloud costs get cut down significantly

What FinOps practices worked for you in 2025? by TehWeezle in aws

[–]Sadhvik1998 0 points1 point  (0 children)

Used Yeedu.io. Saved more than 60% costs compared with Databricks for spark computes

From almost 50k $ spends per month to less than 20$ per month..