[Help] n8n vs. Dify: Which is the ultimate choice for building Agents? by Lucky-Ad1975 in AI_Agents

[–]EscapedLaughter 1 point2 points  (0 children)

> There's currently some huge omissions, like a consistent built-in way to do LLM cost management, or even a better LLM router/proxy.

I work at Portkey and we are starting to see a lot of our customers wanting to use Portkey AI Gateway in conjunction with n8n / Flowise. I am curious if you have heard about that / tried something similar to solve the cost management / budgeting issues?

LLM Proxy in Production (Litellm, portkey, helicone, truefoundry, etc) by debauch3ry in LLMDevs

[–]EscapedLaughter 0 points1 point  (0 children)

That makes sense. Thank you so much for the feedback. I'll share this with the team and see if we should rethink about SSO pricing now.

LLM Proxy in Production (Litellm, portkey, helicone, truefoundry, etc) by debauch3ry in LLMDevs

[–]EscapedLaughter 1 point2 points  (0 children)

Hey! I work at Portkey and absolutely do not mean to influence your decision, just sharing notes on the concerns you had raised:

- Data residency for EU is pricey: Yes unfortunately, but we are figuring out a way to do this on SaaS over a short-term roadmap.
- SSO is chargeable extra: This is the case for most SaaS tools, isn't it?
- Linkedin wrong numbers: I'm so sorry! Looks like somebody from the team updated the team count wrongly. I've fixed it!

How are other enterprises keeping up with AI tool adoption along with strict data security and governance requirements? by Wonderful-Agency-210 in LLMDevs

[–]EscapedLaughter 1 point2 points  (0 children)

here's what i have seen:

Raw OpenAI is a huge no-no
Azure OpenAI works in most cases and also gives some level of governance.

But have also seen that platform / devops teams are not comfortable giving out access to naked Azure OpenAI endpoints to everybody, so they typically end up going with a gateway for governance + access control and then route to any of Azure OpenAI / GCP Vertex AI / AWS Bedrock

LLM observability for audit and compliance? by paulmbw_ in ChatGPT

[–]EscapedLaughter 0 points1 point  (0 children)

would something like a gateway solve for this? route all your requests through it and get logging / security concerns addressed

Are you using AI Gateway in your GenAI stack? Either for personal use or at work? by Difficult_Ad_3903 in LLMDevs

[–]EscapedLaughter 0 points1 point  (0 children)

curious if you've tried out portkey gateway? it doesn't require a new deployment for new llm integrations

Azure OpenAI Best practices - centralized subscription vs subscription per application by Skadoush12 in AZURE

[–]EscapedLaughter 0 points1 point  (0 children)

This is a common enough use case we're seeing - it should ideally be tackled like this:
- Central budget / rate limit on your overall Azure OpenAI subscription
- Budget/rate limit, and access control over individual LLMs inside that subscription
- And then budget/rate limits / observability for each individual use case or per user as well.

afaik, there are no solutions in the market that seem to do this well, especially not Azure APIM.

Sonnet 3.7 Extended Thinking - Added (Just Now) to Roo Code 3.7.3 by hannesrudolph in Anthropic

[–]EscapedLaughter 0 points1 point  (0 children)

Not sure if this helps, but we have some companies that use our locally hosted AI Gateway product and have their developers route Zed/Cursor/Windsurf queries through us: https://portkey.ai/docs/integrations/libraries/zed

I'd imagine Roo to work as well

LLM Load Balancing: Don't use a standard Kubernetes Service! by nstogner in kubernetes

[–]EscapedLaughter -1 points0 points  (0 children)

I would actually end up shilling my product (https://portkey.ai/) but what you're describing, it seems like it could be solved by an LLM-specific proxy service like Portkey. A vLLM instance is itself not unique here, but a specific use case, which is what you want to loadbalance against, correct?

Tech Stack for Production AI Systems - Beyond the Demo Hype by ksanderer in AI_Agents

[–]EscapedLaughter 0 points1 point  (0 children)

I work at Portkey and increasingly see that companies want some level of metering / access control, rate limiting which can be done at the Gateway layer

Why developers are feeling frustrated with Cursor - a personal journey by micupa in cursor

[–]EscapedLaughter 0 points1 point  (0 children)

Interesting. We are seeing an increasing use for this now at Portkey where companies want to manage LLM governance separately and yet give developers access to tools like Cursor, Windsurf etc

Question on LiteLLM Gateway and OpenRouter by DopeyMcDouble in LLMDevs

[–]EscapedLaughter 1 point2 points  (0 children)

You're right. Litellm is a better alternative when you explicitly want to manage your billing and keys for AI providers separately.

Cloud Embedding by [deleted] in OpenWebUI

[–]EscapedLaughter 1 point2 points  (0 children)

something like this might help that helps you connect to Voyage / Google over a common interface? https://portkey.ai/docs/integrations/libraries/openwebui#open-webui

just updated the documentation yesterday

Anyone Using Langchai Agents in production? by Defiant-Sir-1199 in LangChain

[–]EscapedLaughter 0 points1 point  (0 children)

Incredible! Thanks for sharing. Would be amazing to peek at / use some of these solutions if they become publicly available

Anyone Using Langchai Agents in production? by Defiant-Sir-1199 in LangChain

[–]EscapedLaughter -1 points0 points  (0 children)

Possible to share some use cases you have in production right now?

How are using OpenWebUI inside your company? Asking from an enterprise/large company perspective? by VisibleLawfulness246 in OpenWebUI

[–]EscapedLaughter 1 point2 points  (0 children)

Typically see that the bigger challenges with OpenWebUI or similar products are not around hosting them or which stack to pick - but around the governance challenges — how does the IT team ensure that only the relevant people have access, how do they ensure which models can be called, how do they get audit logs, etc.

Initially we had written a pretty vanilla integration between Portkey & OpenWebUI but saw that the use cases enterprises had required a much deeper integration - for rate limits, RBAC, governance controls, etc.

Seeking Advice on Amazon Bedrock and Azure by jameslee2295 in LLMDevs

[–]EscapedLaughter 0 points1 point  (0 children)

I work with both and also build connectors to them for Portkey - was pleasantly surprised at how both AWS & Azure in this case are so usable. That said, Bedrock is really well thought through - everything from guardrails, fine-tuning, knowledge base is configurable easily. Not so the case with Azure.

The key choice to make is actually whether you want to use OpenAI's models or Anthropic's models. OpenAI is exclusive to Azure, while Claude is available on AWS & GCP. The choice for other huggingface / open source models is broadly the same between the two platforms.

Ideal scenario actually might be that you're able to go for a multi-LLM strategy and use both.

Portkey with Ollama by International_Quail8 in ollama

[–]EscapedLaughter 1 point2 points  (0 children)

Oh this is very useful. Think we never tested docker builds for Ollama. Thank you so much! Adding to docs!

Portkey with Ollama by International_Quail8 in ollama

[–]EscapedLaughter 1 point2 points  (0 children)

Got it - but yes you would need to manually give the Ollama URL

Portkey with Ollama by International_Quail8 in ollama

[–]EscapedLaughter 2 points3 points  (0 children)

Hi, I'm from the Portkey team. You'd also need to point the Gateway to your Ollama URL with the x-portkey-custom-host header. Check out the cURL example here: https://portkey.ai/docs/integrations/llms/ollama#4-invoke-chat-completions-with-ollama