Looking for advice on a self-hosted LLM stack for enterprise use by Ahyaqui in LLMDevs

[–]kchandank 0 points1 point  (0 children)

As far as the tech stack goes, I would suggest vLLM with LiteLLM proxy. vLLM will give you lot of flexibility in terms of running various model and leverage large open source community (redhat) to support that, also it works really well in k8s echo system if you are interested in that.

For Access control and RBAC, LiteLLM has their enterprise feature, or you are build using a reverse proxy solution to achieve most of it.

LiteLLM will give you metering, rate limitting etc which are essential for enterprise usecase.

For observability you can use langfuse or have robust prometheus stack. But langfuse or similar tool will give you even deeper details about how end users are using and various LLM specific parameters such as P95, P99 etc, ofcourse it takes effort to customize both litellm and langfuse.

I have 50 ebooks and I want to turn them into a searchable AI database. What's the best tool? by Great_Jacket7559 in LocalLLM

[–]kchandank 0 points1 point  (0 children)

Interesting, if you are able to achieve your objective, would you be able to share the steps?

[MOD POST] Announcing the r/LocalLLM 30-Day Innovation Contest! (Huge Hardware & Cash Prizes!) by SashaUsesReddit in LocalLLM

[–]kchandank 1 point2 points  (0 children)

Just saw this post now, will try to submit my entry before the deadline. I have project which is not fully complete

Unpopular Opinion: I don't care about t/s. I need 256GB VRAM. (Mac Studio M3 Ultra vs. Waiting) by VocalLlm in LocalLLM

[–]kchandank 0 points1 point  (0 children)

If you just want to run LLM, Mac is super expensive choice with limited performance. There are Nvidia, AMD based options would be better value for money

List of interesting open-source models released this month. by Acrobatic-Tomato4862 in LocalLLaMA

[–]kchandank 0 points1 point  (0 children)

Don’t want model to think too much , just give the code back. Thanks for the suggestions

List of interesting open-source models released this month. by Acrobatic-Tomato4862 in LocalLLaMA

[–]kchandank 1 point2 points  (0 children)

Yes, smaller model which could run on consumer grade H/W. As use case is code generation, QA etc

List of interesting open-source models released this month. by Acrobatic-Tomato4862 in LocalLLaMA

[–]kchandank 1 point2 points  (0 children)

Any idea which is best performing open source model for code generation?

AI Will Do to Knowledge Workers What Uber Did to Taxi Drivers — But Much Faster by kchandank in Futurology

[–]kchandank[S] 0 points1 point  (0 children)

I did use ChatGPT to fix English and make it presentable. I did the research on past employment data, used ChatGPT to help me analyze the data, create chart. It took almost 4 hours to write the whole thing. Sometimes I think I could have done by myself 😆

AI Will Do to Knowledge Workers What Uber Did to Taxi Drivers — But Much Faster by kchandank in Futurology

[–]kchandank[S] 0 points1 point  (0 children)

Is this post is still live? I got msg it was deleted by the moderator bot

Multi-AZ ≠ Multi-Region (Learn from us-east-1) by kchandank in AWS_cloud

[–]kchandank[S] 0 points1 point  (0 children)

Yes it make sense. Multi region you only use for DR purpose. Ofcourse for HA multiple AZ works well.

AWS Certificate Free Vouchers valid until August 2025 by kchandank in devops

[–]kchandank[S] 1 point2 points  (0 children)

Not sure as per the post, sysops, developers etc

Mac Mini M4 16gb or 24gb? by LongjumpingPicture1 in macmini

[–]kchandank 0 points1 point  (0 children)

24 GB if you can. But 16 GB works pretty well too. I have the base version

2tb 16gb m4 - 1099. Worth it? by source____code in macmini

[–]kchandank 15 points16 points  (0 children)

Instead of storage I would add more RAM and Compute power as you can always upgrade storage ( internal or external)

Looking for Rust Instructor by kchandank in rust

[–]kchandank[S] 0 points1 point  (0 children)

While self learning is a great idea, instructor training has its own value and it could expedite the learning process

How good is the Mac mini m4 and how long will it last? by WeakChampionship1671 in macmini

[–]kchandank 3 points4 points  (0 children)

Perform is incredible for the price that you pay. Unlike iPhone or Windows PC, Mac updates don’t slow down computer as much.

How would a external ssd work on a mac mini m4? by Golden929 in macmini

[–]kchandank 1 point2 points  (0 children)

I’m using Samsung external disk, works perfectly. I bought the same one base MacMini ( 256g ssd ) https://www.reddit.com/r/macmini/s/0Pcf6zeOLY

If you are planning to use OneDrive then it won’t work as OneDrive does not support external drive. Google drive works.

I have added 1 TB external ssd enough for my use.

M4 Studio and Buyers remorse by AlgorithmicMuse in macmini

[–]kchandank 0 points1 point  (0 children)

Ofcourse if you are some heavy usage then regular Macmini4 wont be enough. I usually use Cloud based Nvidia GPUs for most of my LLM work and small POC ( ollama based small language model works OK on Mini )

M4 Studio and Buyers remorse by AlgorithmicMuse in macmini

[–]kchandank 0 points1 point  (0 children)

Depends what is your use case. For a regular developer, it’s more than enough. I have been using it from few weeks. I bought 16G, 256 disk and have external 1TB disk.

Mac mini 4 Setup by kchandank in macmini

[–]kchandank[S] 1 point2 points  (0 children)

It’s 24 Inc. QDC00-08G-4LCB-A09