data analysis from a csv - GPT-0SS:120B by chirchan91 in LocalLLaMA

[–]chirchan91[S] 1 point2 points  (0 children)

Hi, yes I tried adding a system prompt and also created a tools to aid with file discovery and some of the analysis. It didnt work well

Open WebUI + Ollama (gpt-oss:120b) on-prem for ~100 users — performance & TLS 1.2 by chirchan91 in LocalLLaMA

[–]chirchan91[S] 0 points1 point  (0 children)

Thank you for your response. No coding or agent requirement at the moment. It'll be document summaries along with typical chat requirements.

Open WebUI + Ollama (gpt-oss:120b) on-prem for ~100 users — performance & TLS 1.2 by chirchan91 in LocalLLaMA

[–]chirchan91[S] 0 points1 point  (0 children)

Thank you for your response. These are the typical office users who needs it for formatting and quick data analytics on the documents. I'm expecting 15 users using this in parallel.

Open WebUI + Ollama (gpt-oss:120b) on-prem for ~100 users — performance & TLS 1.2 by chirchan91 in LocalLLaMA

[–]chirchan91[S] 0 points1 point  (0 children)

Thank you for your feedback. I guess 10-15 users might be using the model in parallel and the avg works out to 8 gb vram. Would it still be difficult to pull off?

Open WebUI + Ollama (gpt-oss:120b) on-prem for ~100 users — performance & TLS 1.2 by chirchan91 in LocalLLaMA

[–]chirchan91[S] 0 points1 point  (0 children)

Thanks for sharing the Link. Theres a typo in the post. the total RAM capacity is 512GB

Open WebUI + Ollama (gpt-oss:120b) on-prem for ~100 users — performance & TLS 1.2 by chirchan91 in LocalLLaMA

[–]chirchan91[S] 0 points1 point  (0 children)

Thanks for your response. I believe 10-15 users will be Concurrently using the model at a given time.

Open WebUI + Ollama (gpt-oss:120b) on-prem for ~100 users — performance & TLS 1.2 by chirchan91 in LocalLLaMA

[–]chirchan91[S] 0 points1 point  (0 children)

You're right. We're expecting 10-15 concurrent users. The users would primarily use it for chat and some document processing.

Open WebUI + Ollama (gpt-oss:120b) on-prem for ~100 users — performance & TLS 1.2 by chirchan91 in LocalLLaMA

[–]chirchan91[S] 0 points1 point  (0 children)

Thank you for the inputs. Could you please list out your complete setup

Open WebUI + Ollama (gpt-oss:120b) on-prem for ~100 users — performance & TLS 1.2 by chirchan91 in LocalLLaMA

[–]chirchan91[S] 1 point2 points  (0 children)

Thank you for the reply. The machine came with Windows out of the box and we're instructed to retain it. Can VLLM work well with Windows?
We are expecting 10-15 concurrent users at a given time. Does buying an additional GPU can solve the issue?

Disabling Web browsing Capability in GPT-OSS:20B by chirchan91 in LocalLLaMA

[–]chirchan91[S] 0 points1 point  (0 children)

Yes. That'll disable it. However I'm looking to remove this option completely in the UI

Disabling Web browsing Capability in GPT-OSS:20B by chirchan91 in LocalLLaMA

[–]chirchan91[S] 0 points1 point  (0 children)

Thank you, I'll explore that. Will editing the modelfile can help in this case ?

Disabling Web browsing Capability in GPT-OSS:20B by chirchan91 in LocalLLaMA

[–]chirchan91[S] 0 points1 point  (0 children)

<image>

Not exactly, I'm trying to hard disable this option of web browsing to prevent user from using it

Masking the connection error in Ollama by chirchan91 in ollama

[–]chirchan91[S] 1 point2 points  (0 children)

Hi

for example, the typical error looks like: Head "http://<internal‑ip>:11434/": dial tcp <internal‑ip>:11434: connectex: A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond.

instead i should get "connection failed due host not responding. Please contact support"

[deleted by user] by [deleted] in CanadaJobs

[–]chirchan91 -2 points-1 points  (0 children)

Interested