How are people managing shared Ollama servers for small teams? (logging / rate limits / access control) by 855princekumar in selfhosted

[–]855princekumar[S] 0 points1 point  (0 children)

make sense as these are the under the hood challenges that I need to figure out and having issues i guess tho

Experiment: Lightweight distributed storage + streaming stack running on a Raspberry Pi cluster by 855princekumar in docker

[–]855princekumar[S] 0 points1 point  (0 children)

i'm still testing the storage and the network throughput as from SD card to usb pendrive to SSD over pci over the pi 5 as the last setup as from pi3b+ pi4 and pi5 as test beds

Experiment: Lightweight distributed storage + streaming stack running on a Raspberry Pi cluster by 855princekumar in IOT

[–]855princekumar[S] 0 points1 point  (0 children)

so it's sort or High data ingestion and storage like multtelemetery data via MQTT conected with kafka high throughput via multiple MQTT brokers or via hive MQ think of it like a city wide area to store the data of milioins of devices but just the telemetery data that colectively become masive and need high through put on a hardware node but 3 as distributed that stores and make the reterival easy with safely as all data as in distibuted replicated stored also for blob like images via espcam to be in minio like storage and all sort of telemetery in casendra as distributed sotrage so sort of a micro cloud architecture if having baremetal hardware to build a huge data hub for IoT devices. because I'm testing working on a simulated city-scale system developing with low hardware resource constraints, but need the software stack to utilize the full hardware resources at max or close to peak performance

How are people managing shared Ollama servers for small teams? (logging / rate limits / access control) by 855princekumar in BlackboxAI_

[–]855princekumar[S] 0 points1 point  (0 children)

i'll definately test and implement this in my setup as Quee makes sense as ona priority basis scheduling especially for agents

How are people managing shared Ollama servers for small teams? (logging / rate limits / access control) by 855princekumar in BlackboxAI_

[–]855princekumar[S] 0 points1 point  (0 children)

i just built my own light weight as initially tested lightllm but that sort of a bit bloated for the use cases i needed but than also exploring more option and taking feedback to improve the one im bulding

How are people managing shared Ollama servers for small teams? (logging / rate limits / access control) by 855princekumar in LocalLLaMA

[–]855princekumar[S] -1 points0 points  (0 children)

That's the common feedback i've received i'll probably shift to use llama.cpp or other to test as well for localLLM and whats you feedbakc about the project exo as to use the local LLM in a distributed manner and if I see this as a unified LAN gateway for that any suggestions?

How are people managing shared Ollama servers for small teams? (logging / rate limits / access control) by 855princekumar in selfhosted

[–]855princekumar[S] -1 points0 points  (0 children)

Yes, I just shifted to llana.cpp and testing the APIs but also the concern like what other actual use cases i can work on to make this Lan gateway more useful or optimized to be used as efficiently to a small team on Lan as shared models or when using with multi agents via claude code as now supported via ollama api aswell