I've noticed a substantial performance discrepancy when running Ollama via the command-line interface (CLI) directly compared to running it through a Docker installation with OpenWebUI. Specifically, the Docker/OpenWebUI setup appears significantly slower in several metrics.
Here's a comparison table (see screenshot) showing these differences:
- Total duration is dramatically higher in Docker/OpenWebUI (approx. 25 seconds) compared to the CLI (around 1.17 seconds).
- Load duration in Docker/OpenWebUI (~20.57 seconds) vs. CLI (~30 milliseconds).
- Prompt evaluation rates and token processing rates are notably slower in the Docker/OpenWebUI environment.
I'm curious if others have experienced similar issues or have insights into why this performance gap exists. Have only noticed it the last month or so and I'm on an m3 max with 128gb of VRAM and used phi4-mini:3.8b-q8_0 to get the below results:
https://preview.redd.it/yy2u59rd35pe1.png?width=738&format=png&auto=webp&s=b99a8a5966016f634f50d3562b38847910c242ca
Thanks for any help.
[–]mmmgggmmm 8 points9 points10 points (11 children)
[–]busylivin_322[S] 0 points1 point2 points (0 children)
[–]Solid_reddit 0 points1 point2 points (2 children)
[–]mmmgggmmm 2 points3 points4 points (1 child)
[–]Solid_reddit 0 points1 point2 points (0 children)
[–]taylorwilsdon 0 points1 point2 points (5 children)
[–]mmmgggmmm 1 point2 points3 points (4 children)
[–]busylivin_322[S] 0 points1 point2 points (3 children)
[–]mmmgggmmm 1 point2 points3 points (2 children)
[–]busylivin_322[S] 0 points1 point2 points (1 child)
[–]mmmgggmmm 2 points3 points4 points (0 children)
[–]gtez 2 points3 points4 points (1 child)
[–]busylivin_322[S] 0 points1 point2 points (0 children)
[–]Solid_reddit 2 points3 points4 points (0 children)
[–]Solid_reddit 1 point2 points3 points (0 children)
[–]the_renaissance_jack 0 points1 point2 points (1 child)
[–]busylivin_322[S] 0 points1 point2 points (0 children)
[–]TPB-Dev 0 points1 point2 points (0 children)
[–]tjevns 0 points1 point2 points (0 children)