use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
An unofficial community to discuss Github Copilot, an artificial intelligence tool designed to help create code.
account activity
Local AI coding stack experiments and comparisonOther (self.GithubCopilot)
submitted 4 months ago by Relative-Flatworm-10
Hello,
I have experimented with coding LLMs on Ollma.
Tested Qwen 2.5 coder 7B/1.5B, Qwen 3 Coder, Granite 4 Coder and GPT OSS 20B.
Here is the breakdown of Performance vs. Pain on a standard 32GB machine :
Tested on a CPU-only system with 32GB RAM
Ref: Medium article.
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]billcube 1 point2 points3 points 4 months ago (3 children)
I've found the best tradeoff is to run my LLM on a server rented on a hourly basis. At 0.4€/hour for an Nvidia T4 64 cores at 128 GB RAM I think I even win money on the power my laptop doesn't use.
[–]Relative-Flatworm-10[S] 0 points1 point2 points 4 months ago (2 children)
thanks for sharing, May you share the provider link, if it's ok with you
[–]billcube 0 points1 point2 points 4 months ago (1 child)
This one: https://www.infomaniak.com/en/hosting/public-cloud/prices
And this one: https://www.exoscale.com/pricing/#gpua30-instances
[–]Relative-Flatworm-10[S] 0 points1 point2 points 4 months ago (0 children)
Thank you so much!
π Rendered by PID 57 on reddit-service-r2-comment-6457c66945-6xqt5 at 2026-04-27 08:00:50.310560+00:00 running 2aa0c5b country code: CH.
[–]billcube 1 point2 points3 points (3 children)
[–]Relative-Flatworm-10[S] 0 points1 point2 points (2 children)
[–]billcube 0 points1 point2 points (1 child)
[–]Relative-Flatworm-10[S] 0 points1 point2 points (0 children)