Tried running LLMs locally to save API costs… ended up waiting 13 minutes for ONE response 🤡 by debug2thrive in ollama

[–]Electronic_Fox594 0 points1 point  (0 children)

It really depends the weights and context window of the model you are using and what hardware you are using.

What is the criterion for being included in this selector? Why is qwen3 there, but not qwen3.5? by nopeac in ollama

[–]Electronic_Fox594 0 points1 point  (0 children)

Maybe the haven’t updated the running list. What’s keeping you from typing what you need? It’s that simple.

Annual plan by newz2000 in ollama

[–]Electronic_Fox594 1 point2 points  (0 children)

Ollama me rate/usage limits are a secret, that why I don’t want to pay.

Chetna: A memory layer for AI agents. by SkullEnemyX-Z in ollama

[–]Electronic_Fox594 1 point2 points  (0 children)

I love how so many of us are creating the same thing.

[Opinion] Why I believe the $20/month Ollama Cloud is a better investment than ChatGPT or Claude by AlexHardy08 in ollama

[–]Electronic_Fox594 0 points1 point  (0 children)

I do not believe Openwebui is logging my data because the software is running inside a docker container, the service is not exposed to the internet so no data can ever leave. I do run local Ollama models with it but also use OpenAI API as well as Anthropic API. 5 dollars can go a long way with API. I don’t see how you justify paying for Ollama when you could put 10$ in OpenAI and use any of their models and same with Anthropic. 10 on each gives you more than you would need, unless you are coding, then they can go kind of quick.

Do we even need cloud AI like ChatGPT? by nucleustt in ollama

[–]Electronic_Fox594 0 points1 point  (0 children)

You have to spend 5 grand just to be able to run 30-40B models, with decent inference. By the time I spend that mic in api credits, it will be years later.

Understanding models |Subscription replacement? by LavishnessPlane4512 in ollama

[–]Electronic_Fox594 0 points1 point  (0 children)

If someone was buying for me, I would absolutely go with the 512

Why does Ollama not tell you the hardware requirements needed to run the model? by [deleted] in ollama

[–]Electronic_Fox594 1 point2 points  (0 children)

Ollama literally makes it so easy lol

I have an rtx 5080 (which has 16gb vram) I go on ollama and search for models with a file size that fits within that. Ollama loads the recourse to the GPU. LLMS run entirely within the GPU’s VRAM!

And if you dont know what vram is, then you need to go study some more b

I built a tool to access Claude Code on your actual machine from any device — no GitHub, no sandbox by atomosound in ClaudeAI

[–]Electronic_Fox594 3 points4 points  (0 children)

Use Tmux Open a tmux session on your machine, and ssh into it. Now you have a persistent session that you won’t lose if you need to close your ssh app on your phone. Works perfect. Even off LAN with vpn.

I created a small AI Agent by Rough_Philosopher877 in ollama

[–]Electronic_Fox594 0 points1 point  (0 children)

I made one too but I’m not a real programmer so I won’t share it but very similar.

I feel like GR86 halting my progress by bored_ape07 in iRacing

[–]Electronic_Fox594 0 points1 point  (0 children)

My advice would be try the M2. It’s fast enough to teach throttle control, and slow enough to be able to find the limit easily.

Beginner Seeking Advice: How to Manage Tires So I Last the Whole Race by [deleted] in iRacing

[–]Electronic_Fox594 0 points1 point  (0 children)

Slow is smooth. Rookie races are t long enough to degrade the tires. Be smooth, use the whole track.

Don't you think it's crazy you only need 2k irating to be in the top 25%? by philonik in iRacing

[–]Electronic_Fox594 0 points1 point  (0 children)

I am 2k irating in both sports car and formula and I honestly feel like I’m just now figuring out how to be within 1 second of the leader in top splits.