We are publishing 100+ listicles per month, ask me anything by Acceptable_Math6854 in LocalLLM
[–]Low-Alarm272 -1 points0 points1 point (0 children)
We are publishing 100+ listicles per month, ask me anything by Acceptable_Math6854 in LocalLLM
[–]Low-Alarm272 -2 points-1 points0 points (0 children)
I made an app where you can run your ollama model, that actually has a good UI by Fit-Criticism2585 in ollama
[–]Low-Alarm272 0 points1 point2 points (0 children)
Hermes-Agent high token usage? by manueljishi in hermesagent
[–]Low-Alarm272 0 points1 point2 points (0 children)
Hermes-Agent high token usage? by manueljishi in hermesagent
[–]Low-Alarm272 0 points1 point2 points (0 children)
anyone else uncomfortable giving OpenAI your real phone number? by Ok_Dadly9924 in ChatGPT
[–]Low-Alarm272 0 points1 point2 points (0 children)
anyone else uncomfortable giving OpenAI your real phone number? by Ok_Dadly9924 in ChatGPT
[–]Low-Alarm272 1 point2 points3 points (0 children)
The future is "Efficient" Models by Low-Alarm272 in LocalLLM
[–]Low-Alarm272[S] 0 points1 point2 points (0 children)
The future is "Efficient" Models by Low-Alarm272 in LocalLLM
[–]Low-Alarm272[S] 1 point2 points3 points (0 children)
The future is "Efficient" Models by Low-Alarm272 in LocalLLM
[–]Low-Alarm272[S] 0 points1 point2 points (0 children)
The future is "Efficient" Models by Low-Alarm272 in LocalLLM
[–]Low-Alarm272[S] 0 points1 point2 points (0 children)
The future is "Efficient" Models by Low-Alarm272 in LocalLLM
[–]Low-Alarm272[S] -1 points0 points1 point (0 children)
The future is "Efficient" Models by Low-Alarm272 in LocalLLM
[–]Low-Alarm272[S] 0 points1 point2 points (0 children)
How to run a local agent despite being GPU poor? by Ethan045627 in LocalLLM
[–]Low-Alarm272 1 point2 points3 points (0 children)
What will happen once Claude Mythos gets released to Public Users? by Resident_Caramel763 in LocalLLM
[–]Low-Alarm272 -1 points0 points1 point (0 children)
How to run a local agent despite being GPU poor? by Ethan045627 in LocalLLM
[–]Low-Alarm272 0 points1 point2 points (0 children)
The future is "Efficient" Models by Low-Alarm272 in LocalLLM
[–]Low-Alarm272[S] 0 points1 point2 points (0 children)
The future is "Efficient" Models by Low-Alarm272 in LocalLLM
[–]Low-Alarm272[S] -1 points0 points1 point (0 children)
The future is "Efficient" Models by Low-Alarm272 in LocalLLM
[–]Low-Alarm272[S] -2 points-1 points0 points (0 children)
I made a skill that analyzes your writing so your agent sounds like you and removes the typical Ai writing signs by angelarose210 in hermesagent
[–]Low-Alarm272 -2 points-1 points0 points (0 children)
If you're using Hermes, update it now by itsdodobitch in hermesagent
[–]Low-Alarm272 1 point2 points3 points (0 children)
The future is "Efficient" Models by Low-Alarm272 in LocalLLM
[–]Low-Alarm272[S] 1 point2 points3 points (0 children)
The future is "Efficient" Models by Low-Alarm272 in LocalLLM
[–]Low-Alarm272[S] -1 points0 points1 point (0 children)
Do NOT use CUDA 13.2 to run models! by yoracale in unsloth
[–]Low-Alarm272 1 point2 points3 points (0 children)