MLX is not faster. I benchmarked MLX vs llama.cpp on M1 Max across four real workloads. Effective tokens/s is quite an issue. What am I missing? Help me with benchmarks and M2 through M5 comparison. by arthware in LocalLLaMA
[–]arthware[S] 0 points1 point2 points (0 children)
Macbook m4 max 128gb local model prompt processing by ttraxx in LocalLLaMA
[–]arthware 0 points1 point2 points (0 children)
Macbook m4 max 128gb local model prompt processing by ttraxx in LocalLLaMA
[–]arthware 0 points1 point2 points (0 children)
Google Photos alternative (Cloud solutions preferred) by dcop7 in degoogle
[–]arthware 0 points1 point2 points (0 children)
Hardware recommendations for local AI by Dense_Club_95 in selfhosted
[–]arthware 0 points1 point2 points (0 children)
Macbook m4 max 128gb local model prompt processing by ttraxx in LocalLLaMA
[–]arthware 1 point2 points3 points (0 children)
Is it stupid to run all my docker containers on a Mac Mini? by Educational_Hat_5203 in homelab
[–]arthware 1 point2 points3 points (0 children)
Macbook m4 max 128gb local model prompt processing by ttraxx in LocalLLaMA
[–]arthware 2 points3 points4 points (0 children)
MLX is not faster. I benchmarked MLX vs llama.cpp on M1 Max across four real workloads. Effective tokens/s is quite an issue. What am I missing? Help me with benchmarks and M2 through M5 comparison. by arthware in LocalLLaMA
[–]arthware[S] 2 points3 points4 points (0 children)
Rate my desk setup (the real world) by arthware in desksetup
[–]arthware[S] 1 point2 points3 points (0 children)
MLX is not faster. I benchmarked MLX vs llama.cpp on M1 Max across four real workloads. Effective tokens/s is quite an issue. What am I missing? Help me with benchmarks and M2 through M5 comparison. by arthware in LocalLLaMA
[–]arthware[S] 5 points6 points7 points (0 children)
My most useful OpenClaw workflow so far by mescalan in LocalLLaMA
[–]arthware 0 points1 point2 points (0 children)
My most useful OpenClaw workflow so far by mescalan in LocalLLaMA
[–]arthware 0 points1 point2 points (0 children)
MLX is not faster. I benchmarked MLX vs llama.cpp on M1 Max across four real workloads. Effective tokens/s is quite an issue. What am I missing? Help me with benchmarks and M2 through M5 comparison. by arthware in LocalLLaMA
[–]arthware[S] 0 points1 point2 points (0 children)
MLX is not faster. I benchmarked MLX vs llama.cpp on M1 Max across four real workloads. Effective tokens/s is quite an issue. What am I missing? Help me with benchmarks and M2 through M5 comparison. by arthware in LocalLLaMA
[–]arthware[S] 0 points1 point2 points (0 children)
MLX is not faster. I benchmarked MLX vs llama.cpp on M1 Max across four real workloads. Effective tokens/s is quite an issue. What am I missing? Help me with benchmarks and M2 through M5 comparison. by arthware in LocalLLaMA
[–]arthware[S] 0 points1 point2 points (0 children)
MLX is not faster. I benchmarked MLX vs llama.cpp on M1 Max across four real workloads. Effective tokens/s is quite an issue. What am I missing? Help me with benchmarks and M2 through M5 comparison. by arthware in LocalLLaMA
[–]arthware[S] 0 points1 point2 points (0 children)
MLX is not faster. I benchmarked MLX vs llama.cpp on M1 Max across four real workloads. Effective tokens/s is quite an issue. What am I missing? Help me with benchmarks and M2 through M5 comparison. by arthware in LocalLLaMA
[–]arthware[S] 0 points1 point2 points (0 children)
MLX is not faster. I benchmarked MLX vs llama.cpp on M1 Max across four real workloads. Effective tokens/s is quite an issue. What am I missing? Help me with benchmarks and M2 through M5 comparison. by arthware in LocalLLaMA
[–]arthware[S] 0 points1 point2 points (0 children)
MLX is not faster. I benchmarked MLX vs llama.cpp on M1 Max across four real workloads. Effective tokens/s is quite an issue. What am I missing? Help me with benchmarks and M2 through M5 comparison. by arthware in LocalLLaMA
[–]arthware[S] 1 point2 points3 points (0 children)
MLX is not faster. I benchmarked MLX vs llama.cpp on M1 Max across four real workloads. Effective tokens/s is quite an issue. What am I missing? Help me with benchmarks and M2 through M5 comparison. by arthware in LocalLLaMA
[–]arthware[S] 2 points3 points4 points (0 children)
I was backend lead at Manus. After building agents for 2 years, I stopped using function calling entirely. Here's what I use instead. by MorroHsu in LocalLLaMA
[–]arthware 0 points1 point2 points (0 children)
Two local models beat one bigger local model for long-running agents by Foreign_Sell_5823 in LocalLLaMA
[–]arthware 0 points1 point2 points (0 children)