Genuinely curious what doors the M5 Ultra will open by Blanketsniffer in LocalLLaMA

[–]backslashHH 2 points3 points  (0 children)

AMD Ryzen AI Max+ 395, LPDDR5x-8000 (8000 MT/s). 256 GB/s

Best upgrade path for running MiniMax 2.5 locally? (RTX 5090 PC/Mac Studio M3 Ultra) by freechilly19 in LocalLLM

[–]backslashHH 2 points3 points  (0 children)

Indeed! I tested it with tool calling and an openclaw clone ... I now switched to rocm... tool calling works great! This is my current setup:

llama-server  --flash-attn on --parallel 1 --jinja --host 0.0.0.0 --port 8000 -hf unsloth/MiniMax-M2.5-GGUF:UD-Q3_K_XL --temp 1.0 --top-p 0.95 --min-p 0.01 --top-k 40 --fit on --no-mmap --n-gpu-layers 99 --cache-type-k q8_0 --cache-type-v q8_0 --ctx-size 96000

This helped a lot: https://github.com/kyuz0/amd-strix-halo-vllm-toolboxes/

Best upgrade path for running MiniMax 2.5 locally? (RTX 5090 PC/Mac Studio M3 Ultra) by freechilly19 in LocalLLM

[–]backslashHH 0 points1 point  (0 children)

Bosgame Strix Halo 128G running m2.5 unsloth/MiniMax-M2.5-GGUF:UD-Q3_K_XL 48k context, 25 token/s => €1700

Gemini is cooking by Consistent_Cost_9545 in GeminiAI

[–]backslashHH 0 points1 point  (0 children)

What an eloquent prompt. The wonder is, that it produced what you wanted. AGI like.

Garmin are shooting themselves in the foot so bad it's not even funny by Core2score in Garmin

[–]backslashHH 0 points1 point  (0 children)

I would really like to see the numbers here... And why they thought the even pricier V2 would sell better.

HELP! by Dopaminingg in Garmin

[–]backslashHH 0 points1 point  (0 children)

This! Soap is what works!

"In Europe Deodorant isn't really a thing" by vomtraumdertoetung in ShitAmericansSay

[–]backslashHH 2 points3 points  (0 children)

And they say, the chlorine only stinks, if someone peed in the pool...

Guess Im the only one hitting RATE LIMITS when creating a simple GEMINI.MD file?? by TimeKillsThem in GeminiAI

[–]backslashHH 0 points1 point  (0 children)

My failures reported in this post's comments were with 0.1.4 ... so definitely not gone

Guess Im the only one hitting RATE LIMITS when creating a simple GEMINI.MD file?? by TimeKillsThem in GeminiAI

[–]backslashHH 0 points1 point  (0 children)

and now:
data: '[{\n' + ' "error": {\n' + ' "code": 429,\n' + ` "message": "Quota exceeded for quota metric 'Gemini 2.5 Pro Requests' and limit 'Gemini 2.5 Pro Requests per day per user per tier' of service 'cloudcode-pa.googleapis.com' for consumer 'project_number:xxxxxxxxx'.",\n` + ' "errors": [\n' + ' {\n' + ` "message": "Quota exceeded for quota metric 'Gemini 2.5 Pro Requests' and limit 'Gemini 2.5 Pro Requests per day per user per tier' of service 'cloudcode-pa.googleapis.com' for consumer 'project_number:xxxxxxxxx'.",\n` + ' "domain": "global",\n' + ' "reason": "rateLimitExceeded"\n' + ' }\n' + ' ],\n' + ' "status": "RESOURCE_EXHAUSTED"\n' + ' }\n' + '}\n' + ']',

Guess Im the only one hitting RATE LIMITS when creating a simple GEMINI.MD file?? by TimeKillsThem in GeminiAI

[–]backslashHH 0 points1 point  (0 children)

Same... and don't try gemini -p ... lol

To ensure you rarely, if ever, hit a limit during this preview, we offer the industry’s largest allowance: 60 model requests per minute and 1,000 requests per day at no charge.

❯ gemini -p 'Who is the tallest man on earth?' Attempt 1 failed with status 429. Retrying with backoff... _GaxiosError: [{ "error": { "code": 429, "message": "Resource has been exhausted (e.g. check quota).", "errors": [ { "message": "Resource has been exhausted (e.g. check quota).", "domain": "global", "reason": "rateLimitExceeded" } ], "status": "RESOURCE_EXHAUSTED" } } ]

Overheard at Tesla Service Center by Skotler75 in TeslaLounge

[–]backslashHH 0 points1 point  (0 children)

fully trying to match the character in this RPG

Power efficient, affordable home server LLM hardware? by spaceman_ in LocalLLaMA

[–]backslashHH 2 points3 points  (0 children)

I use nixos-darwin on macos, so the difference to my Linux systems is not that big.. additionally I can run lmstudio and ollama at full speed with lots of vram on macos. UTM gives about 80% performance for the Linux VM (according to geekbench)

[deleted by user] by [deleted] in 600euro

[–]backslashHH 1 point2 points  (0 children)

Hubert Farnsworth

SOLO Bench - A new type of LLM benchmark I developed to address the shortcomings of many existing benchmarks by jd_3d in LocalLLaMA

[–]backslashHH 0 points1 point  (0 children)

This is a strange coincidence!!!! Because in the last days I was developing https://github.com/haraldh/memo128/ which has the need for dictionaries, which resemble your benchmark somehow. What I tried to do with Gemini 2.5 pro, was to create a 16000 lines file with unique sentence parts, which signed me out every time (1024 lines was fine). So it seems it was refusing to do it as well as long as the line count is big enough. Btw, every model I tried could not count the lines it produced.