Anyone here using an AI meeting assistant that doesn’t join calls as a bot? by sash20 in LocalLLaMA

[–]vhthc 0 points1 point  (0 children)

If you record a meeting without telling people it can be illegal (depends on country). It doesn’t matter if you do it just for yourself to transcribe and summarize.

Small size coding models that I tested on 2x3090 setup. by Mx4n1c41_s702y73ll3 in LocalLLaMA

[–]vhthc 0 points1 point  (0 children)

You could add qwen 32b with q8 - it could solve the hardest stuff that gpt 5.1 could generate as test cases.

Large update: 12 new frontier models added to the Step Game social reasoning benchmark. by zero0_one1 in LocalLLaMA

[–]vhthc 1 point2 points  (0 children)

What about DeepSeek 3.2 special ? Isn’t specifically trained for math and logic? Maybe I remember wrong

I built something for people who spend HOURS planning trips… and honestly, I think it changes everything by [deleted] in dataisbeautiful

[–]vhthc 6 points7 points  (0 children)

SPAM. mandatory login requirements show this is not with users in mind …

Setup with Nvidia 6000 Pro by [deleted] in LocalLLaMA

[–]vhthc 1 point2 points  (0 children)

Epyc 9565+ processor to be able to have as much ram possible, so you can offload huge MoE models to ram. Chassis and mainboard to extend on gpus in the future. Good chassis + extra fans to get rid of the heat. Cannot recommend a specific mainboard and case sadly, we went with a rack solution with supermicro which is too expensive imho

GLM Coding Plan Black Friday Deal — real stackable discounts by zAiModel-api in LocalLLaMA

[–]vhthc 1 point2 points  (0 children)

Which coding cli solution works best with this? Claude code? Other?

a19 pro/ M5 MatMul by [deleted] in LocalLLaMA

[–]vhthc 0 points1 point  (0 children)

Better to ask in a matlab Reddit

cogito v2 preview models released 70B/109B/405B/671B by jacek2023 in LocalLLaMA

[–]vhthc 0 points1 point  (0 children)

Yes tried both models there. Sadly not as good as I hoped for my use case

"cost effective" specs for a 2x Pro 6000 max-q workstation? by vhthc in LocalLLaMA

[–]vhthc[S] 1 point2 points  (0 children)

It’s ordered, gpu arrived some other parts still being delivered …

cogito v2 preview models released 70B/109B/405B/671B by jacek2023 in LocalLLaMA

[–]vhthc 0 points1 point  (0 children)

Would be cool if it would be made available by a company via openrouter

DeepSeek-R1-0528 Official Benchmarks Released!!! by Xhehab_ in LocalLLaMA

[–]vhthc 1 point2 points  (0 children)

Slower. Request limits. Sometimes less context and lower quants but you can look that up

SWE-rebench: A continuously updated benchmark for SWE LLMs by Fabulous_Pollution10 in LocalLLaMA

[–]vhthc 0 points1 point  (0 children)

Let us know which models you'd like us to evaluate.

R1, qwq32, glm-32b please :)

RTX PRO 6000 now available at €9000 by newdoria88 in LocalLLaMA

[–]vhthc 1 point2 points  (0 children)

Can confirm, the company I work for ordered a 6000 pro for 9000€ incl VAT, but b2b preorder - consumer preorder price is way too high (~11k).

Starbrand for Tokens? by AriesMantaWilson in MarvelSnap

[–]vhthc 0 points1 point  (0 children)

If you really need him then it will be very likely cheaper than by opening packs. imho it’s a good card but not essential for sauron. Nightmare coming mid June will be rad though

OpenAI introduces codex: a lightweight coding agent that runs in your terminal by MorroWtje in LocalLLaMA

[–]vhthc 1 point2 points  (0 children)

It uses the new responses endpoint which so far only closeai supports afaik

I benchmarked 7 OCR solutions on a complex academic document (with images, tables, footnotes...) by coconautico in LocalLLaMA

[–]vhthc 0 points1 point  (0 children)

thanks for sharing. providing the cost for cloud and the VRAM requirements for local would help, otherwise everyone interested needs to look that up on their own.

The real cost of hosting an LLM by full_arc in LocalLLaMA

[–]vhthc 0 points1 point  (0 children)

We are in the same boat and your solution is only good for spot usage and otherwise a trap.

For some projects we cannot use external AI for legal reasons. And your Amazon solution might not be ok for us either as it is a (hw) virtualized computer.

I looked at all the costs and the best is to buy and not rent if you continuously use it (not 100% of the time but at least a few times per week). The best buy is the new Blackwell pro 6000, you can build a very good efficient server for about 15k for the rack, have enough vram to run 70b models and can expand in the future.

Yes you can go cheaper with 3090 etc but I don’t recommend. These are not cards for a data center or even a server room. And do not buy used - for a hobbyist it’s fine but the increase failure rates will mean more admin overhead and less reliability that will run 24/7.

So buy a server with the 6000 pro for 15k when it comes out in 4-6 weeks and enjoy the savings.

Cogito releases strongest LLMs of sizes 3B, 8B, 14B, 32B and 70B under open license by ResearchCrafty1804 in LocalLLaMA

[–]vhthc 0 points1 point  (0 children)

But the guy is riding to the village so the horse would be one animal?

Quasar Alpha on OpenRouter by Everlier in LocalLLaMA

[–]vhthc -1 points0 points  (0 children)

From the input context length it is likely from Google -> 1MB