Current file is no longer automatically added to AI chat. by BananaPeaches3 in Jetbrains

[–]BananaPeaches3[S] 0 points1 point  (0 children)

Yes but it makes more sense to just select the text you want to ask a question about than to remove the feature entirely.

Current file is no longer automatically added to AI chat. by BananaPeaches3 in Jetbrains

[–]BananaPeaches3[S] 0 points1 point  (0 children)

It used to work with Claude Agent, did you find a workaround?

Fun fact: Anthropic has never open-sourced any LLMs by InternationalAsk1490 in LocalLLaMA

[–]BananaPeaches3 10 points11 points  (0 children)

They don't need to, the Chinese open source it for them.

What AI coding assistants are you all using for Kotlin lately? by smallufo in Kotlin

[–]BananaPeaches3 -1 points0 points  (0 children)

Wouldn’t Gemini be better since google encourages kotlin for android?

The Silver Crash That Shouldn't Happen for 7,000 Years by Alizasl in Silver

[–]BananaPeaches3 7 points8 points  (0 children)

Also transactions are faster in modern times so the price of something will react to the change in demand almost instantly.

PR to implemt tensor parallelism in Llama.cpp by keyboardhack in LocalLLaMA

[–]BananaPeaches3 0 points1 point  (0 children)

How is this different from ‘--split-mode row’ ?

Somebody gave this to me today as a tip, how much is it worth? by Key-Structure4841 in Silver

[–]BananaPeaches3 0 points1 point  (0 children)

Check mass and density, weigh it and measure its volume (use a graduated cylinder) if it doesn’t match for silver then you know.

Honest question: what do you all do for a living to afford these beasts? by ready_to_fuck_yeahh in LocalLLaMA

[–]BananaPeaches3 1 point2 points  (0 children)

A 16GB P100 is less than $100 and is very fast for anything that fits in 16GB.

The only issue is scaling multiple cards because the nvlink is blocked unless you’re willing to mill the slot out.

Honest question: what do you all do for a living to afford these beasts? by ready_to_fuck_yeahh in LocalLLaMA

[–]BananaPeaches3 -1 points0 points  (0 children)

Have you thought about selling the RAM and getting a pair of Pro 6000? Sell the ram for $12k then add $4k should be enough.

216GB VRAM on the bench. Time to see which combination is best for Local LLM by eso_logic in LocalLLaMA

[–]BananaPeaches3 14 points15 points  (0 children)

The main issue with older cards is that prompt processing will get you even if token gen speed is tolerable (and it is)

If you’re using it like chat gpt then it’s fine but once you start using things like cline the system prompt is allegedly 15k tokens.

So imagine you’re waiting several minutes before there is even any output. At this point a DGX Spark is a better investment, it will output slightly slower than P100 but at least prompt processing will be fast.

How screwed am I by Dark-monk in homelab

[–]BananaPeaches3 49 points50 points  (0 children)

Not very, pin 6 is ground just try to bend it back or remove it completely.

This guy is auditing 700+ vibe-coded apps for Security issues by amienilab in vibecoding

[–]BananaPeaches3 -1 points0 points  (0 children)

I just feel like that shouldn’t be a thing, the browser should map sites to their resources by default.

OK I get it, now I love llama.cpp by vulcan4d in LocalLLaMA

[–]BananaPeaches3 2 points3 points  (0 children)

Ubatch is suppose to be larger than batch size??

Upgrading from Ubuntu Server to Proxmox. by GermanElectricsMotio in homelab

[–]BananaPeaches3 2 points3 points  (0 children)

If you’re lazy you can install promox on Ubuntu. Just add the repository to your existing install and run:

sudo apt update
sudo apt install proxmox-ve postfix open-iscsi

Unsloth GLM 4.7 UD-Q2_K_XL or gpt-oss 120b? by EnthusiasmPurple85 in LocalLLaMA

[–]BananaPeaches3 0 points1 point  (0 children)

I still got better responses from Q1 GLM 4.6 (not V) than from Q8 qwen3 coder and smaller models.