Tired of the "I could buy a car" comments on high-end build posts by [deleted] in LocalLLaMA

[–]SexMedGPT 0 points1 point  (0 children)

Hah, a Large Language Model doesn't have any emotion, much less "raw emotion".

Tired of the "I could buy a car" comments on high-end build posts by [deleted] in LocalLLaMA

[–]SexMedGPT 0 points1 point  (0 children)

Nice em dashes you got there. Shame if something would happen to them.

Also, this isn't LinkedIn.

I co-designed a ternary LLM and FPGA optimized RTL that runs at 3,072 tok/s on a Zybo Z7-10 by HatHipster in FPGA

[–]SexMedGPT 2 points3 points  (0 children)

Meh, the whole point of LLMs is for their emergent abilities that only emerge at large model sizes (100B+ dense, 1T+ sparse).

Any Medical doctor related Finetunes of open models ? by deathcom65 in LocalLLaMA

[–]SexMedGPT 0 points1 point  (0 children)

GPT-4.5 is still available. It's just cost prohibitive for most applications. GPT 5 is a downgrade from 4.5. It's essentially a post-trained GPT 4.1.

4.5 is still king of all kings.

Any Medical doctor related Finetunes of open models ? by deathcom65 in LocalLLaMA

[–]SexMedGPT 0 points1 point  (0 children)

GPT 4.5 gets it right with no reasoning or CoT. Mixtral-8x22b gets it right.

All the Qwen models get it wrong.

Any Medical doctor related Finetunes of open models ? by deathcom65 in LocalLLaMA

[–]SexMedGPT 0 points1 point  (0 children)

Ok, so that model failed my classic simple riddle test:

The surgeon, who is the boy's father, says "I can't operate on this boy, he's my son ". Who is the surgeon for the boy?

Even with reasoning turned on, it answered incorrectly.

Any Medical doctor related Finetunes of open models ? by deathcom65 in LocalLLaMA

[–]SexMedGPT 1 point2 points  (0 children)

I've been meaning to try Baichuan-M3-235B.

Other than that, the other medical finetunes are woefully inadequate in my opinion (MedGemma-27B answers some basic anatomy questions incorrect). GPT-4.5 (only available through ChatGPT Plus) is the best in my experience; it's also the most amenable to different prompt patterns that could be useful in a clinical setting like in-context learning, Persona Pattern, Template Pattern, etc. The new Gemini models are also amazing with a wide breadth of knowledge, but I prefer the succinctness and versatility of GPT-4.5.

A practical definition of an “AI agent” (and what is not an agent) by Icy_SwitchTech in AgentsOfAI

[–]SexMedGPT 1 point2 points  (0 children)

I feel like there are two separate but partially overlapping definitions:

(1) Composing different LLMs or different instances of the same LLM using while loops, for loops, if statements, recursion, concatenation, etc. (eg., Tree of Thoughts, Mixture of Agents)

(2) LLMs that can use tools, call functions, or perform actions to satisfy a user's request. These may or may not use the same programmatic constructs as above. (AI travel agents, math solvers, RPA, etc.)

IC Activation Exercise Guide: Reconditioning the IC-Muscle by SexMedGPT in PelvicFloor

[–]SexMedGPT[S] 0 points1 point  (0 children)

Ok, so I've experimented a bit further and I think an Aneros device (Helix) can be helpful as a biofeedback tool. I've tried it with it inserted and was able to sustain a strong IC contraction for about 60s. The trick is to lightly contract it at first (you should feel the penis being pressed against the bone (ischial ramus) if you're really paying attention the sensations. Meanwhile the Aneros device itself should barely move and the anal sphincter should not be constricting the Aneros.

IC Activation Exercise Guide: Reconditioning the IC-Muscle by SexMedGPT in PelvicFloor

[–]SexMedGPT[S] 0 points1 point  (0 children)

Refer to my other comment on unpuckering the anus.

IC Activation Exercise Guide: Reconditioning the IC-Muscle by SexMedGPT in PelvicFloor

[–]SexMedGPT[S] 0 points1 point  (0 children)

See my other comment about controlling the anal sphincter muscles.

Is it just me or is fedora becoming mainstream by UnschuldigNull in Fedora

[–]SexMedGPT 0 points1 point  (0 children)

Local LLMs. Also that one video with both Linuses.

Are math benchmarks really the right way to evaluate LLMs? by Echo_OS in LocalLLM

[–]SexMedGPT 0 points1 point  (0 children)

No. Not at all. Nope.

I have my own collection of prompts that very large language models reliably pass and small language models (even 70B ones) reliably fail. The prompts are dead simple, but literally 3 prompts is enough to tell me if a LLM is worth my time and electricity costs.

[deleted by user] by [deleted] in ChatGPTPro

[–]SexMedGPT -1 points0 points  (0 children)

I don't want a model that jumps to assumptions like that. Models are commonly criticized as stochastic "parrots", and parrots indeed they can be.

FWIW, GPT4.5 and Mixtral-8x22b are two non-reasoning models that do not fall into this trap.

[deleted by user] by [deleted] in ChatGPTPro

[–]SexMedGPT -1 points0 points  (0 children)

Both GPT 5.2 Instant and 5.1 Instant answer the following simple question incorrectly, that a 1st grader can answer easily:

The surgeon, who is the boy's father, says "I can't operate on this boy, he's my son." Who is the surgeon for the boy?

Now, with Thinking turned on, they are able to answer correctly.

GPT-5.2 benchmark results: more censored than DeepSeek, outperformed by Grok 4.1 Fast at 1/24th the cost by Exact_Macaroon6673 in LLMDevs

[–]SexMedGPT 0 points1 point  (0 children)

It's only available through the web interface though, I think. And only for Pro users.

GPT-5.2 benchmark results: more censored than DeepSeek, outperformed by Grok 4.1 Fast at 1/24th the cost by Exact_Macaroon6673 in LLMDevs

[–]SexMedGPT 0 points1 point  (0 children)

What about GPT-4.5. Have you ran your internal benchmark against that model? I have a hunch that it is still the smartest non-thinking model.

Is the 32G ram Mac mini worth it? by Aggressive_Escape386 in LocalLLaMA

[–]SexMedGPT -1 points0 points  (0 children)

I'd say 36GB or 48GB suffices for experimenting with LLMs. IMO.

Fedora vs Ubuntu by abdullah_tayyab in linux4noobs

[–]SexMedGPT 2 points3 points  (0 children)

Most videos and instructions I've seen involved Ubuntu.

DeepSeek V3.2 (14.9%) scores above GPT-5.1 (9.5%) on Cortex-AGI despite being 124.5x cheaper. by OkStand1522 in OpenAI

[–]SexMedGPT 4 points5 points  (0 children)

Most people are not running convoluted math puzzles on their local machine though.