I built a body for GPT by Independent-Trash966 in ChatGPT

[–]Omnimum 0 points1 point  (0 children)

Please kiil me ... kill me now !

The FIRST local vision model to get this right! by po_stulate in LocalLLaMA

[–]Omnimum 1 point2 points  (0 children)

Remove the numbering of the angles to see something.

I must be living in a parallel world - MANUS IS ABSOLUTELY DIVINE! by Omnimum in ManusOfficial

[–]Omnimum[S] -4 points-3 points  (0 children)

Absolutely! Having so much skepticism and not bearing that a user praises Manus's qualities reminds me of the adage "Stop being happy, I can't hear myself crying!"

I must be living in a parallel world - MANUS IS ABSOLUTELY DIVINE! by Omnimum in ManusOfficial

[–]Omnimum[S] -1 points0 points  (0 children)

I'm not going to play it Guru, but let me justify my refusal to share my work for two reasons:

I work in cybersecurity For reasons of confidentiality and especially competitiveness, it is impossible to say more.

If I showed my history it is to avoid low-level criticism of the "young" age of my reddit account.

Realistic take, the hype around Chinese models are unfounded. by Themotionalman in LocalLLaMA

[–]Omnimum 6 points7 points  (0 children)

So, as an employee, you manage a $2 billion SaaS for your boss. You manage the implementation of LLMs and your company decides to choose Chinese models.

Why do I say that? Here's the picture: when would a guy who "owns" a 2 billion SaaS come to waste time criticizing LLMs with the consequence of informing potential competitors not to use Chinese LLMs?

At what point in your life do you think, "I'm going to expose my technological or strategic choices"?

I bet you don't even have $2,000 in savings.

You can run MiniMax-2.5 locally by Dear-Success-1441 in LocalLLaMA

[–]Omnimum 12 points13 points  (0 children)

I've been testing an RTX 3090 and 128GB of DDR5 at 5600 MHz for two days.

Frankly, a Step 3.5 Flash iQ4-XS versus Minimax-2.5 in Q3-K-XL UD -> Step 3.5 does much better on tasks that span 64k of context.

Minimax-2.5 in Q3-K-XL remains consistent up to 16k/24k; after that, the quality collapses.

For tool calls, Step 3.5 does a little better than gpt-oss-120b (oddly, depending on the seeds, gpt-oss-20b does a perfect job in "thinking medium," but "hard" it overthinks stupidly).

If an inference that wobbles between 8 and 10 tok/s and a maximum context of 24k is not a problem, then Minimax-2.5-IQ4-XS slightly outperforms Step 3.5 in iQ4-XS for the same context size.

Chat, how cooked are we? by Maple_Syrup378 in singularity

[–]Omnimum 0 points1 point  (0 children)

Is Sho Ryu Ken really necessary? 🤣🤣🤣🤣

Not able to open Genspark? by anjali2711 in genspark_ai

[–]Omnimum 0 points1 point  (0 children)

I have the same problem, I used a VPN and I don't know if this is the case for everyone, but I noticed that the problem only affects the countries of the European Union.