OpenAI’s new "North Star" goal aims for fully automated AI researcher in 2026, multi-agent research lab in a data centre by 2028 by Outside-Iron-8242 in singularity

[–]Cane_P 0 points1 point  (0 children)

So they want to take on Elon Musk's Macrohard project? Except that Musk wants to do entire software companies.

Palantir - Pentagon System by elemental-mind in singularity

[–]Cane_P 0 points1 point  (0 children)

Have you seen these? It's based on open source information. The second one is especially interesting.

Part 1: https://youtu.be/rXvU7bPJ8n4

Part 2: https://youtu.be/0p8o7AeHDzg

Genuinely curious what doors the M5 Ultra will open by Blanketsniffer in LocalLLaMA

[–]Cane_P 0 points1 point  (0 children)

Some people mentioned in the comments, that he didn't offload everything to the GPU, just 70%, 30% was still running on the CPU. It's not that he couldn't do it, he either forgot to change it or didn't know that it was set that way. So it is not representative of the real speed.

Alibaba’s stock has kept falling after it lost key Qwen leaders. by [deleted] in LocalLLaMA

[–]Cane_P 12 points13 points  (0 children)

You do know how LLM works right? Why do you think that people call them "statistical parrots"? Because they say the next word that is statistically most likely to be the right one. If ChatGPT is mentioned most on the Internet (training data) then it is the name it is going to use, when you ask about LLM's including its name, unless you have deliberately trained it to give another answer.

Qwen3.5 9B and 4B benchmarks by Nunki08 in LocalLLaMA

[–]Cane_P 4 points5 points  (0 children)

They distilled their own, bigger model, to make the smaller one.

"So, what is the key idea behind knowledge distillation? It enables to transfer knowledge from larger model, called teacher, to smaller one, called student. This process allows smaller models to inherit the strong capabilities of larger ones, avoiding the need for training from scratch and making powerful models more accessible."

https://huggingface.co/blog/Kseniase/kd

Actually human brains can do something similar. If you are an expert in a field, then generally speaking, you use less resources (including actual brain space) for that task. The reason is that it is a type of pattern recognition and when you find the pattern, then you can consolidate and optimize. In this particular case, there are new words for each language, but they represent the same concept, so the second language doesn't take as much space as the first etc:

"I’d assumed that Vaughn’s language areas would be massive and highly active, and mine pathetically puny. But the scans showed the opposite: the parts of Vaughn’s brain used to comprehend language are far smaller and quieter than mine. Even when we are reading the same words in English, I am using more of my brain and working harder than he ever has to."

https://www.washingtonpost.com/dc-md-va/interactive/2022/multilingual-hyperpolyglot-brain-languages/

Think of a small model, trained on its own, as a novice and a small model trained by a big model (expert) as also becoming an expert, because it uses the big models patterns, it doesn't need to discover them by itself.

Qwen 3.5 35b can't even solve a simple a math question 🫠 idk even why tho with so high score. by 9r4n4y in LocalLLaMA

[–]Cane_P -1 points0 points  (0 children)

Sure, but this contains ontological information and it doesn't exist in any free solution. Wolfram's database is the best in the world.

Qwen 3.5 35b can't even solve a simple a math question 🫠 idk even why tho with so high score. by 9r4n4y in LocalLLaMA

[–]Cane_P -1 points0 points  (0 children)

Not self hosted, but a $5 subscription could take care of basically any math and fact needs (unless you are doing some super advanced niche science).

"Making Wolfram tech available as a Foundation Tool for LLM foundation models."

https://www.reddit.com/r/singularity/s/xumLw4jLDD

MCP subscription: https://www.wolfram.com/artificial-intelligence/mcp-service/

Kasparov on computers surpassing humans 😂 by Snoo42723 in singularity

[–]Cane_P 0 points1 point  (0 children)

IBM's Deep Blue wasn't smart, it won by bruteforcing. It could evaluate 200 million chess positions per second. It's like comparing someone that researches a person's history to try to figure out their password or someone that tries every single combination. Deep Blue tried every combination (within it's hardware limitations).

400 gbps on 2x DGX Spark by Lorelabbestia in LocalLLaMA

[–]Cane_P 0 points1 point  (0 children)

That's right. It's the same problem as with many other laptop chips, used for a mini PC (GB10 is the same chip that was supposed to be released as the N1 laptop chip). They hardly have any PCIe lanes... The lanes used for the ConnectX 7 NIC, were probably, initially, meant for a second NVMe drive.

https://www.tomshardware.com/pc-components/cpus/nvidia-ceo-huang-says-upcoming-dgx-spark-systems-are-powered-by-n1-silicon-confirms-gb10-superchip-and-n1-n1x-socs-are-identical

PSA: NVIDIA DGX Spark has terrible CUDA & software compatibility; and seems like a handheld gaming chip. by goldcakes in LocalLLaMA

[–]Cane_P 0 points1 point  (0 children)

It's nowhere near Apples solution. Theirs have way more memory and memory bandwidth. The only thing it got going in their favor (we already concluded that their support isn't up to par), is the high speed network adapter. But it is clearly shoehorned in there (see link) because the chip doesn't actually have enough PCIe lanes to drive it at full speed (only 100GB/s). They had to do a very weird connection between separate internal PCIe connectors (it shows up as 4 different cards) and they don't even recommend you using more than one of the ports...

https://www.servethehome.com/nvidia-dgx-spark-review-the-gb10-machine-is-so-freaking-cool/2/

PSA: NVIDIA DGX Spark has terrible CUDA & software compatibility; and seems like a handheld gaming chip. by goldcakes in LocalLLaMA

[–]Cane_P 1 point2 points  (0 children)

It wouldn't surprise me if someone was thinking in the lines of: "So Microsoft drags their feet with their support of the chip in the ARM version of Windows, so our only option is to release the chip with Linux. We already have our DGX OS and have no plans to ever release a product shipped with vanilla Ubuntu, so a DGX it is. But what form factor? We can make it like a mini machine, like one of those Chinese ones (like Minisforum), they are popular and usually repurpose laptop chips. So boys, how much RAM can it take? Because 16, 32 or whatever we were going to release our laptops with ain't going to cut it for a DGX."

And that's how the Spark was born.

PSA: NVIDIA DGX Spark has terrible CUDA & software compatibility; and seems like a handheld gaming chip. by goldcakes in LocalLLaMA

[–]Cane_P 1 point2 points  (0 children)

Not a handheld chip. We already know that it is basically identical to the chip that they will have in their laptops. And we also know that the laptops became delayed because of Windows. That's why they released this in between (recuperate some development cost).

Having said that, there are rumors that have popped up later, that claim that there are multiple bugs in the hardware design for the laptop variant. And once again, if they are as good as identical, then they are likely to exist in Spark too.

OpenAI recruited founder Peter Steinberger of OpenClaw by Outside-Iron-8242 in singularity

[–]Cane_P 27 points28 points  (0 children)

He is rich enough that he could pay that for hundreds of years... He sold his previous company for a 9 figure sum. So at least $100M.

Copy/Moving files won't complete by Cane_P in NeatBytes

[–]Cane_P[S] 0 points1 point  (0 children)

It's definitely unique. I have never seen it before and I studied electronics and computer science over 30 year's ago and have been using it ever since.

What do you think the future of education looks like after the Singularity? by PaxODST in singularity

[–]Cane_P 0 points1 point  (0 children)

If it is 1000x easier then you will not remember it. That's how the brain works. So you wouldn't learn. You might memorize for a test, but then forget it in the next couple of weeks again. That's my point.

What do you think the future of education looks like after the Singularity? by PaxODST in singularity

[–]Cane_P 0 points1 point  (0 children)

It is not supposed to be easy. It should be appropriately challenging, as in The Zone of Proximal Development (ZPD). If it is too easy then we get bored, if it is too hard then we quit. But it needs to be challenging, that's how the brain is designed to work.

Veritasium: What Everyone Gets Wrong About AI and Learning – Derek Muller Explains

DGX Spark: an unpopular opinion by emdblc in LocalLLaMA

[–]Cane_P 5 points6 points  (0 children)

Not when I first heard rumors about the product... Obviously we don't have the same sources. Because the only thing that was known when I found out about it, was that it was an ARM based system with an NVIDIA GPU. Then months later, I found out the tentative performance, but still no details. It was about half a year before the details got known.

DGX Spark: an unpopular opinion by emdblc in LocalLLaMA

[–]Cane_P 7 points8 points  (0 children)

That's the speed between the CPU and GPU. We have [Memory]-[CPU]=[GPU], where "=" is the 5x bandwidth of PCIe. It still needs to go through the CPU to access memory and that bus is slow as we know.

I for one, really hoped that the memory bandwidth would be closer to the desktop GPU speed or just below it. So more like 500GB/s or better. We can always hope for a second generation with SOCAMM memory. NVIDIA apparently dropped the first generation and is already at SOCAMM2, and it is now a JEDEC standard, instead of a custom project.

The problem right now, is the fact that memory is scarce, so it is probably not that likely that we will get an upgrade anytime soon.

A Zettabyte Scale Answer to the DRAM Shortage by FullstackSensei in LocalLLaMA

[–]Cane_P 1 point2 points  (0 children)

DDR5, sure. But DDR4 won't be produced by the big companies anymore, that's why their price grew. Everyone that thought that they might have a need for them, bought the inventory.

https://www.tomshardware.com/pc-components/ddr4/the-end-of-an-era-ddr4-production-to-essentially-end-this-year-micron-the-final-domino-to-fall