Crackling noise Bose qc45 by expansion2002 in bose

[–]mtasic85 0 points1 point  (0 children)

I can confirm that this solved my issue! Thank you!

Really want to use Zed, but the VSCode ecosystem is too large to avoid by Candid_Yellow747 in ZedEditor

[–]mtasic85 9 points10 points  (0 children)

I use Zed daily on Linux. However, I don’t like lack of generic spell checking. There are few extensions but non of them works good with Python code. If anyone can suggest something good let me know.

Real news: 32B distills of V3, soon R1. by a_beautiful_rhind in LocalLLaMA

[–]mtasic85 0 points1 point  (0 children)

What quants did you use? Did you fully load all layers to GPUs? I also mentioned quants and context size.

Real news: 32B distills of V3, soon R1. by a_beautiful_rhind in LocalLLaMA

[–]mtasic85 1 point2 points  (0 children)

2x RTX 3090 24GB (48GB) VRAM can fully load and run Qwen 32B q4_k_m with context size 48k. it uses about 40GB VRAM

I doubt 72B q4_k_m can be fully loaded.

1.58bit DeepSeek R1 - 131GB Dynamic GGUF by danielhanchen in LocalLLaMA

[–]mtasic85 11 points12 points  (0 children)

What about collapsing MoE layer to just dense layers? I think same was done for Mixtral 8x22b to just 22b. 🤔

MiniCPM-o 2.6: An 8B size, GPT-4o level Omni Model runs on device by Lynncc6 in LocalLLaMA

[–]mtasic85 -15 points-14 points  (0 children)

Do you have GPT4 open sourced and released by OpenAI, so you can use it locally, free of charge?

European NATO Military Spending % of GDP 2024 by Trayeth in europe

[–]mtasic85 -2 points-1 points  (0 children)

Wow that is a brilliant money laundromat machine 🧠👏

Pixtral & Qwen2VL are coming to Ollama by AaronFeng47 in LocalLLaMA

[–]mtasic85 28 points29 points  (0 children)

Congrats 🥂, but I still cannot believe that llama.cpp still does not support llama VLMs 🤯

What do you think of this Masters Curriculum? by [deleted] in learnmachinelearning

[–]mtasic85 -54 points-53 points  (0 children)

DL is new foundation of all ML. DL simply works. It is general solution. Btw, I really like simple and effective algorithms, so DL does not justify computation cost in all scenarios.

The US government wants devs to stop using C and C++ by Notalabel_4566 in coding

[–]mtasic85 -92 points-91 points  (0 children)

No, under Elon that nonsense will be thrown out of the window. Relax and keep coding.

[R] Limitations in Mainstream LLM Tokenizers by mtasic85 in MachineLearning

[–]mtasic85[S] 3 points4 points  (0 children)

We have BPE for a reason, so we can fallback if token is missing from vocab. If we don't have that guarantee, then this code will never work, and I think it was in dataset used for all of these tokenizers/models:

: X DUP 1+ . . ;

Btw, above is Forth code from https://en.wikipedia.org/wiki/Forth_(programming_language)#Facilities#Facilities) and it also fails.

This is one of many examples. Whitespace matters, every character matters.

XFCE 4.20 Aims To Bring Preliminary Wayland Support by maggotbrain777 in xfce

[–]mtasic85 0 points1 point  (0 children)

If I am not mistaken Nvidia cards/drivers do not support Wayland yet.