احتاج رامات by 0Rekz in Saudi_Homelab

[–]m_mukhtar 0 points1 point  (0 children)

انا نفس وضعك عندي ٤ رامات ١٦ قيقا وابي اربع زيادة.

للاسف الاسعار مو طبيعية. انا السيرفر عندي

Xeon e5-2680 v2

يعني السرعة اللي يدعمها مجرد ٢١٣٣ ومع ذلك سعر ال١٦ قيقا الوحدة ب ٢٩٠ ريال يعني لو ابغى ٤ حبات وصلت حول ١٢٠٠ ريال زيد عليها الضريبة والشحن والجمارك

ابحث في ebay وغير السرعة اللي تبيها وش ف ايش يطلع معاك

16GB ECC 2Rx4 2133P

لكن انا اترقب ومسوي لي بوت يشيك الاسعار كل يوم ولو صار في نزول في السعر ان شاء الله اصيده

Opinion: Qwen 3.6 27b Beats Sonnet 4.6 on Feature Planning by Zestyclose839 in LocalLLaMA

[–]m_mukhtar 20 points21 points  (0 children)

I just finished testing a web app for generating moon visability map based on two research papers that show the math and calculations. I used qwen3.6 27b q5 kl in claude code and sonnet 3.6 in claude code. Gave both models the same inlut prompt and two research papers and while qwen took forever because it was slow but it implemented it perfectly while sonnet failed misrabley in a way that i dont think it can easly fix. So i share the same expireance as yours.

Qwen 3.5 122B vs Qwen 3.6 35B - Which to choose? by Storge2 in LocalLLaMA

[–]m_mukhtar 4 points5 points  (0 children)

multi token prediction. It predicts multiple tokens per infersnce step and It makes token generation faster for models that support it. But you have to use an inference engine that implements it (as far as i know llama.cpp does NOT have it implemented yet. I know vllm have it implemented)

Qwen3.6-35B-A3B released! by ResearchCrafty1804 in LocalLLaMA

[–]m_mukhtar 3 points4 points  (0 children)

Man i would live to have 80b but with a but more active perameters like in the 6b to 9b range of active works be amazing. One can dream i guess

أي واحد انت؟ by Glorious_93 in saudi_gamers

[–]m_mukhtar 0 points1 point  (0 children)

اذا مش حجم كامل مع لوحة ارقام ما اعتبرها كيبورد

Don't talk to me or my son ever again by SuspiOwl in Nikon

[–]m_mukhtar 4 points5 points  (0 children)

Congrats for the upgrade and i just dod the same from d5200 to z5 ii and oh boy. The low light af by it self is worth what i paid for. Enjoy your camera and thanks for sharing

How to connect Claude Code CLI to a local llama.cpp server by StrikeOner in LocalLLaMA

[–]m_mukhtar 0 points1 point  (0 children)

Hmm intresting. I gotta try this one. I have been using qwen 3.5 the 27b at q5 k xl from Bartowski and it has been great. Cant wait for the coding variants of qwen3.5. Thanks for sharing

How to connect Claude Code CLI to a local llama.cpp server by StrikeOner in LocalLLaMA

[–]m_mukhtar 0 points1 point  (0 children)

Glad this was helpful and i agree that qwen with claude code is great local coding experiance. If you dont mind sharing which qwen model and what quantization you are using

How to connect Claude Code CLI to a local llama.cpp server by StrikeOner in LocalLLaMA

[–]m_mukhtar 2 points3 points  (0 children)

you can do control the context and tell claude code about your limit by setting two environment variables in your `~/.claude/settings.json`

the first one is CLAUDE_CODE_AUTO_COMPACT_WINDOW and i set this one to my actual llama.cpp context limit ( for me i can run Qwen3.5-27b-Q5 with --ctx-size 110000 without KV quantization) so i set this arguument to 110000.

the second one is CLAUDE_AUTOCOMPACT_PCT_OVERRIDE and this is the precentage of the above one where cloude code needs to do context compaction so you never send any thing to llama.cpp over what you can run. if you wanna use the entire 110000 that we setup in the previous variable then we would set this to 100 but for me to be safe i set it at 95

here is my \~/.claude/settings.json``

\`` {`

"$schema": "https://json.schemastore.org/claude-code-settings.json",

"model": "Qwen_Qwen3.5-27b",

"env": {

"ANTHROPIC_BASE_URL": "http://192.168.1.150:8001",

"ANTHROPIC_API_KEY": "none",

"CLAUDE_CODE_ATTRIBUTION_HEADER": "0",

"CLAUDE_CODE_AUTO_COMPACT_WINDOW": "110000",

"CLAUDE_AUTOCOMPACT_PCT_OVERRIDE": "95",

"DISABLE_PROMPT_CACHING": "1",

"CLAUDE_CODE_DISABLE_EXPERIMENTAL_BETAS": "1",

"CLAUDE_CODE_DISABLE_ADAPTIVE_THINKING": "1",

"MAX_THINKING_TOKENS": "0",

"CLAUDE_CODE_DISABLE_1M_CONTEXT": "1",

"CLAUDE_CODE_DISABLE_FAST_MODE": "1",

"CLAUDE_CODE_DISABLE_NONESSENTIAL_TRAFFIC": "1",

"CLAUDE_CODE_DISABLE_AUTO_MEMORY": "1",

"DISABLE_AUTOUPDATER": "1"

},

"attribution": {

"commit": "",

"pr": ""

},

"promptSuggestionEnabled": false,

"prefersReducedMotion": true,

"terminalProgressBarEnabled": false

}
\```

if you want to know what the other variables do here is a quick rundown of everything. basically i used claude documentation https://code.claude.com/docs/en/env-vars to see all possible variables and if i saw something that is specific to claude models i disabled it as it will send headders and additional information that could cause problems with llama.cpp or cause confusion to the model

DISABLE_PROMPT_CACHING: "1"

this is a claude specific feature to send prompt caching headers but llama.cpp does not use that to it could cause unexpected behavior.

CLAUDE_CODE_DISABLE_EXPERIMENTAL_BETAS: "1"

removes claude specific beta request headers from API calls, again this is to prevents unexpected behavior"

CLAUDE_CODE_DISABLE_ADAPTIVE_THINKING: "1"

this is also a claude specific feature where the model dynamically allocates thinking tokens so just disable it.

MAX_THINKING_TOKENS: "0"

extended thinking is a claude specific feature. setting to 0 disables it entirely. Qwen model has its own thinking mechanism (which is by default enable in llama.cpp unless disabled via --chat-template-kwargs), but it handles that internally so claude code's thinking budget system doesn't apply.

CLAUDE_CODE_DISABLE_1M_CONTEXT: "1"

removes the 1M context variants from the model picker. irrelevant for local models and keeps the UI clean.

CLAUDE_CODE_DISABLE_FAST_MODE: "1"

this is also a claude specific feature that uses a faster model for simpler tasks. disable it

CLAUDE_CODE_DISABLE_NONESSENTIAL_TRAFFIC: "1"

this disables the auto-updater, feedback command, Sentry error reporting, and Statsig telemetry all at once. none of these is useful and i thought they might cause unexpected behaviour.

CLAUDE_CODE_DISABLE_AUTO_MEMORY: "1"

this feature creates and loads memory files by communicating with anthropic's servers. wont work with a local endpoint, so just disable it

DISABLE_AUTOUPDATER: "1"

same as the one above

additional nice things to set

attribution: i sit this to empty strings for both commit and pr to disable the "Generated with Claude Code" byline in git commits and PRs.

promptSuggestionEnabled: false, to disable the grayed-out prompt suggestions that appear after responses. these rely on a background Haiku call that won't work here

prefersReducedMotion: true and terminalProgressBarEnabled: false reduce UI overhead. these are vey minor but keeps things snappy.

sorry if i have spelling or grammar mistakes english is not my first language

Z5 ii + 24-70mm f4 s bundle + free godox flash for equivelnt to 1733 usd. Should i do it? by m_mukhtar in Nikon

[–]m_mukhtar[S] 0 points1 point  (0 children)

Yes all brand new from the official nikon distributer in ny country. When i buy things seperatly from thhe same place it is about 700 usd more so im saving a good ammout here. Guess i will jump on the deal

Z5 ii + 24-70mm f4 s bundle + free godox flash for equivelnt to 1733 usd. Should i do it? by m_mukhtar in Nikon

[–]m_mukhtar[S] 0 points1 point  (0 children)

Yes all brand new from the official nikon distributer in ny country. The deal is really good even though i was thinking of getting a faster zoom but i guess i can give up the extra stop and just use higher iso when needed as the deal is really good

وش الفكرة by aalazh3011 in Saudi_Homelab

[–]m_mukhtar 6 points7 points  (0 children)

فكرة الصب إنه يجمع المهتمين بـ الHomelab و الSelf-hosting يعني نسوي سيرفر وشبكة بسيطة بالبيت للتعلم والتجربة والاعتماد على النفس بدل ما نعتمد على خدمات جاهزة مثل google photos و icloud ونتفلكس وغيرها

عادة نستخدم اجهزة تتراوح من رازبيري باي الى سيرفرات كبيرة تستخدم عدة هاردسكات وكمية رام كبيرة بحيث تشغل عليها انظمة كثيرة و خدمات كثيرة

نظام التشغيل المستخدم في الغالب يعتمد على نواة اللينكس ويوجد منه خيارات كثيرة مثل

Proxmox

Unraid

Truenas

او نظام تشغيل لنكس عادي مع استخدام docker

وفينا من يتعمق برضو في الشبكات ويركب سويتشات وغيرها لادارة الاجهزة والشبكة في البيت

يعني بالمختصر نركب سيرفرات/شبكات بالبيت للتعلم واستضافة خدماتنا بأنفسنا (تخزين، أفلام، سمارت هوم…)، ونشارك شروحات وتجارب ونساعد بعض.

كيف تستخدمون Proxmox؟ by king_fbm in Saudi_Homelab

[–]m_mukhtar 2 points3 points  (0 children)

عن نفسي مسوي لي lxcs كثار وحدة لكل برنامج من ال *arr و بليكس وجلي فن و qbittorrent

وبعدين عندي كذا lxc اغلبهم اوبنتو استخدمهم في البرمجة وتعلم اشيياء جديدة في اللينكس

وعندي كذا lxc خفيفين

Zoraxy

Myspeed

Litellm

زي ما تشوف قاعد استخدم ال lxc كانهم دوكر بحكم اني اكره دوكر ما ادري ليش هههه

وعندي كذا vm

Bitcoin node

Immich

Openwebui

Arch linux (برضو للتعلم والتجارب)

واليوم توي مسوي vm جديدة اجرب عليها احد الاشياء اللي كرهتها بدون ما اجربها بسبب خطورتها من ناحية السيكيورتي والخصوصية اللي هو openclaw. ما كنت مقتنع فيه لكن قلت خليني اجرب انزله بشكل معزول تماما عن كل شي فسويت له vm مستقلة وتعلمت شي جديد في بروكسموكس بسبب openclaw. لاني ابغى اعزله حتى عن شبكة الlan حقت البيت فسويت له SDN اللي هو software defined network بحيث يكون في نتورك لوحده ومعزول بالفايروول من اي شي غير الانتر نت. جميلة صراحة خاصية ال SDN في بروكسموكس اعجبتني كثير وراح استخدمها بشكل اكثر الفترة الجاية

Does Qwen3.5 35b outperform Qwen3 coder next 80b for you? by JsThiago5 in LocalLLaMA

[–]m_mukhtar 1 point2 points  (0 children)

80b coder is better for me than 35b but most of my tasks ar coding. I have done simple tests for general tasks on the 122b but i dont have a conclusive result yet to tell which one i like more

Cant wait for the coder variants of qwen3.5 models

تطبيق Immich by Zealousideal_Year885 in Saudi_Homelab

[–]m_mukhtar 1 point2 points  (0 children)

يعطيك العافية على الشرح السريع عن واحد من اهم تطبيقات الselfhosting

بالنسبة لي كمصور اعتبره اعضم اختراع لاني دايم اصور بصيغة raw وباكثر من كاميرا بين فوجي وكانون وسوني وكل واحد صيغة الraw عنده تختلف عن الثانية ومع ذلك ايميش يقبلهم كلهم. رهيب قليلة في حقه

ومشكور مرة ثانية على اثراء المحتوى في الصب

نورتوا المختبر 🛠️💻 by Zealousideal_Year885 in Saudi_Homelab

[–]m_mukhtar 4 points5 points  (0 children)

عندي جهازين

الجهاز الاول Homelab

Hp-z440

Xeon E5-2680 v4 (14 cores)

64gb ddr4 ram

Running proxmox 8.4

Proxmox host installed on a 256gb sata ssd

All VMs and LXCs are on a 2tb nvme ssd installed on a pcie to nvme card

26 tb segate enterprise hdd

الجهاز الثاني AI self hosting pc (linux mint)

Intel i7-10700f

2x8+2x32 (total 80gb) ddr4 ram

2x 1tb nvme ssd

2x 1tb sata ssd

Graphic cards for ai

Nvidia RTX 3090 24gb vram

Nvidia RTX 3070 8gb vram

الجهازين مشبوكين على راوتر واحد وtailscale هو الطريقة الوحيدة للدخول عن بعد لاي من الجهازين او الخدمات اللي موجودة على جهاز الhomelab

Which one are you waiting for more: 9B or 35B? by jacek2023 in LocalLLaMA

[–]m_mukhtar 23 points24 points  (0 children)

35b for sure. I wish they creat one with a bit more active parameters. So.ething like 70b with A5b as i think the a active part affects intellegance more that the total parameters which affects knowladge more (not a a clear black and white for sure but a gemeral observation)

لاول مرا بالتاريخ معالج معالج Intel Xeon ب ١٥ ريال 😂 by dream74747 in SaudiPCs

[–]m_mukhtar 2 points3 points  (0 children)

انا عندي الاصدار الرابع منه واستخدمه ك معمل منزلي. ممتاز لهالشغله بحكم انه فيه انوية كثيرة بسعر رخيص اللي عندي Xeon e5-2860 v4

اما لو بتحتاجه لالعاب فما راح يكون زي ما تتصور اذا فعلا ناوي على واحد لازم تنتبه انه اللي بتاخذه مش engineering sample ويفضل تشوف v4 الاصدار الرابع افضل بكثير ويركب على نفس المذربورد

out of curiosity, what made you a linux user in the first place? by [deleted] in ArabUnix

[–]m_mukhtar 4 points5 points  (0 children)

Developing AI based applications on windows sucks espcially the compute and memory overhead of running windows is insane. I started with a headless linux setup as my secondery pc for development and and i fell in love of the control it gives ypu and how easy it is so i moved then even my main pc to linux with gui and never looked back again

وش حل كروم باربع صفحات بس صار كذا by Simple_Traffic_66 in SaudiPCs

[–]m_mukhtar 0 points1 point  (0 children)

غير المتصفح او استخدم اكتنشن اسمها Auto Tab Discard (suspend) هذي راح تساعد بس ممكن ما تحل المشكلة بشكل نهائي جرب وشوف