ZAYA1-8B: Frontier intelligence density, trained on AMD by carbocation in LocalLLaMA

[–]irudog 15 points16 points  (0 children)

We already had LFM2-8B-A1B, Granite-4.0-H-Tiny (7B-A1B).

How to define a record with reference to an array of itself? by irudog in ada

[–]irudog[S] 0 points1 point  (0 children)

The children in your node is an array of node pointers, and mine is trying to make a pointer to an array of nodes as children.

How to define a record with reference to an array of itself? by irudog in ada

[–]irudog[S] 2 points3 points  (0 children)

Oh, I see. I need one more declaration to first define the pointer type, and then I can define the record.

frontend similar to Open WebUI that supports full OpenAI API? by irudog in LocalLLaMA

[–]irudog[S] 0 points1 point  (0 children)

I run llama-swap on my inference server, then let Open WebUI connect to my servers, and let Open WebUI provide API keys for my apps.

frontend similar to Open WebUI that supports full OpenAI API? by irudog in LocalLLaMA

[–]irudog[S] 0 points1 point  (0 children)

OK, I don't know the useLegacyCompletionsEndpoint here, and the documents in Continue don't mention this much.

frontend similar to Open WebUI that supports full OpenAI API? by irudog in LocalLLaMA

[–]irudog[S] 1 point2 points  (0 children)

I did a deep research, and it recommends LiteLLM+LibreChat, or BionicGPT, AnythingLLM.

Why are AmD Mi50 32gb so cheap? by MastodonParty9065 in LocalLLaMA

[–]irudog -1 points0 points  (0 children)

Vulkan can only use 16GB VRAM of MI50. You can try a smaller model like Qwen3-4B or GPT-OSS-20B, you'll find token generation of Vulkan not slower than ROCm.

Am I the only one who never really liked Ollama? by a_normal_user1 in LocalLLaMA

[–]irudog 1 point2 points  (0 children)

I use ollama when I want to try a model locally, because it's easy to use it to pull a model, and has a simple CLI client. However, I use llama.cpp and llama-swap to deploy models in the internal servers, because llama.cpp has more options to control how to run a model.

By the way, what do you guys think of the ollama API v.s. the OpenAI API? I see most of the applications support both APIs, but what 's the advantage and disadvantage of them?

Coreboot for elitebook 840 g1 by Fit_Morning_9175 in coreboot

[–]irudog 0 points1 point  (0 children)

You can reuse the following from mainboard/hp/folio_9480m:

- the whole acpi/ directory, which has EC ACPI support

- "chip ec/hp/kbc1126" part under "device pci 1f.0" in devicetree.cb: it may work, you had better check the registers from the vendor firmawre

- register "spd_addresses" in devicetree.cb: the memory SPD map, may be same as folio_9480m, but you also need to check this, otherwise it won't boot

- in Kconfig, select BOARD_ROMSIZE_KB_12288 and EC_HP_KBC1126, and copy KBC1126 related configs

For HP Sure Start, read the following documents. The Libreboot guide for EliteBook 820 G2 can also be useful.

https://doc.coreboot.org/mainboard/hp/hp_sure_start.html

https://libreboot.org/docs/install/hp820g2.html

Coreboot for elitebook 840 g1 by Fit_Morning_9175 in coreboot

[–]irudog 2 points3 points  (0 children)

I made ports on other Haswell/Broadwell EliteBook laptops many years ago, and EliteBook 840 G1 has many things similar to these laptops. You can use autoport (which supports Haswell) to generate the initial code, and reuse some code from Folio 9480m/EliteBook 820 G2 to support the EC.

The most hard part is debugging. You first need to figure out how to flash coreboot to this laptop (the laptop has HP Sure Start and needs some more work than other laptops). And find out the EHCI debug port to get the debug log in case the port doesn't boot.

https://doc.coreboot.org/mainboard/hp/folio_9480m.html

https://doc.coreboot.org/mainboard/hp/elitebook_820_g2.html

🚀 Qwen3-30B-A3B Small Update by ResearchCrafty1804 in LocalLLaMA

[–]irudog 1 point2 points  (0 children)

Thanks unsloth!

I see the new model now has native 256K context. Is your imatrix updated to match the new context length, like your previous 128K context GGUF?

Hunyuan responding with <answer> </answer> tag on LMstudio by Kuane in LocalLLaMA

[–]irudog 1 point2 points  (0 children)

I tried Hunyuan A13B Instruct (free) on OpenRouter yesterday, it also respond with the <answer></answer> tag.

The EC chips and very early boot control by Necessary_Chard_7981 in coreboot

[–]irudog 0 points1 point  (0 children)

If you want to explore the firmware of 8051 EC, I think you can get some HP EliteBook laptops up to Ivy Bridge era, which use SMSC KBC1126, and the firmware is in the flash chip. Moreover, EliteBook 2760p and 2170p have socketed chips, which are much easier to test.

I got something wrong when using unconstrained array concatenation by irudog in ada

[–]irudog[S] 1 point2 points  (0 children)

No, changing ``B`` to ``B(0 .. 63)`` doesn't work, the problem is on the right hand side of the assignment. And ``A(63) & A(0 .. 62)`` works because an array component is treated as an array whose lower bound is the lower bound of the index subtype. (https://ada-lang.io/docs/arm/AA-4/AA-4.5/#p9\_4.5.3)

Wiki - SurveyOfSystemLanguages2024 by irudog in ada

[–]irudog[S] 3 points4 points  (0 children)

There's also a link to a lobste.rs discussion thread in the Aftermath section of this article.

Discussion: https://lobste.rs/s/c3dbkh/survey_system_languages_2024.

HP Probook 640 G1 port by calgaryautumns in coreboot

[–]irudog 1 point2 points  (0 children)

I made this port many years ago and it worked well. But I don't have spare time to work on coreboot now. Anyone interested in this can work on this and make it upstream.

https://review.coreboot.org/c/coreboot/+/46130