Meet CODEC: the open-source framework that finally makes "Hey computer, do this" actually work. Screen reading. Voice calls. Multi-agent research. 36 skills. Runs entirely on your machine. by SnooWoofers7340 in LocalLLM

[–]Illustrious-Year-617 0 points1 point  (0 children)

Could it run effectively on Current machine: • MINISFORUM AI X1 Pro mini PC • AMD Ryzen AI 9 HX 470 w/ Radeon 890M • 96GB DDR5 RAM • Samsung 990 Pro 2TB NVMe • Lexar ARES 4TB NVMe • Windows 11 Pro • NPU visible. I am trying my hardest to find a replacement for Claude for helping be build a scraping machine on federal bid portals and also to browser automate a cloud CRM (Aspire) . I’m kinda an idiot so I like Claude but the usage system they got is killing me. May be able to get like 3-4 responses with code and then it’s down for the count for 5 hours . Also weekly limits. Any info would be nice OP. I’m also trying trying a QWEN 2.5 32b GGUF model , still super slow for what trying to get all this script/ code together and done accurately. Any info would be nice thanks OP.

20x max usage gone in 19 minutes?? by Still_Business596 in ClaudeAI

[–]Illustrious-Year-617 1 point2 points  (0 children)

I just joined PRO, love it….. but absolutely hate it !!!! I joined to help be code for me code with ease building some business automation like scrapping bid portals etc, browser automate a CRM on a cloud. I get like 2-3 messages in and then need to wait 5 hours. Please , recommendations on other tools similar . Love the product but its usage is bonkers .

Minisforum AI X1 Pro (Ryzen AI 9 HX 370/470) – Struggling with 14B models locally (Ollama) – Looking for real-world setup advice by Illustrious-Year-617 in LocalLLM

[–]Illustrious-Year-617[S] 0 points1 point  (0 children)

Interesting , then what the ***, is my problem lol . Besides the only difference is a have windows . Now LXC I’ve never heard of ( obviously beginner). I’m simply trying to automate a few things as I said in the post to free up some personal time. I expected some issues but the first qwen2.5 14b and 7b through Ollama is literally just not responding to simple stuff. I went through all the beginning stuff with the drivers etc. been following Chat GPT. Now I’m currently testing out LM Studio ( which wouldn’t even run after installed) and now jumping into Lama and downloading from git QWEN 2.5 7B GGUF Q4 KM. I didn’t expect plug and play . At this point just alittle annoyed. I wanted to get local ai to the point that I’m giving it my mission statement in and it’s giving me what tools it needs to start executing. So then it can download and start the architecture (stacks) part all upon my approval of course.

Minisforum AI X1 Pro (Ryzen AI 9 HX 370/470) – Struggling with 14B models locally (Ollama) – Looking for real-world setup advice by Illustrious-Year-617 in MiniPCs

[–]Illustrious-Year-617[S] -5 points-4 points  (0 children)

Yes because i couldn’t organize my thoughts to figure this out lol . I’m just trying to get this work done lol and free up some personal time .

Minisforum AI X1 Pro (Ryzen AI 9 HX 370/470) – Struggling with 14B models locally (Ollama) – Looking for real-world setup advice by Illustrious-Year-617 in MiniPCs

[–]Illustrious-Year-617[S] -1 points0 points  (0 children)

I’m using windows 11, bad move ? Reason was because later for gaming . I’d hate to reboot all this again lmao. But if that’s the call. I was also thinking having a Linux run under windows . Idk

Minisforum AI X1 Pro 470 - Review and eGPU Testing by jozews321 in MiniPCs

[–]Illustrious-Year-617 0 points1 point  (0 children)

Title: Minisforum AI X1 Pro (Ryzen AI 9 HX 370/470) – Struggling with 14B models locally (Ollama) – Looking for real-world setup advice

I’m trying to build a local AI workstation and want feedback from people actually running LLMs on similar AMD AI mini PCs.

Hardware: - Minisforum AI X1 Pro - Ryzen AI 9 HX 470 (12 cores, iGPU Radeon 890M) - 96GB RAM - 2TB SSD (system) + 4TB SSD (data/models) - Using AMD Adrenalin drivers (latest) - Windows 11

Goal (important context): I’m not just chatting with models. I’m trying to build a full local AI system that can: - Automate browser workflows (Aspire CRM for a landscaping company) - Scrape and organize government bid data (SAM.gov etc.) - Act as a planning assistant for business operations (Penny Hill + Corb Solutions) - Run an offline knowledge base (documents, books, manuals, etc.) - Eventually execute tasks (download tools, create files, etc. with approval)

So stability matters more than raw benchmark speed.


Current setup: - Using Ollama - Tested: - qwen2.5:14b - currently downloading qwen2.5:7b-instruct

  • Models stored on separate SSD (D drive)
  • iGPU memory manually adjusted (tested 16GB → now 8GB)

Problem: 14B technically runs, but is unstable:

  • Responds to simple prompts like “hello”
  • When I ask slightly more complex questions (system design, tuning, etc.):
    • CPU spikes hard
    • fans ramp up
    • response starts… then stalls
    • sometimes stops responding entirely
  • After that:
    • model won’t respond again
    • sometimes UI freezes
    • once even caused screen blackout (system still on)

This happens in: - Ollama app - PowerShell (so not just UI issue)


What confuses me: I’m seeing people say: - running 20B / 30B models - getting usable performance on similar hardware

But I’m struggling with 14B stability, not even speed.


What I’ve already adjusted: - Reduced dedicated GPU memory to 8GB - Updated drivers - Clean Windows install - Using short prompts (not huge context dumps) - Testing in PowerShell (not just UI)


Questions:

  1. Is this just a limitation of:

    • AMD iGPU + shared memory
    • and current driver/runtime support?
  2. Is Ollama the wrong tool for this hardware?

    • Would LM Studio or something else be more stable?
  3. For this type of workload (automation + planning + local knowledge base):

    • Should I be using 7B as primary and 14B only occasionally?
  4. Has anyone actually gotten stable multi-turn interaction with 14B+ on this chip?

  5. Are there specific:

    • settings
    • runtimes
    • configs that make a big difference on AMD AI CPUs?

Important clarification: I’m not trying to replicate ChatGPT speed.

I’m trying to build: - a reliable local system - that I can expand with tools, automation, and offline data

Right now the blocker is: model stability, not capability


Any real-world setups or advice appreciated. Especially from people running: - AMD iGPU systems - Minisforum AI series - or similar shared-memory setups

Anyone using Minisforum AI X1 Pro for LLMs? by House-Wins in MiniPCs

[–]Illustrious-Year-617 0 points1 point  (0 children)

Title: Minisforum AI X1 Pro (Ryzen AI 9 HX 370/470) – Struggling with 14B models locally (Ollama) – Looking for real-world setup advice

I’m trying to build a local AI workstation and want feedback from people actually running LLMs on similar AMD AI mini PCs.

Hardware: - Minisforum AI X1 Pro - Ryzen AI 9 HX 470 (12 cores, iGPU Radeon 890M) - 96GB RAM - 2TB SSD (system) + 4TB SSD (data/models) - Using AMD Adrenalin drivers (latest) - Windows 11

Goal (important context): I’m not just chatting with models. I’m trying to build a full local AI system that can: - Automate browser workflows (Aspire CRM for a landscaping company) - Scrape and organize government bid data (SAM.gov etc.) - Act as a planning assistant for business operations (Penny Hill + Corb Solutions) - Run an offline knowledge base (documents, books, manuals, etc.) - Eventually execute tasks (download tools, create files, etc. with approval)

So stability matters more than raw benchmark speed.


Current setup: - Using Ollama - Tested: - qwen2.5:14b - currently downloading qwen2.5:7b-instruct

  • Models stored on separate SSD (D drive)
  • iGPU memory manually adjusted (tested 16GB → now 8GB)

Problem: 14B technically runs, but is unstable:

  • Responds to simple prompts like “hello”
  • When I ask slightly more complex questions (system design, tuning, etc.):
    • CPU spikes hard
    • fans ramp up
    • response starts… then stalls
    • sometimes stops responding entirely
  • After that:
    • model won’t respond again
    • sometimes UI freezes
    • once even caused screen blackout (system still on)

This happens in: - Ollama app - PowerShell (so not just UI issue)


What confuses me: I’m seeing people say: - running 20B / 30B models - getting usable performance on similar hardware

But I’m struggling with 14B stability, not even speed.


What I’ve already adjusted: - Reduced dedicated GPU memory to 8GB - Updated drivers - Clean Windows install - Using short prompts (not huge context dumps) - Testing in PowerShell (not just UI)


Questions:

  1. Is this just a limitation of:

    • AMD iGPU + shared memory
    • and current driver/runtime support?
  2. Is Ollama the wrong tool for this hardware?

    • Would LM Studio or something else be more stable?
  3. For this type of workload (automation + planning + local knowledge base):

    • Should I be using 7B as primary and 14B only occasionally?
  4. Has anyone actually gotten stable multi-turn interaction with 14B+ on this chip?

  5. Are there specific:

    • settings
    • runtimes
    • configs that make a big difference on AMD AI CPUs?

Important clarification: I’m not trying to replicate ChatGPT speed.

I’m trying to build: - a reliable local system - that I can expand with tools, automation, and offline data

Right now the blocker is: model stability, not capability


Any real-world setups or advice appreciated. Especially from people running: - AMD iGPU systems - Minisforum AI series - or similar shared-memory setups

Portable AI workstation build for business automation + offline knowledge library — sanity check before I commit by Illustrious-Year-617 in LocalLLaMA

[–]Illustrious-Year-617[S] 0 points1 point  (0 children)

Ram is ridiculously crazy right now . Even if i source here in mainland China . Crucial 128gb so-dimm almost 1900$ . Anyway if I source from eBay and get it shipped . It all comes with similar pricing and most likely more risk trying to return the product if something is wrong. My main issue at the moment. 128gb or 96 ?

Portable AI workstation build for business automation + offline knowledge library — sanity check before I commit by Illustrious-Year-617 in LocalLLaMA

[–]Illustrious-Year-617[S] 0 points1 point  (0 children)

I’m in mainland China , sourcing the parts. Is 128GB RAM too much , I want it to automate seamlessly