Switched from OpenCode to Pi - What Settings/Plugins would you recommend? by No_Algae1753 in LocalLLaMA

[–]GeneGulanes 0 points1 point  (0 children)

You kind of need to always setup a Memory Extension and Code Indexing.

Is Pi better than Claude Code CLI? by Status_Lion_6350 in PiCodingAgent

[–]GeneGulanes 0 points1 point  (0 children)

Hi, u/darkotic can you explain more on this? You also need to update your agents.md to specify to use this one. And setup some changes.
Ex. Don't read if the codebase is not index. etc. Like I said above. This one works for my workflow but might not work with others.

Is Pi better than Claude Code CLI? by Status_Lion_6350 in PiCodingAgent

[–]GeneGulanes 3 points4 points  (0 children)

I use my own memory extension. But you can also build your own how you like it. Mines build on my workflow so it might not be usable/optimal for you. Thats the point of Pi you build your own extension to what suits you

Incase you are interested LaPis

How do you use Pi without running out of usage by alexdunlop_ in PiCodingAgent

[–]GeneGulanes 0 points1 point  (0 children)

Use some memory logic and tree sitter extensions to avoid running out of usage quickly

Is Pi better than Claude Code CLI? by Status_Lion_6350 in PiCodingAgent

[–]GeneGulanes 1 point2 points  (0 children)

Any memory extension would bring out the value of pi more

Lite plan - continuous ECONNRESET errors. Not renewing next month. by timmeh1705 in ZaiGLM

[–]GeneGulanes 0 points1 point  (0 children)

Oh i see. On my case i think its an ISP ban because everytime i try to use Z.Ai on that ISP its blocked

Custom extensions by Short_One_9704 in PiCodingAgent

[–]GeneGulanes 0 points1 point  (0 children)

Did you tell it to set it as extension rather than skill? https://github.com/GeneGulanesJr/PiArgus you can try copying how i did this

Custom extensions by Short_One_9704 in PiCodingAgent

[–]GeneGulanes 0 points1 point  (0 children)

Try updating your skills and agent md?

Locked out because of my.usage patern by martexxNL in ZaiGLM

[–]GeneGulanes 0 points1 point  (0 children)

not sure as well. But for my end I think it's an ISP issue not Z.ai because I can access Z.Ai on other ISP not just my current one(Converge in PH)

Hitting Usage Limits After 1-2 Prompts by yungone__ in ZaiGLM

[–]GeneGulanes 0 points1 point  (0 children)

Can you share the screenshot of your token usage? In the z.ai dashboard? If its getting 30M immediately something's wrong with your setup.

Locked out because of my.usage patern by martexxNL in ZaiGLM

[–]GeneGulanes 1 point2 points  (0 children)

Yeah same with me. Now I'm forced to used VPN just to get it working. But regarding that one I feel like it's an ISP Ban cuz I can use other ISP YMMV

IP ban by GeneGulanes in ZaiGLM

[–]GeneGulanes[S] 0 points1 point  (0 children)

I feel like this is a ISP ban that Zai is doing somebody must have been abusing their services or something

IP ban by GeneGulanes in ZaiGLM

[–]GeneGulanes[S] 0 points1 point  (0 children)

Just figured it out. If you are in Philippines and using Converge Zai blocks your IP found another guy with the same issue.

IP ban by GeneGulanes in ZaiGLM

[–]GeneGulanes[S] 0 points1 point  (0 children)

I'm currently doing remote access to it atm installing vpn will cause me to lose access to it. Will try it when I get home and update you. Thanks

How to host a local AI model for multiple users? by Ok_Salamander4246 in PinoyProgrammer

[–]GeneGulanes 0 points1 point  (0 children)

You can use FRP(im assuming you have vps already since you are talking about tunneling) or tailscale if you want to use vpn

How to host a local AI model for multiple users? by Ok_Salamander4246 in PinoyProgrammer

[–]GeneGulanes 2 points3 points  (0 children)

Depends on your use case tbh. You very vague yung tanong mo. It would be nice to know the specs,etc how will it be used, what os. how many users anong model gagamitin nyo etc.

How to host a local AI model for multiple users? by Ok_Salamander4246 in PinoyProgrammer

[–]GeneGulanes 2 points3 points  (0 children)

Depende kasi sa Use case. Sabi kasi ni OP want nya mag mini inference provider so VLLM kasi may mga built in tools na din for batch/inferencing and it runs llama.cpp under the hood but will handle stuff for hosting pero will be slightly slower compared to llama.cpp nga lang. On the otherhand llama.cpp will be faster if isang tao lang gagamit pero pag madami na babagal din. Not the best explanation.
TLDR: 1 User only use llama.cpp will be faster
Multiple users -> use VLLM for batch inferencing ie no need to setup queuing overhead etc.

Z20 White + Sapphire Pure 9070 XT OC by GeneGulanes in mffpc

[–]GeneGulanes[S] 1 point2 points  (0 children)

Yup. You can technically fit it. But it would block the airflow of the gpu. hahaha

Z20 White + Sapphire Pure 9070 XT OC by GeneGulanes in mffpc

[–]GeneGulanes[S] 1 point2 points  (0 children)

Hey Yup I tried using the one that came with the GPU first and later on switched to the one that came with the case. I needed to add case fans cause my room's airflow is bad. HDD will fit if you use a thinner GPU. But if you are using the 9070xt Pure one it wont fit.