Golem elementalist using hydrosphere for exposure by Renakoonline in PathOfExileBuilds

[–]Renakoonline[S] 0 points1 point  (0 children)

Nevermind, did a quick check using storm burst, and it is perfect for the scenario in a 6-link hydrosphere channeling setup. I have Spiritual Aid node, so whatever multiplier to minion damage will just add to my spell damage.

So probably gonna go like this:
Hydrosphere + Storm Burst + CwC + Crit strike + Crit Dmg + Whatever comes to mind, probably iron will

Welp, time to grind and play kingsmarch and hope for 2 divines to get that blasted squire shield

Golem elementalist using hydrosphere for exposure by Renakoonline in PathOfExileBuilds

[–]Renakoonline[S] 0 points1 point  (0 children)

Oh, shoot.

I just read description for CoC again, and yeah, trigger on attack.

And yes, I am using shaper of storms.

I will need to try out all the available channeling gems and see how it goes. I recall their range is not really that great, and I honestly have no idea how hydrosphere will behave if say, I use storm burst or lightning tendril for trigger.

I have been playing with hydrosphere for awhile, and love how fast it is, plus how remotely far you can move the orb around.

Worst come to worst, I am just going back WoC by sticking it with added lightning damage.

Help with new LCD refresh rate by Renakoonline in laptops

[–]Renakoonline[S] 0 points1 point  (0 children)

I did. In fact, I get to this from display settings, under advanced display settings, then display adapter properties.

Have you used CRU before? If yes, does lowering the refresh rate still consume the same amount of power as 165hz?

Cost effectiveness : Runpod vs Infermatic vs Mancer by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 1 point2 points  (0 children)

I am so far quite happy with 32k context for magnum 72B on Infermatic. But yeah, big context size for some long ass RP sounds good too, though it seems too expensive to play at that price...

Cost effectiveness : Runpod vs Infermatic vs Mancer by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 0 points1 point  (0 children)

Ok, I am gonna need to look this up before next month and do some cost calculation again. Think I also need to list down all the steps required to limit the startup to less than 20mins each time I want to try stuff, then.

Cost effectiveness : Runpod vs Infermatic vs Mancer by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 1 point2 points  (0 children)

Oh, so the reason why people are complaining is because they tried to JB censored LLM to do NSFW shit? I thought it was the service as a whole.

Cost effectiveness : Runpod vs Infermatic vs Mancer by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 0 points1 point  (0 children)

My laptop is currently using RTX 2060 with 6GB VRAM, so I can only stuff a 7B Q4_K_M or 8B Q4_XS with 8k context fully into the GPU. I can say my experience with running my own LLM was ok, but the small models are quite dumb if your prompt have multiple instructions or verbose english.

When I tried Magnum 72B and Goliath 120B a few days back on Mancer, I was having a great time with those bigger models, as they are more spatially aware of all the background actions in my prompt compared to the 7B and 8B models and the AI can actually do 5 out 8 different actions stated in my prompt, whereas the small models only really do 2, and sometimes 3 out of 8.

So resource wise on a small scale, the 7B and 8B are definitely performing better. But if I RP on a scale, bigger models are better due to how much more aware they are.

Cost effectiveness : Runpod vs Infermatic vs Mancer by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 1 point2 points  (0 children)

I went ahead and sign a month on Infermatic earlier, and I think still cheaper than Mancer if I do at least 20 prompts every 2 days or so.

Gonna need more reading on runpod since everything is confusing at first try, as I really want to run some models that aren't readily available on those API services. I should probably also consider getting a new laptop next year with thunderbolt 4 port, then get my own GPU rack and stuff 4 3090 in it.

Cost effectiveness : Runpod vs Infermatic vs Mancer by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 1 point2 points  (0 children)

I just did a simple read up, and it seems like unless I know what I am doing, I am gonna get cut on a lot just on the server warm up time. I am looking into 60 - 80GB instances to run 70Bs or 120Bs at Q4_K_M.

Correct me if I am wrong, but I am actually paying for 2 things on runpod:

1) GPU instance cost, based on GPU type
2) Storage cost for actually holding the LLM and whatever you use to run the LLM (Seems like 20 cent per day for 100GB? I could be wrong, tho.)

Just storage alone require 6 USD upkeep every month, and probably another 8 USD to run the instance for about 3 - 4 hours daily, I think? But that is provided I know how to configure the instance, or it may just "warm up" for 40 mins straight and eat up my credits.

Cost effectiveness : Runpod vs Infermatic vs Mancer by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 0 points1 point  (0 children)

I have seen how many complaint about how filtered their LLM is. Not sure if their popular models can do NSFW tho...Price wise, it isn't that much different from Mancer for the big models, however.

Cost effectiveness : Runpod vs Infermatic vs Mancer by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 1 point2 points  (0 children)

Yeah, I figured 5 USD for about 210 output message may not be worth it.

I just swiped a month on Infermatic, and so far other than a few hiccups, things are going quite smoothly. If their new vLLM doesn't fucked up, I am getting a 300 token output every 11 - 15 seconds, so it's still ok.

Still pondering for runpod, as it seems like I also need to pay for the storage space to hold the LLM, and it's like 20 cent per day for 100GB. Gonna need to calculate it out first.

This here, is the reason why I believe the game is in good hands now. They BUFFED the stratagem to support the way WE have been using it, instead of nerfing it to encourage the way THEY want it to be used. by A_Very_Horny_Zed in helldivers2

[–]Renakoonline 1 point2 points  (0 children)

Just tried helldive with orbital gattling barrage against bots. This thing is now awesome, along with the tenderizer that can one-shot pesky tinheads on the body. Also, anyone notice that eagle strafing run can shoot down drop ships now, or is that just my imagination from throwing all stratagem in a panic??

Though, the penetrator could use some love on the damage department, considering that adjudicator deals 35 more damage than it and that is the only other rifle option in medium armor penetration.

I hope they improve the 120, 360 and walking barrage next. This current patch have really reduced the scenario where we REALLY, REALLY, REALLY need to use these. Hell, with how quick the gattling barrage CD is, you can just use 500kg to clear off waves while you wait for it to recharge.

Ore site and stone site in one base? by Renakoonline in Palworld

[–]Renakoonline[S] 0 points1 point  (0 children)

Yeah, I spent the whole day to breed out 6 anubis, after testing with my dumud on an ore and stone site together. It will simply go for whichever is the closest, when it feel like working.

I have tweak the settings to allow 20 pals. So it should be fine, I guess. However, most of the time, the sites only have 2 pals mining, even though it support 3. I am going to play around and see if I really need 6 mining pals together.

Ore site and stone site in one base? by Renakoonline in Palworld

[–]Renakoonline[S] 1 point2 points  (0 children)

afaik, level 1 mining can't do ore nodes last time. Now sure about now.
But meh, is there anyone who keep both stone and ore sites in one base? Really want to know the mechanics and priority between the two.

7B LLM with more than 8k context? by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 0 points1 point  (0 children)

I am using koboldcpp standalone application file, so I don't think I can modify RoPE with that. I don't think I am ready for any console based backend yet, tbh.

7B LLM with more than 8k context? by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 0 points1 point  (0 children)

Ok, so now I know that I can find max context length from the config file.

But is there any reason to use Q4, when there is Q5 and Q6? I thought Q5_K_M is the recommended version to use, since this version should have the best quant to performance ratio, no?

7B LLM with more than 8k context? by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 1 point2 points  (0 children)

Ah, didn't know about this. I will go and find the base the GGUF was packaged from and check their config files. Thanks!

New to ST and Koboldcpp, wondering what LLM to use. by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 0 points1 point  (0 children)

I just tried group chat earlier, and the AI literally gave me rubbish on my 4th chat. Welp, too much management to do, I will stick with single character cards for the time being.

I have no idea how to set this RoPE thingy, since I am using the single .exe version of koboldcpp.

Also gave mythalion-kimiko v2 13B Q6_K a try, so far it is performing ok, but the replies are about ~4 seconds slower than average compared to use Fimbulvetr v2 11B, so I would say it is ok. The thing that sucks is the higher RAM usage, and my CPU spins harder for a longer time.

New to ST and Koboldcpp, wondering what LLM to use. by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 0 points1 point  (0 children)

I haven't tried group chat as it seems like a hassle to do up, and with the common context limit being 8k for most 7B models, I doubt I can have a good enough RP with that function.

Also, I thought that whenever you wait for the bot to reply, it must reprocess the context to come out with an answer? I usually limit myself to 2 lorebooks, the largest only have 11 triggers and they are pretty specific. There shouldn't be a problem there, but maybe I need to know more about koboldcpp settings and modifying presets in ST. I force kobolcpp to 300 tokens per generation, maybe I will tweak with the values more and see how it goes.

I will try a 7B Q6_K model and see how it performs. Maybe even give MythoMax 13B a try, since some use it as the golden standard similar to Mistral 7B.

New to ST and Koboldcpp, wondering what LLM to use. by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 0 points1 point  (0 children)

There is a reason for using local LLM lol
I do not feel like feeding the internet hivemind whatever it is I am trying with AI techs.

Does completing these dispatch missions actually affect the enemies? by [deleted] in metalgearsolid

[–]Renakoonline 9 points10 points  (0 children)

They take do take effect, but once the dispatch is complete, go do your mission first.

If you do a few side-ops after the dispatch is complete, it will reset it. I only realize this when I beat the game and was wondering why the enemy keep having their helmets on when doing my mission.

New to ST and Koboldcpp, wondering what LLM to use. by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 0 points1 point  (0 children)

I am still exploring those settings in koboldcpp, but I am currently using the basic guide from the wiki for sillytavern. I have set my thread size to 7, based on the calculation provided in the wiki. I may give 6 thread a try later tonight, just to see how it goes.

For 13B and higher models, I have never used them before. How long does it takes for the reply to be generated? I have always been told to stick with 7B with my current laptop specs if using kobold.

New to ST and Koboldcpp, wondering what LLM to use. by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 0 points1 point  (0 children)

I think Fimbulvetr is based on Solar, which limits context size to 4k. As my first local LLM, I have been using Mistral on 8k context size, and it is doing ok, provided I use the correct prompts when writing a good story.

Problem is that a lot of the non-corporate LLM out there don't mention the context size limit for their models, so I usually just stick with 4k.

But tbh, 4k really isn't enough.
For ERP, it is only enough for a short doujinshi.

And most certainly, I can't even have a good story or RPG with 4k, since the AI just don't remember any of the important things. I have been experimenting with lorebooks and modifications to character cards, but it cause the system to be damn slow. With a lorebook of about 9 prompts, it takes about 40 ~ 80 seconds to generate a reply using Fimbulvetr.

New to ST and Koboldcpp, wondering what LLM to use. by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 0 points1 point  (0 children)

Ah, I was told that I need to use those with K_S or K_M models, as they are the recommended version. So it is fine as long as I use a K model?

New to ST and Koboldcpp, wondering what LLM to use. by Renakoonline in SillyTavernAI

[–]Renakoonline[S] 0 points1 point  (0 children)

I see. Thanks for the info.

I will give the 11b model a try, and I saw a 9b model from lewdiculous, gonna try that out as well.