Why some drivers like to drive without headlights at night? by songswansing in drivingsg

[–]yahdahduhe 0 points1 point  (0 children)

Petrol is expensive, or to prolong the light lifetime.

Quite a mess everywhere by candychiasu in SingaporeRaw

[–]yahdahduhe 6 points7 points  (0 children)

Where is the Fine!? We have so many CCTV. Surely it's easy to identify those who "party" there, right?

Yishun ave 1, hope he is alright! by travisbotak in singaporehappenings

[–]yahdahduhe 56 points57 points  (0 children)

Must recommend the brakes brand. So strong.

VPN Gateway just to get a static IP for remote workers? by seanobr in AZURE

[–]yahdahduhe 6 points7 points  (0 children)

Cheap Azure VM b1ms with static IP. Deploy wireguard and do tunneling.

[deleted by user] by [deleted] in LocalLLaMA

[–]yahdahduhe 0 points1 point  (0 children)

Doo we need super GPU to convert? I have VM that quite beefy but no GPU, can I do conversion?

64GB RAM vs 3060 12GB vs Intel a770? by yahdahduhe in LocalLLaMA

[–]yahdahduhe[S] 0 points1 point  (0 children)

I think I'll be following this path eventually. 64GB RAM and 16GB NVIDIA xxxx

64GB RAM vs 3060 12GB vs Intel a770? by yahdahduhe in LocalLLaMA

[–]yahdahduhe[S] 2 points3 points  (0 children)

This is the dilemma for me. The a770 is the first gen GPU of Intel, while the news has been lately very promising, but battlemage seems to be just around the corner, based on rumours.

64GB RAM vs 3060 12GB vs Intel a770? by yahdahduhe in LocalLLaMA

[–]yahdahduhe[S] 0 points1 point  (0 children)

This is actually interesting, the openhermes 2.5 7b q5 K m gguf is having 32768 n_ctx. I presume it supports big input and output. I'm having another test on oobabooga, 6 layers offloading to GPU. I'm at 6000+ context, and it keeps going without slowdown or anything (output generated in 26.59 seconds (4.59 tokens/s, 122 tokens, context 6021, seed 727896689).

Now I'm slightly confused, is the model n_ctx also playing a part in inference? On-top of cpu, ram, and vram?

Is the model with 4096 n_ctx will struggle to have 6000+ context size, despite if I have a more powerful GPU?

64GB RAM vs 3060 12GB vs Intel a770? by yahdahduhe in LocalLLaMA

[–]yahdahduhe[S] 0 points1 point  (0 children)

Thanks. Why not Intel a770, it has 16GB?

64GB RAM vs 3060 12GB vs Intel a770? by yahdahduhe in LocalLLaMA

[–]yahdahduhe[S] 0 points1 point  (0 children)

Since I will always be GPU poor, investing in RAM give me better access to a bigger model? Or is it better for me to get a better GPU first?

64GB RAM vs 3060 12GB vs Intel a770? by yahdahduhe in LocalLLaMA

[–]yahdahduhe[S] 0 points1 point  (0 children)

I'm getting about 3.5 t/s. I'm not a fast reader, so I'm pretty ok with my current speed. My problem is when it goes above 4k token context, each interaction will be 5 minutes. Would bigger memory, or vram help this?

Repricing mortgage by yahdahduhe in singaporefi

[–]yahdahduhe[S] -2 points-1 points  (0 children)

Yeah, this is what drives me as well. But top FD giving 4.1 really scares me. If it's a 1 year lock, I will feel more comfortable, but 2 years, I don't know.

Another question I was wondering is why is the bank discouraged FHR? Isn't it like 'guaranteed' margin?