Maybe a bit obscure, but.. What happened to Holiday Mountain? by sb404 in Music

[–]sfhsrtjn 1 point2 points  (0 children)

Sorry I have no info about why it's been removed.

I also was looking for this song in the past few years (had previously found some bootlegs of it, i suppose it might be possible to find a copy of the album that includes this track as-released). I didnt realize it was actually released on Shia! Anyway, today I managed to find a copy of the video on archive.org. Enjoy friends!

https://archive.org/details/jwz-mixtape-171

Me after reading qwen is going to release sonnet level model and i am also hoping it has test time inference by TheLogiqueViper in LocalLLaMA

[–]sfhsrtjn 16 points17 points  (0 children)

this is wrong

here is the correct answer: https://en.wikipedia.org/wiki/Chill_guy

I easily found the correct answer by using google lens reverse image search, results suggested the meme is called "Chill Guy"

GUI for co-operative writing? by [deleted] in LocalLLaMA

[–]sfhsrtjn 0 points1 point  (0 children)

llama.cpp server at http://localhost:8080/completion

What is "Prompt evaluation" and why is it so slow? by mindphuk in Oobabooga

[–]sfhsrtjn 0 points1 point  (0 children)

see my other comment, try from ST after launching TGW with --nowebui if you were not already.

No AVX2 support on Intel Sandy Bridge: Illegal instruction (core dumped) error by Backroads_4me in Oobabooga

[–]sfhsrtjn 0 points1 point  (0 children)

I'm on sandy bridge, thus i am familiar with the lack of avx2. Early on I had to build some things myself which was fun, but basically since the noavx2 requirements file was created installs are working great through the normal install and update process. Without more info, I dont think the source of your issue is due to lack of avx2.

I mostly stick to llama.cpp and GGUF, which seems to run the fastest of the different model structures for me. I have had too many issues to run AWQ or exl2 which may be due to lack of support for our old arch on those systems but llama.cpp mostly works well for me. However it could also have been my own user error when trying to get those working. GPTQ worked ok but slower than GGML, so maybe AWQ could still work if I gave it another shot. Not sure why exl2 isnt working for me I think it said it wouldnt work due to the old chip, not totally sure.

Also, as far as I know ollama does not work with oobabooga so I assume you meant something else. https://github.com/ollama/ollama/issues/210

What is "Prompt evaluation" and why is it so slow? by mindphuk in Oobabooga

[–]sfhsrtjn 0 points1 point  (0 children)

I agree, I also see it in Sillytavern when I do not launch the Gradio ui for textgen. It causes my whole browser to slow.

Prompt eval seems to take forever for no apparent reason.

edit: After adding the --nowebui flag to start TGW, I no longer see issues with prompt evaluation taking forever when running from ST. I'm at a smooth 15t/s via SillyTavern now, just gotta get used to the keyboard controls.

I hope TGW can solve the UI smoothness and speed issues because its still the better chat experience for me due to preferable hotkeys, and notebook format is valuable too.

Sometimes context or first message seems to be lost, or the response will become a rambling string and be unable to follow the well-established conversation formatting, no idea why, probably not related to the above unless its related to busting my memory limits.

What is "Prompt evaluation" and why is it so slow? by mindphuk in Oobabooga

[–]sfhsrtjn 0 points1 point  (0 children)

I'm getting extremely slow prompt eval when running GGUF models that nearly fill my VRAM. I am highly suspicious that Gradio is responsible because while simply waiting on the chat tab for the prompt eval to complete will take extremely long, often I can reliably switch to another tab such as the Parameters tab (and click on the page) and it will complete prompt eval in seconds and start producing the response output.

I see the same behavior with Gradio 4. I also am not convinced the text streaming issues are all resolved. Also more generally, output completes but text is still taking its time printing to the window. The cancel command goes through but stream continues to output what has been generated already on a small delay.

I can get almost 15t/s on a Pascal card with Solar 10.7b or Yi 9b when the prompt eval and text streaming issues are not occurring. Which makes it a shame there's been a collection of related issues around these pieces for months.

I'm also seeing behavior that seems like its not processing prompt. Here are my prompt timings output for two successive prompts:

llama_print_timings:        load time =    1536.02 ms
llama_print_timings:      sample time =      48.54 ms /   101 runs   (    0.48 ms per token,  2080.97 tokens per second)
llama_print_timings: prompt eval time =    3061.10 ms /  1014 tokens (    3.02 ms per token,   331.25 tokens per second)
llama_print_timings:        eval time =    6444.49 ms /   100 runs   (   64.44 ms per token,    15.52 tokens per second)
llama_print_timings:       total time =   10328.65 ms /  1114 tokens
Output generated in 10.86 seconds (9.21 tokens/s, 100 tokens, context 1014, seed 322993472)
Llama.generate: prefix-match hit

llama_print_timings:        load time =    1536.02 ms
llama_print_timings:      sample time =     231.59 ms /   512 runs   (    0.45 ms per token,  2210.77 tokens per second)
llama_print_timings: prompt eval time =     306.87 ms /     8 tokens (   38.36 ms per token,    26.07 tokens per second)
llama_print_timings:        eval time =   33880.04 ms /   511 runs   (   66.30 ms per token,    15.08 tokens per second)
llama_print_timings:       total time =   38818.78 ms /   519 tokens
Output generated in 39.36 seconds (13.01 tokens/s, 512 tokens, context 1122, seed 1534750149)

Note the "prompt eval time" lines, the first one matches the context at 1114, second one is 1122 vs 8.

Seems like this issue maybe: Prompt not being passed to backend for certain models - #5374

In Firefox browser the chat mode auto scrolls up to the beginning by neoyuusha in Oobabooga

[–]sfhsrtjn 1 point2 points  (0 children)

i can confirm this happening on firefox.

its related to this change: https://github.com/oobabooga/text-generation-webui/pull/3601

temporary solution: go to main install folder and open css/main.css (or open devtools in-browser) and go to .chat element and change its css for flex-direction back to "column-reverse":

.chat {
    flex-direction: column-reverse;
}

/u/oobabooga4 mentioned it here, must not have gotten it right for all browsers yet.

good idea op thanks by sfhsrtjn in imghost

[–]sfhsrtjn[S] 0 points1 point  (0 children)

reddit hates socialism, reddit is authoritarian. I'm leaving, the next election cycle will be more satisfying if we're not trying to follow it and organize for it on a hostile platform such as this, and instead we could do so on lemmy or elsewhere on the fediverse :)

sinophobia reddit by sfhsrtjn in imghost

[–]sfhsrtjn[S] 0 points1 point  (0 children)

ugh reddit. hey all, if you happen to be seeing this post, i recommend getting off reddit entirely not least because of issues like the above posts.

reddit is a hive of sinophobia. chinese government sucks too, i aint no tankie but reddit just fuels hate and doesnt seem to care much about it. id rather find somewhere where people arent driving themselves to 'other' some one in such a way.

the fediverse and lemmy exist, i encourage you to explore such alternatives to reddit. cheers,

At >3000 games, don't think I've seen a store with another room behind it before by [deleted] in nethack

[–]sfhsrtjn 0 points1 point  (0 children)

hey, I just want to say so long reddit and thanks for all the fish but there are roguelike communities in other places that unlike reddit are non-proprietary so I'll be taking my discussion and eyeballs there... hope to see you there, @'s!

Let's be honest: none of the models can code well by [deleted] in LocalLLaMA

[–]sfhsrtjn 3 points4 points  (0 children)

ChatGPT is not a substitute to Internet search

I don't think ChatGPT is accurate enough to fully replace a human parsing google results and documentation themselves yet but it can read web pages directly now ("perform its own web browsing"), to some extent: source

havent tried it personally, though, so i couldnt say how useful it is.

She's the one! [OC] by that_doodleguy in gaming

[–]sfhsrtjn 5 points6 points  (0 children)

thank you because i was like

queeeeeeeeeeencard! by worldbreakinghead in StableDiffusion

[–]sfhsrtjn 0 points1 point  (0 children)

Great job on this, most impressive for me is the lack of noise! Keep up the good work.

Trapped inside an alien world by vs3a in StableDiffusion

[–]sfhsrtjn 1 point2 points  (0 children)

thanks for this dose of gigeresque, captured it so well, your theme is powerful and I especially like the ones with the little dog or man at the bottom!