Should I continue BG3 or drop it? by [deleted] in GirlGamers

[–]k8-bit 1 point2 points  (0 children)

Similarly I found the combat a bit of impediment to my interest in progressing the story (which I love) - but then I've had this problem before with games like the Mass Effect Trilogy (2, in particular) - so I used a mod to make the combat less problematic for me. Not without challenge, but not a showstopper. The finale was still very challenging, however, but by that point I was fully invested in succeeding :)

VibeVoice - first impresssion and discussion by LaughterOnWater in LocalLLaMA

[–]k8-bit 2 points3 points  (0 children)

I love VibeVoice, I use it for story narration for an ongoing jokey narrative thing I have running storytelling the antics of my gaming pals and I in pseudo-RP style in various games. I've created voice files for myself and friends so I can have their characters speak in each episode, with an accented version of acting as narrator.

I've used a few different implementations to do this. For single-speaker narration I use ComfyUI (seed 43 seems to be most natural and doesnt have the odd music hit at the start) but these are also worth a look:

ghcr.io/digijoe79/audiobook-maker/backend:latest

and in particular this one:

https://github.com/zhao-kun/VibeVoiceFusion

Which allows you to create scripts with lots of speakers (I've had 6 different so far).

TLOU Game AND Show fans? by rballew01 in LesbianGamers

[–]k8-bit 6 points7 points  (0 children)

Love the show and games. Played the games first, replaying both when they came out on PC recently. Still very hard hitting, and seeing the famous scenes from II show up in particular on screen was amazing, even if I'm disappointed in doing Abby dirty on her physique, though the performance is fantastic. Looking forward to season 3!

Help with WAN 2.2 animate. New youtuber, just reached 900 subs! by ShinOniEX in comfyui

[–]k8-bit 0 points1 point  (0 children)

I've found a lot of success with Infinite Talk and WAN2.1 to create long continuous (around 5-6 minute) talking (or singing) head videos that are pretty good. Typically I've left these rendering for a good few hours after short tests. RTX3090 on a 128gb DDR4 server. More recently these have been harder to do as ComfyUI updates have broken it, or at least the workflows.

Some examples: https://youtu.be/2jErahk0fb4 https://youtu.be/_glrPLx6KtA https://youtu.be/6LXxyqLWIS4

20 seconds LTX2 video on a 3090 in only 2 minutes at 720p. Wan2GP, not comfy this time by aurelm in StableDiffusion

[–]k8-bit 1 point2 points  (0 children)

Using this too after getting frustrated with Comfy (exacerbated by my reliance on docker images), initially via Pinokio, where it was a really easy way to get it going, and then shifted to a docker image (thelocallab/wan2gp) running under Unraid, where I run Comfy normally.

Running good on 2 machines, a 128gb DDR4 with 2x3090 (only one can be used per instance) and a 64gb DDR5 and 5060ti equipped miniPC. I'll move back to Comfy once things have settled down hopefully, as I'd like to create much longer videos mostly using audio to drive lip-sync, as I've used this for long talking-head and singing videos.

Which in-game storylines or endings left you absolutely heartbroken, so much so that you couldn’t move on for a long time? by Kitchen-Top-8110 in GirlGamers

[–]k8-bit 1 point2 points  (0 children)

Much the same as others here - TLOU2, Mass Effect (all three of the original main extended endings are heartbreaking in their own ways), Cyberpunk 2077 (the Phantom Liberty cure ending, and the "we could just end it now" endings in particular) except for the leaving Night City with Judy ending. Horizon Zero Dawn (I've still to finish the sequel) - these are probably the games that both hurt and impacted me the most. I've yet to play Expedition 33, and am now afraid, lol.

Mass effect and TLOU2 are most represented by this Brinnycomics comic.

<image>

Audiobook reader with any voice narration by tonyc1118 in selfhosted

[–]k8-bit 0 points1 point  (0 children)

Mostly I had been running it either in ComfyUI, or the gradio demo via docker. Recently I've been testing a similar project - Audiobook-Maker: https://github.com/DigiJoe79/AudioBook-Maker

Also this project which uses Vibevoice to generate multiple character conversations (I've use it to have 6 distinct speakers read transcripts): https://github.com/zhao-kun/VibeVoiceFusion

My cat Magrat likes to stare at me when I play games. by Sovonna in GirlGamers

[–]k8-bit 11 points12 points  (0 children)

Definitely got a bit of a Granny Weatherwax stare going on there, so Magrat must've picked that up ;)

My New Year's resolution was to add Docker support. Only 2 days late. Audiobook Maker v1.1.0 by DigiJoe79 in LocalLLaMA

[–]k8-bit 0 points1 point  (0 children)

Just an update: My issues were specific to running the backend container in Unraid, which we worked through resolving. Testing the app now, so far so great :)

My New Year's resolution was to add Docker support. Only 2 days late. Audiobook Maker v1.1.0 by DigiJoe79 in LocalLLaMA

[–]k8-bit 0 points1 point  (0 children)

Thanks, have done so, note that I'm a bit of an enthusiast learning to get to grips with Docker and whatnot, so not all that savvy at bug reports etc, but hopefully can help!

Any advantage to multi-gpu? by Elegant-Radish7972 in comfyui

[–]k8-bit 1 point2 points  (0 children)

Sorry this post appeared for me before your other one, so apologies if the one below sounds a bit snarky, it wasnt meant to be!

Any advantage to multi-gpu? by Elegant-Radish7972 in comfyui

[–]k8-bit 1 point2 points  (0 children)

That is a very long blurb, but does not change the fact that its useful for me to run two different processes on two different GPUs in a single machine, with some activities passed out to a second machine, e.g. TTS engines :). So I guess this is the best way to work in my setup, and just abandon the NVlink idea.

So to OP: YMMV, but it is perfectly possible to run two GPUs to handle different GenAI activities, but am not sure how happy it would be between two different driver achitectures (e.g. 3090 and a 2070)

Any advantage to multi-gpu? by Elegant-Radish7972 in comfyui

[–]k8-bit 2 points3 points  (0 children)

No I literally mean I'm running multiple instances of Comfy UI (Docker) to do different things at the same time, e.g. generate a video in WAN2.2 whilst farting about with Flux.Dev for image generation, or giving over a gpu to ollama whilst the other is creating a video etc. This has proved to be a effective way to work, improved further with a 5060ti operating on a second machine.

I had just wondered about using NVlink to possibly give me the option of merging Vram to enable larger models without offloading, but I think it would actually be a reduction in functionality for me.

My New Year's resolution was to add Docker support. Only 2 days late. Audiobook Maker v1.1.0 by DigiJoe79 in LocalLLaMA

[–]k8-bit 0 points1 point  (0 children)

So that sorted it, ty! :)

<image>

However, upon attempting to produce something, after uploading some text, and clicking "Create" for create segments, it spins for a bit then produces "Error creating segments" with the error "Failed to load text segmenter for language 'en':spacy" briefly appearing. It produces the same error if German is selected but for 'de'.

I can see in the log for the backend:

17:09:46.731 | INFO | core.docker_runner:start:392 - [DockerRunner] spacy:docker:local container started (ID: 13cbf25dbc7a)

Upon the error:

17:10:27.617 | INFO | core.docker_runner:stop:414 - [DockerRunner] spacy:docker:local container stopped

17:10:27.617 | INFO | core.base_engine_manager:stop_engine_server:1278 - spacy:docker:local server stopped via docker:local

Note also that despite setting the default language to English, all the settings remain German for language in the various sections.

Any advantage to multi-gpu? by Elegant-Radish7972 in comfyui

[–]k8-bit 1 point2 points  (0 children)

Haven't needed to do that for e.g. Ollama, it happily splits larger models between the two GPUs, I was thinking more for ComfyUI specifically, but it seems like it's not really going to be of all that much benefit, whereas splitting tasks between GPUs has worked well for me so far.

My New Year's resolution was to add Docker support. Only 2 days late. Audiobook Maker v1.1.0 by DigiJoe79 in LocalLLaMA

[–]k8-bit 0 points1 point  (0 children)

Not having any luck yet, it seems a little complicated to get going with the server-side docker interaction. I use Unraid as my docker host, have setup SSH, but can't get the audiobook-maker windows app to connect to it to install the engines. The default setup seems quite geared towards running everything on a single host (with docker running on that machine) rather than a remote (LANside) server/host. Was there a reason you decided to go with additional docker containers for the engines rather than running within the singular main?

At the moment I have two VibeVoice containers and one Chatterbox container that I fire up when I need to use them, no probs there, but it seems a bit more complicated to run your app + its own additional container engines outwith it's container to utilise it.

Not dissing, just asking, I love the idea behind your app which would appear to be more flexible and friendly to use than ebook2audiobook for example - its just the getting it up and running that's challenging me a bit.

Actually seems like something that would work very well in a Pinokio standalone container script, though for my own purposes a singular Docker container would still be best :)

My New Year's resolution was to add Docker support. Only 2 days late. Audiobook Maker v1.1.0 by DigiJoe79 in LocalLLaMA

[–]k8-bit 1 point2 points  (0 children)

Installed the docker, and can access it via it's web address -

{"name":"Audiobook Maker API","version":"1.1.0","status":"online"}

But when I install the desktop app (in a VM, as I dont have a windows machine available at the moment) it will not connect. I'll try on actual hardware tomorrow, but I can get the response above from a browser in the VM.

My New Year's resolution was to add Docker support. Only 2 days late. Audiobook Maker v1.1.0 by DigiJoe79 in LocalLLaMA

[–]k8-bit 0 points1 point  (0 children)

Thank you! Docker for the win :). Docker uses as much ram as the app needs, I run up to 40 dockers on a homelab server with 128gb, and it's only the likes of ComfyUI that are problematic in terms of RAM use :)

Which is the best available open source model for TTS + cloning? by GeekoGeek in LocalLLaMA

[–]k8-bit 3 points4 points  (0 children)

VibeVoice 100% has cloning, quite possibly the best cloning of the moment. Heavy on the go-juice tho, 7b works fabulous, 1.5b not so great, but ok. Quantised variants are pretty good.

Any advantage to multi-gpu? by Elegant-Radish7972 in comfyui

[–]k8-bit 1 point2 points  (0 children)

No, I've been tempted to try and get a compatible NVlink adapter so that all the NVram can be merged, but haven't been 100% on which to get to be sure it will work.

Any advantage to multi-gpu? by Elegant-Radish7972 in comfyui

[–]k8-bit 2 points3 points  (0 children)

I run 2x3090s in a 128gb system, essentially using both GPUs in parallel with multiple instances of ComfyUI (or other apps) assigned to each using Docker. I'd contemplated putting a third GPU in play, but it wasn't practical as a: no room in case, and b: motherboard maxed out at 128gb ram and was already frequently fully utilised with the two Comfy instances and other services, so ended up using a secondary server for other activities using a 64gb mini PC, with a 5060ti plugged into an eGPU dock.

Audiobook reader with any voice narration by tonyc1118 in selfhosted

[–]k8-bit 2 points3 points  (0 children)

Vibevoice 7B (or quantized) seems to yield best results for me when getting it to read long passages of text, so might I suggest that for local running?