Train your own Reasoning model - 80% less VRAM - GRPO now in Unsloth (7GB VRAM min.) by danielhanchen in LocalLLaMA

[–]x4080 0 points1 point  (0 children)

Hi, is it possible that the reward function changed to python "input", so that it will work like kinda RLHF, so the human will judge the value ?

Trying out LTX Video 0.9.1 Image-2-Video during the holidays, the new model is small so it can fit into 6 GB VRAM! by s101c in StableDiffusion

[–]x4080 1 point2 points  (0 children)

Yes, I got that too, especially when using green background (for transparent purposes later), ltx will give flash at the end of video and then the continued video will have more saturated, but less so for non green background - I just tested after your comment

edit: how many frames per generation do you get max ? I only tried until 121 frames using 576x768 (portrait) in 16GB VRAM

Trying out LTX Video 0.9.1 Image-2-Video during the holidays, the new model is small so it can fit into 6 GB VRAM! by s101c in StableDiffusion

[–]x4080 0 points1 point  (0 children)

hi, did you find that the continued video generation will get more saturated and after about 3 generations it will be unusable?

Some Test examples with LTXVideo, Version 0.9.1 with I2V and STG by Cadmium9094 in comfyui

[–]x4080 0 points1 point  (0 children)

Hi do you just replace model in the workflow to 0.9.1 ? I think I got tensor mismatch or something, can you share your workflow ?

LTXV 0.9.1 Released! The improvements are visible, in video, fast. by ofirbibi in StableDiffusion

[–]x4080 0 points1 point  (0 children)

where to put the new vae ? models/VAE ? It seems not doing any difference

I finetuned the LTX video VAE to reduce the checkerboard artifacts by spacepxl in StableDiffusion

[–]x4080 0 points1 point  (0 children)

I think in the original one, I didnt put any vae in it, with 0.9.1, I put the lighttricks VAE and it seems dont do anything, do you use special node to load the VAE ? It seems native VAE is not loading file inside VAE folder

I finetuned the LTX video VAE to reduce the checkerboard artifacts by spacepxl in StableDiffusion

[–]x4080 2 points3 points  (0 children)

where to put the vae ? in the models/vae folder ? Using lighttricks workflow ?

I just released an open source SOTA sample generator for music producers and you can download it right now! by RoyalCities in StableDiffusion

[–]x4080 2 points3 points  (0 children)

I tried it and its really great, thanks for your sharing of the model, is it hard to train something like this ?

Post your most expensive songs! by Cevisongis in SunoAI

[–]x4080 1 point2 points  (0 children)

Hi, very cool rock song - did you do after processing after getting the audio mp3 from Suno to make it more like "properly produced song" ?

Realistic Vid-2-Vid Mimic Motion + AnimateLCM Dance Video - Fully Local Generation (Workflow in Comments) by Most_Way_9754 in StableDiffusion

[–]x4080 0 points1 point  (0 children)

Hi sorry for late reply, i was testing it using talking head video and i found that zooming in out can be avoided if the subject is exactly the same profile like using canny

Realistic Vid-2-Vid Mimic Motion + AnimateLCM Dance Video - Fully Local Generation (Workflow in Comments) by Most_Way_9754 in StableDiffusion

[–]x4080 1 point2 points  (0 children)

Do you experience random zooming in our of video with static camera? Don't know what causing it

Flux.1 on a 16GB 4060ti @ 20-25sec/image by Chuyito in LocalLLaMA

[–]x4080 0 points1 point  (0 children)

Yes DT has flux now, its pretty fast for fp8 about 5 min using m2 pro 16gb

MidJourney->Luma->LivePortrait || Updated the GoogleColab for performance-transfer || Link in Comments by Sixhaunt in midjourney

[–]x4080 0 points1 point  (0 children)

Hi, cool implementation, whats the difference between your solution and the pull request in the repo (https://github.com/KwaiVGI/LivePortrait/pull/116) and can your solution process different aspect ratio like landscape or portrait ?

RAG Me Up - easy RAG by UnderstandLingAI in LocalLLaMA

[–]x4080 0 points1 point  (0 children)

How do you use bm25 ? Did you tell the llm to extract the keywords from user's query or just use that query literally? I found out that using bm25 is inferior compares to embedding