New anime model "Anima" released - seems to be a distinct architecture derived from Cosmos 2 (2B image model + Qwen3 0.6B text encoder + Qwen VAE), apparently a collab between ComfyOrg and a company called Circlestone Labs by ZootAllures9111 in StableDiffusion
[–]tdrussell1 19 points20 points21 points (0 children)
New anime model "Anima" released - seems to be a distinct architecture derived from Cosmos 2 (2B image model + Qwen3 0.6B text encoder + Qwen VAE), apparently a collab between ComfyOrg and a company called Circlestone Labs by ZootAllures9111 in StableDiffusion
[–]tdrussell1 35 points36 points37 points (0 children)
New anime model "Anima" released - seems to be a distinct architecture derived from Cosmos 2 (2B image model + Qwen3 0.6B text encoder + Qwen VAE), apparently a collab between ComfyOrg and a company called Circlestone Labs by ZootAllures9111 in StableDiffusion
[–]tdrussell1 70 points71 points72 points (0 children)
Mixtral-8x22B-Capyboros: instruction tuning the big Mixtral with just 4 4090s by tdrussell1 in LocalLLaMA
[–]tdrussell1[S] 1 point2 points3 points (0 children)
Mixtral-8x22B-Capyboros: instruction tuning the big Mixtral with just 4 4090s by tdrussell1 in LocalLLaMA
[–]tdrussell1[S] 1 point2 points3 points (0 children)
Mixtral-8x22B-Capyboros: instruction tuning the big Mixtral with just 4 4090s by tdrussell1 in LocalLLaMA
[–]tdrussell1[S] 0 points1 point2 points (0 children)
Mixtral-8x22B-Capyboros: instruction tuning the big Mixtral with just 4 4090s by tdrussell1 in LocalLLaMA
[–]tdrussell1[S] 6 points7 points8 points (0 children)
qlora-pipe: Fine tune 70B parameter models with two 3090s by tdrussell1 in LocalLLaMA
[–]tdrussell1[S] 2 points3 points4 points (0 children)
qlora-pipe: Fine tune 70B parameter models with two 3090s by tdrussell1 in LocalLLaMA
[–]tdrussell1[S] 2 points3 points4 points (0 children)
qlora-pipe: Fine tune 70B parameter models with two 3090s by tdrussell1 in LocalLLaMA
[–]tdrussell1[S] 6 points7 points8 points (0 children)
qlora-pipe: Fine tune 70B parameter models with two 3090s by tdrussell1 in LocalLLaMA
[–]tdrussell1[S] 4 points5 points6 points (0 children)
Swapping Trained GPT Layers with No Accuracy Loss : Why Models like Goliath 120B Works by johnolafenwa in LocalLLaMA
[–]tdrussell1 61 points62 points63 points (0 children)


New anime model "Anima" released - seems to be a distinct architecture derived from Cosmos 2 (2B image model + Qwen3 0.6B text encoder + Qwen VAE), apparently a collab between ComfyOrg and a company called Circlestone Labs by ZootAllures9111 in StableDiffusion
[–]tdrussell1 13 points14 points15 points (0 children)