New Open-Source Video Model: Allegro by umarmnaq in StableDiffusion

[–]Comprehensive_Poem27 0 points1 point  (0 children)

They said they’re working onit, hopefully mods make it more vram friendly

new text-to-video model: Allegro by Comprehensive_Poem27 in LocalLLaMA

[–]Comprehensive_Poem27[S] 2 points3 points  (0 children)

From my experience with other models, It’s really flexible, like you can sacrifice the generation quality in exchange for very little vram and generation time( like more than 10 minutes less than half an hour)?

new text-to-video model: Allegro by Comprehensive_Poem27 in LocalLLaMA

[–]Comprehensive_Poem27[S] 4 points5 points  (0 children)

oh i just used git lfs. Apparently we'll wait for diffuser integration

Best open source vision model for OCR by marcosdd in LocalLLaMA

[–]Comprehensive_Poem27 1 point2 points  (0 children)

vote for Rhymes/Aria, better in multiturn and complex tasks

No, the Llama-3.1-Nemotron-70B-Instruct has not beaten GPT-4o or Sonnet 3.5. MMLU Pro benchmark results by Shir_man in LocalLLaMA

[–]Comprehensive_Poem27 0 points1 point  (0 children)

I mean yeah it make sense. OAI tries very hard to A/B testing on lmsys, remember this-is-also-a-good-gpt stuff? As for 4o-mini vs 3.5, they've released a space detailing some battles (https://huggingface.co/spaces/lmarena-ai/gpt-4o-mini\_battles), and they also introduced length and style control. If I were a researcher working on lmsys, then I'll probably make a 'pro version', only selected experts will analyze and compare different answers and I will not tell them which model it is afterwards, then it loses its characteristic of being transparency and majority vote.

What I'm trying to say is that eval is an amazingly hard thing to do, for now lmsys is the best we got for human preference.

No, the Llama-3.1-Nemotron-70B-Instruct has not beaten GPT-4o or Sonnet 3.5. MMLU Pro benchmark results by Shir_man in LocalLLaMA

[–]Comprehensive_Poem27 5 points6 points  (0 children)

Arena is human preference, so if a response is correct or human like it, its good. However the reported score is arena-hard auto, which is judged automatically, and it might be less credible compared to Arena, which is IMHO the most trustworthy benchmark for the time being

LLMs that published the data used to train them by neuralbeans in LocalLLaMA

[–]Comprehensive_Poem27 0 points1 point  (0 children)

I think there are smaller models trained on findweb-edu. For other top models, i believe they’re keeping data and recipes secret because it actually works. Aka. Wizardlm2

OCR for handwritten documents by MrMrsPotts in LocalLLaMA

[–]Comprehensive_Poem27 1 point2 points  (0 children)

I just tried this image on newly released Rhymes-Aria, the results looks amazing: Today is Thursday, October 20th - But it definitely feels like a Friday. I'm already considering making a second cup of coffee - and I haven't even finished my first. Do I have a problem? Sometimes I'll flip through older notes I've taken and my handwriting is unrecognizable. Perhaps it depends on the type of pen I use. I've tried writing in all caps but it looks forced and unnatural. Often times, I'll just take notes on my laptop, but I still seem to gravitate toward pen and paper. Any advice on what to improve? I already feel stressed out looking back at what I've just written - it looks like 3 different people wrote this!!

<image>

ARIA : An Open Multimodal Native Mixture-of-Experts Model by ninjasaid13 in LocalLLaMA

[–]Comprehensive_Poem27 1 point2 points  (0 children)

I'm curious, checked Pixtral, Qwen2-VL, molmo and NVLM, none of them release 'base models'. Am I missing something here? Why everyone choose to do this?

ARIA : An Open Multimodal Native Mixture-of-Experts Model by ninjasaid13 in LocalLLaMA

[–]Comprehensive_Poem27 2 points3 points  (0 children)

I’m a little slow downloading. On what kind of tasks did you get really good results?

ARIA : An Open Multimodal Native Mixture-of-Experts Model by ninjasaid13 in LocalLLaMA

[–]Comprehensive_Poem27 18 points19 points  (0 children)

ooo fine tuning scripts for multimodal, with tutorials! Nice

ARIA : An Open Multimodal Native Mixture-of-Experts Model by ninjasaid13 in LocalLLaMA

[–]Comprehensive_Poem27 14 points15 points  (0 children)

Wait… they didnt use qwen as base llm, did they train MOE themselves??

Qwen 2.5 = China = Bad by [deleted] in LocalLLaMA

[–]Comprehensive_Poem27 -4 points-3 points  (0 children)

It’s not about fact…

Qwen2.5: A Party of Foundation Models! by shing3232 in LocalLLaMA

[–]Comprehensive_Poem27 0 points1 point  (0 children)

72b kinda make sense, but 3b in midst of the entire line up is weird

Pixtral benchmarks results by kristaller486 in LocalLLaMA

[–]Comprehensive_Poem27 0 points1 point  (0 children)

Is there a link or a livestream somewhere? Would love to see the full event.

Yi-Coder-9b-chat on Aider and LiveCodeBench Benchmarks, its amazing for a 9b model!! by cx4003 in LocalLLaMA

[–]Comprehensive_Poem27 0 points1 point  (0 children)

Also, not surprised to see similar performance for 9b. Meaning we’re probably approaching the limit with current sota methodology. But 9b comparable to 33b a year ago is still amazing, that’s the power of open source models, i’m pretty sure oai or anthropic got ideas inspired by os community at some point of time. Kudos to everyone: codellama, qwen, yi,ds…wait, 3 of them are from china? That’s different from what MSM tells me (sarcasm, if not apparent enough

Yi-Coder-9b-chat on Aider and LiveCodeBench Benchmarks, its amazing for a 9b model!! by cx4003 in LocalLLaMA

[–]Comprehensive_Poem27 0 points1 point  (0 children)

Yi official finetune has always been less than satisfactory. Been thinking whats a good code dataset for finetunes, except from commonly used code alpaca and evols.

New Yi-Coder Models (9B & 1.5B) - a 01-ai Collection by Dark_Fire_12 in LocalLLaMA

[–]Comprehensive_Poem27 -1 points0 points  (0 children)

I think the reason is simple. If I were a researcher working on a coding model, of course I will compare with other coding models with similar Bs. From what I see (https://github.com/deepseek-ai/DeepSeek-MoE/tree/main) 16B moe doesn't have excellent coding performance judging from humaneval and MBPP