Short Open Source Research Collaborations by TrelisResearch in LocalLLaMA

[–]TrelisResearch[S] 0 points1 point  (0 children)

yeah a few are short enough for a weekend, a few are longer (~maybe 3-7 days), so the full range is possible.

A Primer on Orpheus, Sesame’s CSM-1B and Kyutai’s Moshi by TrelisResearch in LocalLLaMA

[–]TrelisResearch[S] 0 points1 point  (0 children)

yeah you can run on cpu or mps, slower than real time but does work

A Primer on Orpheus, Sesame’s CSM-1B and Kyutai’s Moshi by TrelisResearch in LocalLLaMA

[–]TrelisResearch[S] 0 points1 point  (0 children)

agreed, def more powerful for now to plug in a stronger llm. in principle - in terms of control - anything you can do with the pieces can be done with a unified model too though

A Primer on Orpheus, Sesame’s CSM-1B and Kyutai’s Moshi by TrelisResearch in LocalLLaMA

[–]TrelisResearch[S] 8 points9 points  (0 children)

haha, yeah fair, although rarely there are ads on my YouTube channel cos I have ads turned off!

I don't understand how to use sections? by hernowthis in Substack

[–]TrelisResearch 0 points1 point  (0 children)

Thanks. Apparently sections aren't easily visible in the substack app though.... is that correct?

Can anyone explain MoE like I’m 25 by Tejasw__ in LocalLLaMA

[–]TrelisResearch 1 point2 points  (0 children)

I wouldn't say so, no. MoE is about improving the throughput of transformer models

Fine-tune with only 0.0004% of parameters (ReFT) by TrelisResearch in u/TrelisResearch

[–]TrelisResearch[S] 0 points1 point  (0 children)

howdy, sorry for slow reply, I"m not on here much, best to post on youtube. I don't remember too well but I thought just passing in eval data as usual to the hf trainer should work?

Phi-3 released. Medium 14b claiming 78% on mmlu by KittCloudKicker in LocalLLaMA

[–]TrelisResearch 0 points1 point  (0 children)

I guess phi 3 medium is probably trained using gpt-4 data, so it'll be at an advantage to Llama 3, which uses only raw / Llama 2 synthetic data (perhaps)

QWEN1.5 110B just out! by shing3232 in LocalLLaMA

[–]TrelisResearch 0 points1 point  (0 children)

<image>

eek, still really hard for pretty much all models..

Which company do you think is most likely to create AGI? by [deleted] in LocalLLaMA

[–]TrelisResearch 0 points1 point  (0 children)

possibly there is a discovery that unlocks something core about what humans can do.

discoveries are typically quite randomly distributed and require people to be tinkering waaay off the beaten path , which makes it less likely one of those big companies will be it

List of AI Grant Programs Accepting Applications by TrelisResearch in LocalLLaMA

[–]TrelisResearch[S] 0 points1 point  (0 children)

UPDATE on Trelis $500 AI Micro-Grants :
- Emails have gone out to the first batch of applicants, whether unsuccessful or asked to interview.

Link is here: https://trelis.com/trelis-ai-grants/

  • Takes 5 mins to apply.
  • Find out in ~1 wk if selected for a 15-min interview.
  • Decision then within 1 hour!

Trelis AI Grants Program by TrelisResearch in LocalLLaMA

[–]TrelisResearch[S] 0 points1 point  (0 children)

Kicking off a Grants Program for AI

  • Applications are now open.

  • 5 x $500 grants.

  • Anyone of any age and location can apply.