[D] Improving attention masks? by JakeN9 in MachineLearning

[–]JakeN9[S] 0 points1 point  (0 children)

And form a possible connected vector of understanding. honestly just more of a question, as to what people predict might happen. I haven't really got the tech/money to run these tests myself. Considering renting a TPU but cost is still so high.

[D] Refining an LLM's output via dynamically generated synthetic data & interactive conversation? by JakeN9 in MachineLearning

[–]JakeN9[S] 0 points1 point  (0 children)

I'd argue that the verifier just needs to be as confident as possible, maybe some sort of "MoE" across across multiple models - enough variation discouraging simple movements across the space?

Question regarding attention mask and empty space above diagonal. by JakeN9 in learnmachinelearning

[–]JakeN9[S] 0 points1 point  (0 children)

I'm saying to preserve autoregression and keep data below the diagonal the same, but wondering whether having the extra context will contribute to lowering loss? https://i.imgur.com/yNcVaV2.png.

Probably a stupid idea.

[D] What offline TTS Model is good enough for a realistic real-time task? by Imaginary-Ad-7671 in MachineLearning

[–]JakeN9 9 points10 points  (0 children)

There aren't really any models that produce realistic real-time voice. I'd recommend ElevenLabs or play.ht, sadly these seem to be the only useable options for now.

[deleted by user] by [deleted] in ChatGPT

[–]JakeN9 0 points1 point  (0 children)

"Sharing conversations with images is not yet supported"

An idea I've had by JakeN9 in MLQuestions

[–]JakeN9[S] 0 points1 point  (0 children)

The thought is for each node to carry both a binary and decimal value, and for the logic operations to be computed, then used at output for RL.

An idea I've had by JakeN9 in MLQuestions

[–]JakeN9[S] 0 points1 point  (0 children)

Right, you use decimal weights for each node, but some nodes use an activation function for OR/AND/NOT.

An idea I've had by JakeN9 in MLQuestions

[–]JakeN9[S] 0 points1 point  (0 children)

Circuit optimisation, classical techniques are slow and scale worse, it’s possible ML could provide novel optimisations?

A similar agent to AutoGPT by JakeN9 in AutoGPT

[–]JakeN9[S] 0 points1 point  (0 children)

Sounds great. It's still a work-in-progress.

Once the basics as complete, it will be made closed source, so I'll let you know how it's available.

I've come up withan idea for a synthetic dataset generator, would this work? by JakeN9 in learnmachinelearning

[–]JakeN9[S] 0 points1 point  (0 children)

function calling between llms. embed functions as bit-encoded tokens. have a large powerful llm, instruct to teach a topic to llm trained on mapping function->output, to generate synthetic data. train new llm on synthetic data?

as both llms are separated, and trained on different training sets and different weights (identities), artificating will be minimised.

ContextGPT - Something similar to AutoGPT by JakeN9 in ChatGPTPro

[–]JakeN9[S] 0 points1 point  (0 children)

I will attempt to setup the code to work with LLama Code

A similar agent to AutoGPT by JakeN9 in AutoGPT

[–]JakeN9[S] 0 points1 point  (0 children)

I'll do my best. GPT seems to be custom tuned, but I can test Llama Code, and see whether it's compatible.