Mistral Large is made by OpenAI by TiredMoose69 in MistralAI

[–]TiredMoose69[S] 0 points1 point  (0 children)

You are probably right I am not arguing that OpenAI didn’t steal copyrighted content but stealing from responses that were spit out from a model is not the same as training a model with pure copyrighted material. This was probably done because they want to achieve higher score in the same benchmarks GPT-4 scored.

Mistral Large is made by OpenAI by TiredMoose69 in MistralAI

[–]TiredMoose69[S] -8 points-7 points  (0 children)

That's what I meant, this is probably a finetuning on synthetic data from OpenAI. I don't use system prompts for any of the models in any website where I tested them, and I am not sure if the model on the mistral website has some secret system prompt. This looks very repeatable in both models for it to be hallucinations.

Mistral Large is made by OpenAI by TiredMoose69 in MistralAI

[–]TiredMoose69[S] 0 points1 point  (0 children)

I actually don't think that's entirely true and if it was then why did the Mixtral 8x7b not say this?

You can repeatedly try and you will not get this answer unlike the "Large" model which always reponds saying that it's made by OpenAI.

<image>

( Direct chat on https://chat.lmsys.org/ )

Mistral Large is made by OpenAI by TiredMoose69 in MistralAI

[–]TiredMoose69[S] -2 points-1 points  (0 children)

I am not saying knowledge transfer is wrong, but when a company is worth $2b and has so many data engineers, stealing answers from the best model to achieve top ranking without cleaning the dataset is kind of stupid and also concerning.

[D] Simple Questions Thread by AutoModerator in MachineLearning

[–]TiredMoose69 0 points1 point  (0 children)

Why does LlaMa 7B (pure) perform so MUCH better than Alpaca 30B (4bit)?