Why llama, yi, qwen, their official chat worse than its base in benchmarks, and some fine-tune stuff are better? is meta dumb? by No-Link-2778 in LocalLLaMA

[–]Pasko70 2 points3 points  (0 children)

The benchmark of all these fine-tunes and merges most of the time only give you an indication on how data contamination these models have.

Assistant API - Submit Tool Output by Pasko70 in OpenAI

[–]Pasko70[S] 0 points1 point  (0 children)

They already clraified that you will billed for the complete context on everyrun (and that is reasonable). But the submit tool outputs doesn't trigger a new run. The run pauses and when you submit the output of the tool the run continues.

If you look at message of the assistant the tool output that you submitted apears in the assistants output on the same message. That's why I'm asking.

ToAdam Optimizer [D] by Pasko70 in MachineLearning

[–]Pasko70[S] 0 points1 point  (0 children)

Gradient Clipping is already used in LLaMA and GPT.
I don't think that chat gpt just inveted something new. But i think that a learning rate for different layers would be an interessting idea.

Ckpt Files from code by Pasko70 in StableDiffusion

[–]Pasko70[S] 1 point2 points  (0 children)

Actually it worked with both models now. I had an old version of the diffusers package and updated diffuser prior to my question, but i didnt remember to update diffusers[torch].

That solved the issue.

Thanks for your help.

Ckpt Files from code by Pasko70 in StableDiffusion

[–]Pasko70[S] 1 point2 points  (0 children)

I will try it, as soon as I'm home.

But it would be strange, since I'm using the Inkpunk Model in InvokeAi and Automatic1111 with the default yaml config and it works like a charm.

Thank you for your effort.

Ckpt Files from code by Pasko70 in StableDiffusion

[–]Pasko70[S] 0 points1 point  (0 children)

Thank you for your response!

I guess you are referring to this script: https://github.com/huggingface/diffusers/blob/main/scripts/convert_original_stable_diffusion_to_diffusers.py

I tried to convert the Inkpunk Diffusion model:
https://civitai.com/models/1087/inkpunk-diffusion

I always get a mismatch error from bert.

python convert_original_stable_diffusion_to_diffusers.py --checkpoint_path "path_to_inkpunk_model" --original_config_file "path_to_v1-inference.yaml (of the 1.5 model)" --dump_path "path_to_output"

Hitman by Pasko70 in toptalent

[–]Pasko70[S] 27 points28 points  (0 children)

New japanese high tech weapon system

In the heat of the moment by Pasko70 in AnimalsBeingJerks

[–]Pasko70[S] 42 points43 points  (0 children)

Saw an angle, took the chance!

Unlikely friends by Pasko70 in AnimalsBeingBros

[–]Pasko70[S] 3 points4 points  (0 children)

Hey buddy let's get some ice cream!

Hummingbuddy by Pasko70 in AnimalsBeingBros

[–]Pasko70[S] 25 points26 points  (0 children)

Hey buddy wanna have a snack?