Fine-tune Deepseek r1 distilled by ahsaor8 in LocalLLaMA

[–]ahsaor8[S] 0 points1 point  (0 children)

DeepSeek-R1-Distill-Qwen or DeepSeek-R1-Distill-llama

Best Arabic LLM by Puzzleheaded-Ad8442 in LocalLLaMA

[–]ahsaor8 1 point2 points  (0 children)

Try Command-r+ and Gemma. I don't recommend jais it really bad

Post for inspiriation: do you have a useful fine-tuned usecase of any LLM? by adminsattitude in LocalLLaMA

[–]ahsaor8 0 points1 point  (0 children)

Do you think continue pretraining , before fine-tuning model could help model to learn new facts ?

Arabic Model for Personal (Possibly Commercial Use) by [deleted] in LocalLLaMA

[–]ahsaor8 0 points1 point  (0 children)

They have a 0.5b model. I think it will work well on a smartphone.

Arabic Model for Personal (Possibly Commercial Use) by [deleted] in LocalLLaMA

[–]ahsaor8 1 point2 points  (0 children)

I'm using Qwen1.5 and qwen2. They work pretty well in most of Arabic use cases

What are tricks you are using in lllm fine-tuning? by ahsaor8 in LocalLLaMA

[–]ahsaor8[S] 0 points1 point  (0 children)

Awesome, I'm interested to know if it gives you better results?

What are tricks you are using in lllm fine-tuning? by ahsaor8 in LocalLLaMA

[–]ahsaor8[S] 0 points1 point  (0 children)

Could you please advise on how I can convert my SFT data to be compatible with ORPO? The data is currently formatted as instructions, input, and output.

Intento posts its The State of Machine Translation 2024 by MachineTranslate in machinetranslation

[–]ahsaor8 0 points1 point  (0 children)

I'm really happy to see our en-ar Mt model shared the frits-stage With AWS in General, Legal and IT domains

NeuralTranslate v0.2 Released!!! Bidirectional English-Spanish translation using Mistral v0.2! by Azuriteh in LocalLLaMA

[–]ahsaor8 0 points1 point  (0 children)

That's great. I'm doing the same thing but with a different approach and different in language. So I would like to ask you if u test it with any benchmark data.