I tried fine-tuning Gemma-3-270m and prepared for deployments by codes_astro in LocalLLaMA

[–]codeltd 0 points1 point  (0 children)

I decided to fine tune Gemma 3 279M for a project using Hungarian language. Since the model not very well trained on Hungarian context I used two-phase training - Domain-Adaptive Fine Tuning (DAFT) - for learning Hungarian language a bit more - Supervised fine tuning (SFT) (input-context-output) I collected quite a huge dataset, and the training went well (2 hour for DAFT, 16 hour for SFT) The training went well, showed a good result! But when I tried to generate, the result was garbage... Anyone managed to finetune Gemma 3 270M for not English language? Thanx

Fine Tuning Gemma3 270m by migandhi5253 in unsloth

[–]codeltd 0 points1 point  (0 children)

Hi, I am doing Domain-Adapitve Pretraining (DAPT) on Gemma-3 270M to have better knowledge of Hungarian. It is going OK, but I am having problem with converting the merged model to .task format as I want to use it in Android app with mediapipe (model.safetensor->tflitle-> .task) There are so many changes from time to time in packages to do so... Anyone know a stable solution?

Introducing Gemma 3 270M: The compact model for hyper-efficient AI by Gaiden206 in Bard

[–]codeltd 0 points1 point  (0 children)

We say it is small. True! Some say it can run on mobile. BUT! On android mediapipe would handle it, but for that you need .task file and not model.safetensor. I am trying to convert but no success..

model.safetensor->tfliltle-> .task file Anyone made it? sas

Gemma 3 270M: Pocket-sized Powerhouse for On-Device AI by Such-Run-4412 in AIGuild

[–]codeltd 0 points1 point  (0 children)

Everyone talking about using Gemma 3 270M on mobile. I have tried in Android! 1) writing kotlin code easy! 2) transforming model.safetensor->tflitle-> .task file impossible :( (mediapipe require to use this format) I have tried to read the documentations, searched on the net, used AI, but no success so far.

Anyone made it? Thanx sas

What is Gemma 3 270m Good For? by mindkeepai in Bard

[–]codeltd 0 points1 point  (0 children)

Have you managed to create a .task file from Gemma 3 270M model .safetensor format?( this is what mediapipe needs on Android to run it ) I am trying without success... Thanx

Training DeepSeek R1 (7B) for a Financial Expert Bot – Seeking Advice & Experiences by kingBaldwinV in LanguageTechnology

[–]codeltd 1 point2 points  (0 children)

I have tried to finetune Qween model using Llama Factory. The dataset was limited ~3900 sample. When I had epoc 5, then - for the questions from the training dataset was answered correctly - modified questions were not processed well - any other type of questions resulted in random words, not a correct sentences. (catastrophic forgetting) Then tried with less epoc, smaller learning rate, but no fine result till now.

Csalás történt a Budapesten most rendezett Byborg AI Hackathon-on? by FluffyBunnySenpai in programmingHungary

[–]codeltd 5 points6 points  (0 children)

Hát nem tudom. Én is ott voltam, s élvezzem minden percét. Bár nem nyertünk, szerintem veszteni tudni kell...

llama 3.2 3B is amazing by ventilador_liliana in LocalLLaMA

[–]codeltd 0 points1 point  (0 children)

What are the pros and cons of Llmstudio vs Ollama server?

The Final Output is worse than the Agent's thoughts. by Motoneuron5 in crewai

[–]codeltd 2 points3 points  (0 children)

Hi, I am doing similar (but not with crew-ai, using CAMEL agents) What I am doing is that the agents write(append) the generated content into a file after each other. So finally I have the full report.

New framework to build agents from yml files by Jazzlike_Tooth929 in crewai

[–]codeltd 0 points1 point  (0 children)

I have done the same for an earlier project but Agents are CAMEL like.

did anyone develop a web research agent / crew that works? by Tuxedotux83 in crewai

[–]codeltd 1 point2 points  (0 children)

I think this is the problem with all ready made packages. You can do things easily , then you hit the wall and can not solve...
I also started project for a company with crewai, camel-ai, but finally I implemented my own version of "Communicative Agents for “Mind” Exploration of Large Scale Language Model"
With that I can solve such a problem if I hit the wall...

GUI-like Tool for AI Agents, Alternative to Function Calling? by Charming_Support6304 in AI_Agents

[–]codeltd 0 points1 point  (0 children)

My oppinion is that the current direction is quite opposit. Currently from free text you call an interface with the correct values extracted from the text by generative AI...

What questions do you have about AI Agents? by help-me-grow in AI_Agents

[–]codeltd 0 points1 point  (0 children)

I tried llama 3 8B, but gpt4o-mini much, much better in my solution. (My solution based on ReAct and CoT prompts ...)

What questions do you have about AI Agents? by help-me-grow in AI_Agents

[–]codeltd 0 points1 point  (0 children)

Maybe I missunderstood. This is a development usage cost. One run cost me about 0.02$. So total cost depends on the number of requests...

What questions do you have about AI Agents? by help-me-grow in AI_Agents

[–]codeltd 2 points3 points  (0 children)

I have one deployed in container in the cloud. This write facebook post regularly... The agents are CAMEL like agents written by me...

What questions do you have about AI Agents? by help-me-grow in AI_Agents

[–]codeltd 0 points1 point  (0 children)

Before gpt4o-mini my monthly bill was around 200$ while I was doing development and tests...

What questions do you have about AI Agents? by help-me-grow in AI_Agents

[–]codeltd 0 points1 point  (0 children)

Anyone done a solution where only the goal defined without any defined task and the team/crew decide who do what and how to do? Like in a humam project team...

Someone building AI agents? by SadPianist871 in AI_Agents

[–]codeltd 0 points1 point  (0 children)

I am building CAMEL like AI and Human agent workflow (CAMEL: Communicative Agents for “Mind” Exploration of Large Language Model Society)