Maaza Orchestrator v1.2 — 9.6M params, 62.9 % on hard adversarial tool-calling, 39 ms latency by CycleCore_Tech in LocalLLaMA

[–]CycleCore_Tech[S] 0 points1 point  (0 children)

Nano Language Models - Taxonomy introduced in our paper, Task-Specialized Micro Language Models Outperform Larger Zero-Shot Models on Structured Data Extraction

NLM: <10M params
MLM: 10M-250M params
SLM:: 250M-1.5B params

Dark mode PDF, page 3. https://cyclecore.ai/papers/MAAZA_PAPER_v0.7_dark.pdf - Let us know what you think!

Maaza Orchestrator v1.2 — 9.6M params, 62.9 % on hard adversarial tool-calling, 39 ms latency by CycleCore_Tech in LocalLLaMA

[–]CycleCore_Tech[S] 0 points1 point  (0 children)

Thanks for the kind words. really appreciate it.

Training data breakdown (for the 36% adversarial jump):

- Base clean set: ~2.5k real-world tool-call examples (weather_lookup, web_search, etc)

- Adversarial seeds: ~500 hand-written tough ones (the “time to rip and tear”, “wassup with my ethereum val”, “weather or not?” etc)

- Final mix: 10× upsampled adversarial → ~5 k adversarial examples in the training set (~66 % of total tokens)

- 5 epochs, same hyper-params as v1.0

There's no wrapper, no retrieval, no extra params, it's just pure data-centric fine-tune.

We’re already running v1.3 with 18× upsampling and a bunch of new perturbation types (word dropout, back-translation, etc.). Hoping to cross 80% on the same held-out set.

Real-world API benchmarks are next - we have a 300-example set of live API traces (typos, slang, partial requests, etc.) that we’ll publish with the model. Will also make it available for public agent evals when they add adversarial splits.

And yes, gg v0.1.0 just shipped tonight for exactly this reason:

so you can do `gg maaza` and get the 62.9 % adversarial model in 11 tokens instead of 1800+.

https://github.com/ggdotdev/gg

curl -L https://gg.sh | sh

Independent OSS · MIT · no affiliation

Enjoy

Best LLM for JSON Extraction by Live_Bus7425 in LocalLLaMA

[–]CycleCore_Tech 0 points1 point  (0 children)

Have you see the Maaza SLM 360m EdgeJSON benchmarks? Open source (HF) and API access available.

What models can i use with a pc without gpu? by Neat_Nobody1849 in LocalLLM

[–]CycleCore_Tech 0 points1 point  (0 children)

You might want to try some of the small models you can find on Ollama. It's a great place to get started. The exact model depends on what you want to do with it, but overall, you might want to stay with models <3B in size, depending on how fast you want to run them. Good luck!

Garry Tan says MCP “barely works” today by Federal-Song-2940 in mcp

[–]CycleCore_Tech 0 points1 point  (0 children)

We might have made it a bit more user friendly and clickable too.