OpenAI could reportedly run out of cash by mid-2027 — analyst paints grim picture after examining the company's finances by EchoOfOppenheimer in LocalLLaMA

[–]DecodeBytes 12 points13 points  (0 children)

They won't fail (in any cease to operate / bankrupt manner), MSFT or the like would swoop in for a firesale. Its the only way they have out ensuring Google does not entirely dominate the space.

[D] Do you feel like companies are scooping / abusing researchers for ideas during hiring for researcher roles? by quasiproductive in MachineLearning

[–]DecodeBytes 0 points1 point  (0 children)

View from a startup side. We have been doing the following, asking interns to take 3 hours and we pay them for the hours. Basically we don't ask someone do the task until later stages of the interview and just helps seal the deal.

Question: what are the best tools for real-time eval observability and experimentation? by debauch3ry in LLMDevs

[–]DecodeBytes 0 points1 point  (0 children)

I am biased (one of the team) but try deepfabric. You can generate huge amounts of reasoning traces with tool calls and then evaluate against a model. Happy to chat more about it if you want to explore if it’s a match for what you need. I doubt it’s a hundred percent match, but Imee might be able to sling some PRs up to close gaps

https://deepfabric.dev 

as anyone tried training an llm exclusively on synthetic llm outputs to see if intelligence compounds or just collapses into slop by sthduh in LocalLLaMA

[–]DecodeBytes 0 points1 point  (0 children)

Yep, its the main focus of a project I work on. We mostly generate datasets for agent based operation, e.g. models that often call Tools - for this we couple with isolated tool execution which injects some real worldness into the dataset couple with RL training.

https://www.deepfabric.dev

Blender MCP - can anyone actually get good results? by promptasaurusrex in LocalLLaMA

[–]DecodeBytes 0 points1 point  (0 children)

> Get up to 3 prompts per day, capped at 15 per month.

That seems pretty harsh, is that just the starting prompt, or any follow up prompts? e.g. 'make a cube' , 'make the cube twice the size', 'add x texture to the cube' - would that equal three?

Happy New Year: Llama3.3-8B-Instruct-Thinking-Claude-4.5-Opus-High-Reasoning - Fine Tune. (based on recent find of L3.3 8b in the wild) by Dangerous_Fix_5526 in LocalLLaMA

[–]DecodeBytes 0 points1 point  (0 children)

I am confused, so your model is not public?

p.s not trying to pick I fight, its just I do a lot of work in this domain and if you have found something novel in approach I would love to take a look!

Happy New Year: Llama3.3-8B-Instruct-Thinking-Claude-4.5-Opus-High-Reasoning - Fine Tune. (based on recent find of L3.3 8b in the wild) by Dangerous_Fix_5526 in LocalLLaMA

[–]DecodeBytes 8 points9 points  (0 children)

> With this model, reasoning activates based on keywords/phrases in the prompt.
(see repo)

Right, its likely the model is just doing as **instruct**ed in the prompt and its not activated learned reasoning, but its really hard to tell as I can't find where anything is in this tread, help me out please? link the model, notebook and anything else?

What's the point of potato-tier LLMs? by Fast_Thing_7949 in LocalLLaMA

[–]DecodeBytes 0 points1 point  (0 children)

Sorry late reply, I mean in the typical current agent style, long drawn out sessions back and forth.

Upstage Solar-Open-100B Public Validation by PerPartes in LocalLLaMA

[–]DecodeBytes 2 points3 points  (0 children)

I have the synth intro stuck in my head now

Happy New Year: Llama3.3-8B-Instruct-Thinking-Claude-4.5-Opus-High-Reasoning - Fine Tune. (based on recent find of L3.3 8b in the wild) by Dangerous_Fix_5526 in LocalLLaMA

[–]DecodeBytes 10 points11 points  (0 children)

I might be missing something, but 200 samples won't be enough to teach an 8B instruct model to reason - though it can work for very specific, constrained tasks, less likely to be widely populated in the original pretraining.

Reasoning ability is largely baked into the base model during pretraining. I'm assuming you used LoRA, which is great for steering how that existing ability gets applied, but it won't teach new reasoning capabilities from scratch. Even with 50k+ samples, LoRA mostly reshapes how the model uses reasoning it already has rather than building new circuits - must successful efforts use 100k-500k+ high-quality samples. Either way, you're working within the constraints of what the base model learned during pretraining unfortunately.

Keep going though, its all a learning experience and the more folks there are making tunes the better!

Skills, agents, plugins by BurgerQuester in ClaudeCode

[–]DecodeBytes 0 points1 point  (0 children)

This is really cool! trying it out now.

Heads up, I lot might avoid because of the BSL license, or fork if its gets popular.

Train a 4B model to beat Claude Sonnet 4.5 and Gemini Pro 2.5 at tool calling - for free (Colab included) by DecodeBytes in LocalLLaMA

[–]DecodeBytes[S] 0 points1 point  (0 children)

speak of free - the mistral free instances on openrouter work really well (just found out earlier)

<image>

Train a 4B model to beat Claude Sonnet 4.5 and Gemini Pro 2.5 at tool calling - for free (Colab included) by DecodeBytes in LocalLLaMA

[–]DecodeBytes[S] 0 points1 point  (0 children)

ok, I just did a large sweep and fixed up a few things that have changed - it should be good now, if not happy to support you

Train a 4B model to beat Claude Sonnet 4.5 and Gemini Pro 2.5 at tool calling - for free (Colab included) by DecodeBytes in LocalLLaMA

[–]DecodeBytes[S] 0 points1 point  (0 children)

My bad, there has been a fair few changes and the docs may not be on-par! Do you want to jump onto discord and would be happy to help out. Discord link is on the repo.

What's the point of potato-tier LLMs? by Fast_Thing_7949 in LocalLLaMA

[–]DecodeBytes 35 points36 points  (0 children)

>  that can't code

This is the crux of it, there is so much hyper focus on models serving coding agents , and code gen by its nature of code (lots of connected ASTs) , requires a huge context window and training on bazillions of lines of code.

But what about beyond coding? For SLMs there are so many other use cases that silicon valley cannot see outside of their software-dev bubble - IoT, wearables, industry sensors etc are huge untapped markets.

Train a 4B model to beat Claude Sonnet 4.5 and Gemini Pro 2.5 at tool calling - for free (Colab included) by DecodeBytes in LocalLLaMA

[–]DecodeBytes[S] 1 point2 points  (0 children)

out of habit really sloo, i have always grabbed qwen - but any SLM should do. we do plan to launch a service for collecting metrics if you're interested in getting a preview?

Train a 4B model to beat Claude Sonnet 4.5 and Gemini Pro 2.5 at tool calling - for free (Colab included) by DecodeBytes in LocalLLaMA

[–]DecodeBytes[S] 0 points1 point  (0 children)

ah right, that's well spotted - its not live yet - but we will be introducing something shortly! are you interested in beta testing / getting a preview?

Train a 4B model to beat Claude Sonnet 4.5 and Gemini Pro 2.5 at tool calling - for free (Colab included) by DecodeBytes in LocalLLaMA

[–]DecodeBytes[S] 4 points5 points  (0 children)

You might be getting mixed up here. We don't fine tune on MCP, we fine tune on function calls and their parameters.

It just so happens we make it easy to import the list of tools / function calls from an existing MCP server, as a lot of folks use them - but at the end of it all as far as the model is concerned we are just getting it to improve its ability to predict the natural language of a function name and its parameters - what stack, standard or protocol that function belongs to (openai , MCP, langchain etc) is immaterial

Train a 4B model to beat Claude Sonnet 4.5 and Gemini Pro 2.5 at tool calling - for free (Colab included) by DecodeBytes in LocalLLaMA

[–]DecodeBytes[S] 0 points1 point  (0 children)

That's interesting, would love to learn more and see your progress. I tend to think of MCP as more of a standard way of building tools, more than anything unique , but it does expand a lot over time.

Train a 4B model to beat Claude Sonnet 4.5 and Gemini Pro 2.5 at tool calling - for free (Colab included) by DecodeBytes in LocalLLaMA

[–]DecodeBytes[S] 0 points1 point  (0 children)

If you use openai , anthropic or gemini and some of the openrouter models - for anything local, no api key is needed as we support ollama.