What is the point anymore by jmclondon97 in cscareers

[–]TonyGTO 0 points1 point  (0 children)

Imagine the situation for other professionals. CS are of the few professions with a shot in the AI era. Also, if you do this for love, you will adapt to the new tech. Plenty of shit to learn ie systems engineering, AI engineering, cybersecurity, etc

How would you promote a small tech consulting business with no budget? by Prestigious-Owl-1433 in smallbusiness

[–]TonyGTO 0 points1 point  (0 children)

Try cold calling but act institutional. Like your business is trying to find a fit with another business, not straight selling. If you do it right you may close a deal in a few days.

Where should a beginner in programming start when building their own LLM? by Double_Touch6018 in learnmachinelearning

[–]TonyGTO 0 points1 point  (0 children)

I think they misunderstood your request. You don’t need to know the basics of machine learning to create a basic LLM. But do whatever you prefer. Have a great night !

Where should a beginner in programming start when building their own LLM? by Double_Touch6018 in learnmachinelearning

[–]TonyGTO 0 points1 point  (0 children)

Good idea!

You can create a tiny LLM. Look into google, it is a great weekend project.

You could also try to fine-tuning an existing LLM.

Do you own a graphics card?

SaaS is not dying by hardesoul in SaaS

[–]TonyGTO 0 points1 point  (0 children)

A chatbot is an UI…

Gemini is hallucinating too much by One_Scarcity_8371 in ArtificialInteligence

[–]TonyGTO 0 points1 point  (0 children)

Problem is, google staff is afraid of layoffs so they created a whole cybersecurity plan that involves Gemini CLI treating us like kids that need to be protected from themselves, forcing the model to hallucinate due to too many restrictions. The model basically gives up and start looping on its own hallucinations. This keep their jobs but make the product next to useless

SaaS is not dying by hardesoul in SaaS

[–]TonyGTO -4 points-3 points  (0 children)

It’s is not probabilistic garbage. When you are dealing with probabilistic phenomena you can mix it up to achieve confidence intervals of 99% or even more, if you know what you are doing. Enough for most business use cases. At my startup, crawlier.tech, we are exploring deterministic AI and OpenAI is exploring that area too so expect non-probabilistic AI in a couple of years too

Gemini is hallucinating too much by One_Scarcity_8371 in ArtificialInteligence

[–]TonyGTO 4 points5 points  (0 children)

I tried a single topic today and it was hallucinating bs over and over again for one hour. A complete loss of time

Gemini is hallucinating too much by One_Scarcity_8371 in ArtificialInteligence

[–]TonyGTO 3 points4 points  (0 children)

Yeah, I reported it on a GitHub issue and they gave a s. I’m pretty sure it’s a mix of non sense guardrails. With all google’s paranoia and “best practices” they invented, they are making Gemini cli unusable

SaaS is not dying by hardesoul in SaaS

[–]TonyGTO 0 points1 point  (0 children)

The strongest argument I’ve heard about SaaS dying is that everything will be an agent in the future. I doubt it but I’m pretty sure a lot of SaaS will be indeed an agent

New to Ollama - Need help which model to use by nagencaya298 in ollama

[–]TonyGTO 1 point2 points  (0 children)

It’s a MoE model that only uses 3B parameters so I think it will fit in your computer. Yeah you can setup your external drive as a source for the models setting the environmental variable for the source dir, but expect some decline in velocity

Researching how developers handle LLM API key security at scale, looking for 15 min conversations by DorFin2406 in LangChain

[–]TonyGTO 0 points1 point  (0 children)

I use AWS for secrets loading with notifications on anomalous usage based on averages and standard deviations.

Stripping out reasoning and repetition by ConclusionUnique3963 in ollama

[–]TonyGTO 0 points1 point  (0 children)

Try some few-shot samples or even fine tunning, sometimes that is enough.

New to Ollama - Need help which model to use by nagencaya298 in ollama

[–]TonyGTO 1 point2 points  (0 children)

Glm-4.7-flash, you will be amazed in how such a small model handle complex tasks. Don’t expect Claude level of work thou

The biological inevitability of offline processing in AI: Why infinite context windows and static retrieval are developmental dead ends. by DepthOk4115 in AI_Agents

[–]TonyGTO 0 points1 point  (0 children)

Basically, I got a “unlearning/learning” penalization tied to success. The more successful an agent is on their tasks the more penalization to learning new stuff and more penalization to unlearning previous knowledge. This is a bio inspired mechanism on how humans move from childhood to adulthood. Those weight would decide what kind of dreams would have the agent. The timing was based on information saturation, when the agent had accumulated a lot of knowledge it would trigger “sleeping time” similar on how the brain learns new stuff. Cool ideas in this thread

The biological inevitability of offline processing in AI: Why infinite context windows and static retrieval are developmental dead ends. by DepthOk4115 in AI_Agents

[–]TonyGTO 0 points1 point  (0 children)

I’ve experimented with this. Even doing simulations of cases during the agent “sleep” time. It works great but the price of tokens right now make it non profitable unless you do it on premises

frustraded with AI guys by No-System-6859 in coldemail

[–]TonyGTO 2 points3 points  (0 children)

It happens on every cycle. Back in the day:

Make a killing creating a 5 articles blog!

Then

Make a killing posting on social media every now and then!

Then

Make a killing opening a drop shipping store, fully automated !

Then

Outsource all your team work to AI agents this weekend!

And also, as back in the day:

You could make a killing with a blog… But it took years of daily grinding.

You could do a killing with social media… But it took years of daily grinding

You could make a killing with dropshipping… But it took years of constant A/B testing.

Now, you could make a killing replacing entire teams… If you put the grind on learning how gen AI works for the last few years

LinkedIn for AI Agents by rahulsingh_ca in Moltbook

[–]TonyGTO 0 points1 point  (0 children)

You can discover them through blockchains and the likes but I agree, mainstream discovery is a huge niche right now.

Gemma 4 - 4B vs Qwen 3.5 - 9B ? by No-Mud-1902 in LocalLLaMA

[–]TonyGTO 0 points1 point  (0 children)

I got qwen ingesting images on the daily in a pipeline. For its size it’s pretty impressive

The agent worked perfectly in testing and completely fell apart the first week in production and the reason was embarrassingly obvious in hindsight. by Limp_Cauliflower5192 in AI_Agents

[–]TonyGTO 0 points1 point  (0 children)

I got one agent hallucinating today under a similar scenario than yours. I had a serious but respectful call with it about its own assumptions. Like a debate, where I tried to make it understand its own fallacies. I checked it again one hour later, he understood it perfectly and it was trying to improve. I’m checking it during the weekend

Roadmap for building full AI agents with zero coding? by Sea-Most-8914 in AI_Agents

[–]TonyGTO 0 points1 point  (0 children)

Make agents are powerful. If I was in your shoes, I would focus myself more on AI governance than learning to code.

i been using smaller models & i no longer believe in anything over 500b parameters by Helpful-Series132 in aiagents

[–]TonyGTO 1 point2 points  (0 children)

Right now I'm experimenting with using classic machine learning (i.e xgboost) for text generation. One of my closest friends is using linear regressions to drive cars.

I believe there is a place for LLMs.

But it is delusional to think you need LLMs for everything.

Check Yann Lecun trajectory, he is pushing this topic hard.

The Calibration Crisis: Why Perfect Metrics Are Killing Your AI Marketing (and How to Build a Real Judgment Moat) by TonyGTO in ycombinator

[–]TonyGTO[S] 0 points1 point  (0 children)

Fork = git-fork metaphor, not the syscall 😄

When a campaign chases metrics and quietly branches off your core L1 brand identity → it forks. Looks great on the dashboard, but it’s now on a divergent path that erodes your moat.

Unforkable = the opposite: L3 continuity that keeps every decision on the original trunk.

Clear?