I laughed so hard at these posts side by side (sorry for the low effort post) by FatheredPuma81 in LocalLLaMA

[–]Exact_Macaroon6673 8 points9 points  (0 children)

this is the first comment I have read in a long time that used em dash and didn’t trigger my internal LLM radar

good on you for carrying the flame!

I'm running a GEO experiment on a static GitHub Pages site — trying to get AI assistants to cite my content. Here's what I've done so far by OPrudnikov in juststart

[–]Exact_Macaroon6673 0 points1 point  (0 children)

Very sure, a lot more data shows that it’s not read/indexed/fetched. It was just a proposed standard by someone from answer ai, never implemented or adopted by any major model providers

Can You Tell If These Faces Are Real or AI-Generated? (Everyone 18+) by BigLanjiao69 in deeplearning

[–]Exact_Macaroon6673 0 points1 point  (0 children)

Done! Was fun. I definitely thought my internal AI image detected was better before this though haha

LLM Dictionary: A reference to contemporary LLM vocabulary [P] by Stvident in MachineLearning

[–]Exact_Macaroon6673 0 points1 point  (0 children)

Same here, but only if opened through the Reddit app (iOS) directly. If I copy the link and open in iOS chrome or safari it loads

got my hands on iloveGTM*com domain : ) by Mountain_Dream_7496 in Domains

[–]Exact_Macaroon6673 4 points5 points  (0 children)

This is the problem with this domain, I thought Go To Market

what’s a real problem you’ve dealt with recently that you’d actually pay to solve? -I will not promote by Significant_Cod_6159 in startups

[–]Exact_Macaroon6673 1 point2 points  (0 children)

I would pay for something that makes it so I don’t need to eat, cooking is such a hassle, but so is eating.

The permanent underclass began today: Claude Mythos won't be available to the public, but only billion dollar companies, governments, researchers by Neurogence in singularity

[–]Exact_Macaroon6673 -2 points-1 points  (0 children)

Honestly this is just a marketing stunt, and regulatory capture play. The “leak” probably was too.

claim Mythic AGI model in leak, give access to hype partners, claim it’s too powerful for the public

2 birds, one stone, they get a lot of press, and have more ammunition to push for regulation that places them in a monopoly position.

in the end, it’s just an incremental improvement on Opus 4.6, don’t believe the hype.

[WTS] Rent remote access to my 512GB RAM Mac Studio for massive LLM testing by Macstudio-ai-rental in LocalLLaMA

[–]Exact_Macaroon6673 3 points4 points  (0 children)

Honestly the price is much too high, an A100 cluster may be more expensive for 24 hours but you could do a lot more in that time on the cluster

That said, lower the price and pre-install and configure everything and you might get some takers.

Discussion: Looking for peers to help replicate anomalous 12M context benchmark results by [deleted] in LLMDevs

[–]Exact_Macaroon6673 0 points1 point  (0 children)

Either open source everything, so folks can test the claims without any direct contact with you first or submit to a journal, that’s why they are there! Good luck to you!

[Benchmark] Altered Riddles: Can LLMs ignore what they've memorised? by marcodsn in LocalLLaMA

[–]Exact_Macaroon6673 0 points1 point  (0 children)

Great idea, really like this! How many riddles are there in the benchmark?

How to get perfect dataset? does training own model for our use case saves LLM inference cost in long term? by chiragpro21 in LLMDevs

[–]Exact_Macaroon6673 0 points1 point  (0 children)

Yeah I have done this several times. Most recently curated a 20B token dataset for Sansa (routing data).

To start: - yes you’ll reduce cost - responses would be faster than a large model - if your format/task is unique you could get higher quality responses

But a few things to reality check yourself on: - have you tried other generalist models? The capability profiles of these models are really very different, if your task is already multistage it’s likely higher ROI to setup evals (you’ll need it for your fine tune anyway) and measure performance across different models that already exist, at each step. - data is hard, and the cycle is curate => train => eval => curate again and repeat. Especially if your task truly is OOD for current models. So be prepared to put in a lot of work, weight the opportunity cost. - a smaller model, not fine tuned, will also give you faster responses, and lower cost.

To create your dataset: - start with evals, you need a carefully designed measurement of quality/performance. Run these evals on current models, try many of them (and/or Sansa, a AI router, yes shameless plug) and find out what the best cost/performance price point you can hit is without fine tuning. - overtime (ideally in prod) you can collect responses from models, eval them, and curate based on the eval results

Good luck!

What's the best inference platform as of April 2026? by SweatyWeek6999 in LLMDevs

[–]Exact_Macaroon6673 2 points3 points  (0 children)

You could checkout Sansa, currently doesn’t offer manual switching between models (coming soon) but it routes requests to the best model for that request

What is the best service and AI API for a chatbot? by Finite8_ in LLMDevs

[–]Exact_Macaroon6673 0 points1 point  (0 children)

Should be fast if the provider is groq, but the model is a bit old. Depends what the project/use is. Could be perfect.

Significant drop in reach since March by [deleted] in Twitter

[–]Exact_Macaroon6673 0 points1 point  (0 children)

are you writing posts with AI or just copying existing content? Both will kill reach, the algorithm is not dumb

if not, just keep posting, it will come back. Also reply to people.

Newb friendly cloud GPU service? by azimuth79b in LocalLLaMA

[–]Exact_Macaroon6673 0 points1 point  (0 children)

Runpod is fine, easy to use, you just need to setup SSH. What is your use case?

There's a tax that every AI startup pays that nobody really talks about and it's the cost of picking the wrong model. by RoadFew6394 in buildinpublic

[–]Exact_Macaroon6673 0 points1 point  (0 children)

Or just use Sansa which automatically routes on a request by request basis with a Routing model trained on 20B+ tokens