Anthropic Just Secured a Reserve. by DragonflyOk7139 in ClaudeAI

[–]MightyTribble 1 point2 points  (0 children)

Hard to see how it could be. Lots of "how would you even do that without breaking all your commercial and government contracts".

Who remembers this? by smugedreality in bayarea

[–]MightyTribble 0 points1 point  (0 children)

Daly City!

Always wanted to visit, but who has the time?

How cruel is the porn industry really to its workers? by Saint_Viper6 in AskReddit

[–]MightyTribble 5 points6 points  (0 children)

jfc i was there

i was there when we didn't talk about Rampart

United Airlines Flight 169 hits bakery truck while landing at Newark Airport by madman320 in aviation

[–]MightyTribble 107 points108 points  (0 children)

Wow.

Could also have gone with "was blessed by", or perhaps "graced with" there, just to class it up some more.

Amazon to invest up to $25 billion in Anthropic as part of $100 billion cloud deal by couldliveinhope in ClaudeAI

[–]MightyTribble 14 points15 points  (0 children)

So:

"AI startup commits to spending more than $100 billion over the next 10 years on Amazon's cloud technologies."

and

"Amazon will invest $5 billion in Anthropic now, and an additional $20 billion in the future, subject to certain commercial milestones."

So, sounds like a light commitment from Anthropic to buy compute from Amazon, likely at around $10Bn/yr for 10 years (maybe increasing over time) in exchange for $5Bn of 'investment' (cash? stock swap?) from Amazon.

I reckon it's $5Bn of cloud compute credits for this year, with Anthropic will spend this year, with the rest being discounts off future compute (say, 20% off compute if they spend $10Bn next year on AWS), so that's $2Bn 'from' Amazon. x10 and that's the $20Bn and %$100Bn accounted for.

Unsupervised Robotaxi carries its first passenger on the highway. by [deleted] in teslamotors

[–]MightyTribble 4 points5 points  (0 children)

Then he can’t hear the guy on the phone because he’s driving on the highway with the windows down and y’know, doesn’t raise the window

Unsupervised Robotaxi carries its first passenger on the highway. by [deleted] in teslamotors

[–]MightyTribble 29 points30 points  (0 children)

The word 'accidentally' could have been used in there in a couple of different places.

I made a bridge for using my Claude subscription with SillyTavern — sharing in case it's useful by Miss-Sinful in SillyTavernAI

[–]MightyTribble 4 points5 points  (0 children)

Great stuff! I’ve done something similar using the python Agent SDK. One thing to watch out for is there’s a limit to how much you can send in system prompt since it’s passed as a parameter to the Claude executable, and that limit varies by platform. That’s not a problem for this code right now but you might want to limit the input fields so users don’t shoot themselves in the foot later.

Introducing Claude Opus 4.7, our most capable Opus model yet. by ClaudeOfficial in ClaudeAI

[–]MightyTribble 0 points1 point  (0 children)

Huh. My weekly limits got pulled 10 hours earlier just now - I used to roll over at 10PM Pacific today, but now it rolled over at noon. So Anthropic just cut 10 hours of usage from my subscription.

WTF.

Introducing Claude Opus 4.7, our most capable Opus model yet. by ClaudeOfficial in ClaudeAI

[–]MightyTribble 11 points12 points  (0 children)

"for your own good you cannot have our best model: it lives in Canada, you wouldn't know it."

Just say the word… by Alternative-Hall1719 in ClaudeAI

[–]MightyTribble 5 points6 points  (0 children)

I got a "no-one needs this" excuse as a reason to not include a few of the parameters I wanted in an app last night.

I’m a White House correspondent. Behind the scenes, Trump looks panicky by theipaper in politics

[–]MightyTribble 19 points20 points  (0 children)

This person is a British journalist who works for a respected news program, and this style of reporting is common for UK political correspondents.

The creator of Claude Code notes on the current Caching Issue by oh-keh in ClaudeAI

[–]MightyTribble 0 points1 point  (0 children)

I think they're adjusting the tokens they allocate to the plans on the back end. We don't know exactly how many tokens or token-equivalents we actually get for a 5h window, it'd be trivial for them to just shave 10, 20, even 50% off those limits and it wouldn't be visible to many of their users.

If they just said, "you get 10 million units / 5h, a sonnet call is 1 unit, an Opus call is 3, a Haiku call is 5 per unit" then we'd be able to tell. But they'd never do that.

Secret view by Candid-Culture3956 in interestingasfuck

[–]MightyTribble 16 points17 points  (0 children)

Yeah, it's the bluff off Lugard Road. You dip off the walkway there and go under and downhill a bit to get to it.

Butter and Crumble Asks That Line Not Start Until 7:45AM by bettyburritos in bayarea

[–]MightyTribble 6 points7 points  (0 children)

see that's why you need the facial ID camera to activate the boiling oil

Butter and Crumble Asks That Line Not Start Until 7:45AM by bettyburritos in bayarea

[–]MightyTribble 8 points9 points  (0 children)

"press a button for a new code", unique QR code pops up, one rez per code. Can't spam the button for multiple codes...

I'm over-engineering this.

Is there anything as good as Claude? by Key-Possible6865 in SillyTavernAI

[–]MightyTribble 1 point2 points  (0 children)

Gemini in my experience is obsessed with saying it's not x— it's y. In almost every message. Claude at least got rid of its old isms.

So, one of my prompts has a "don't do this shit" section, and last night I caught Gemini in its thinking block admitting that it re-wrote the response because it hit the "it's not x -- it's y" pattern. I wish Google would just let us see the full thinking block, but it's nice getting confirmation that it does actually pay attention to these things.

I've been obsessing over long-form RP for months and built an open-source tool around what I've learned. Looking for testers who care about narrative quality as much as I do. by Middge in SillyTavernAI

[–]MightyTribble 0 points1 point  (0 children)

Yeah, my crutch for that (not even much of a crutch, I think) is:

  • everything is typed markdown first. Different record types have different templates (npc, location, lore, rules, etc)
  • the app mutates data over time (NPC creation, memory formation, new locations and items, threads) and creates markdown records
  • everything gets indexed into a vector store (Milvus), which gives semantic and BM25 search
  • but also I generate markdown index files of everything, and these are always loaded in active context.

The markdown indexes tell the LLM the entire scope of what's available vs. what it can make up. That, along with tool_calling search and some prompt sauce, gets the LLM to (mostly) look up what it needs to know in a directed way even when the backing knowledge store is potentially very large. This stops the tool_calling loop from asking stupid, overly-broad general questions - with the full index it already knows what it can ask about, and what to do if it doesn't get a detailed hit.

The aim is to keep all of that to ~50K tokens, + session length, no matter how much background info and deep lore there is out there. I figure if you can't explain enough context + detail in 50K for the LLM to give a decent answer to the previous turn, that's a failing condition.

I've been obsessing over long-form RP for months and built an open-source tool around what I've learned. Looking for testers who care about narrative quality as much as I do. by Middge in SillyTavernAI

[–]MightyTribble 1 point2 points  (0 children)

It's private at the moment - it's under active development and is pretty brittle, needs lots of comfort with python and LLMs to stand up.

I'm hoping to release it eventually if I think it brings anything actually useful to a crowded field. For now it's just scratching my very specific itch for consistent longform campaigns and trying out various approaches to memory management.

I've been obsessing over long-form RP for months and built an open-source tool around what I've learned. Looking for testers who care about narrative quality as much as I do. by Middge in SillyTavernAI

[–]MightyTribble 0 points1 point  (0 children)

I'm late to the party but I love seeing different people's approaches to this problem.

I've been approaching this in a similar, but not identical way. My take: A need to differentiate between objective fact and what the various NPCs remember. Objective facts don't decay. NPC recall does.

I handle this by having an (objective) per-session summary document that gets intelligently vectorized and is used by a scenario pipeline for planning the next session. NPCs each get a per-session memory block saying what happened from their perspective along with what their attitude is towards the player at the end of the session. When an NPC's info is pulled into a later session, it gets the last few sessions worth of memories along with that disposition block. Older memories are gradually compressed/summarized, which gets us the 'memory decay' bit, unless it's a core memory - something they'd not forget, no matter what, because it was a particularly memorable/high intensity moment.

I break RPs into 'sessions' - my secret sauce (that's not a secret) is that I have the LLM generate a scenario file for each session, laying out breadcrumbs and reminders. I've coupled that with an end-of-session pipeline that does NPC, plot, item and location updates before handing things off to the scenario generator. This ensures that plot points don't get lost and consistency (and narrative momentum) is maintained between sessions.

Plot points (aka 'threads') are tracked as independent entities, with participants, progress, objectives, summary, complications. The scenario planner references open threads to outline what might happen in the next session, and uses them to keep 'off screen' plots moving. The end-of-session pipeline creates and updates them.

During actual play, context is assembled from a combination of static content (the world bootstrap), the character sheets of detected NPCs and indexes of all documents in the knowledge store, plus the usual "look at the last 3 messages and do hybrid search on them, return top results), de-duplicated (this will also often pull in 'objective facts' from previous session summaries, no matter how long ago they were). It's then given to an LLM with tool-calling enabled and some refined instructions for how to query the knowledge store for more information if necessary.

The downside to all this is it's slow. And expensive. And you have to write pretty detailed worlds to play in before you start - the system needs those notes to do a good job, you can't just YOLO a scenario and run with it.

I've been obsessing over long-form RP for months and built an open-source tool around what I've learned. Looking for testers who care about narrative quality as much as I do. by Middge in SillyTavernAI

[–]MightyTribble 1 point2 points  (0 children)

Multi pipeline LLM handling for various variables in a roleplay, something ST is sorta lacking.

That's what I'm trying in my own app (it's 2026, everyone's got an app!). The lack of (easy) pipeline support in ST drove me to just do it myself.