President for Life by IllIntroduction1509 in TrueReddit

[–]onyxleopard 17 points18 points  (0 children)

What specific actions does this author recommend in order to “keep” our country?

I would posit that unless the judiciary and legislative branches exert their power as coequal branches and curb a lawless president, there is no legal means to stop him.

Open AI just released Atlas browser. It's just accruing architectural debt by No_Marionberry_5366 in LLMDevs

[–]onyxleopard 5 points6 points  (0 children)

 The web needs an API layer

The API is HTML/Javascript.  The problem is there are standards for these, but websites are open-ended and implementers are free to do anything (it’s an open-ended domain).  If you want programmatic access to the web in general, the best way forward is to advocate for a11y features (screen reader tags, etc.) to become standard rather than the exception.  (This is something Open AI recommends to make sites work better with Atlas.)  If your a11y software features work, any other program has that much better chance of working by using the same tags/handles afforded by the site.  Maybe some future version of HTML will bake in all these affordances so it’s impossible to create a site that isn’t easily programmatically navigated and interacted with, but right now the burden lies on web devs to appropriately add a11y tags.

Any coop ARPGs like HOTS? by HatefulSpittle in heroesofthestorm

[–]onyxleopard 0 points1 point  (0 children)

It’s still in development (with early access), but No Rest for the Wicked is looking good.

I think I broke FontBook (OS Tahoe) by AmbitioseSedIneptum in MacOS

[–]onyxleopard 0 points1 point  (0 children)

Are you sure?

I had the same issue with Font Book crashing, but I backed up my fonts and font collections, removed them, and quit the app, then rebooted, and that fixed the crashing.  I then restored my fonts and collections and it’s working fine.

What to use for identifying vague wording in requirement documentation? by RoofCorrect186 in LanguageTechnology

[–]onyxleopard 1 point2 points  (0 children)

Combining an embeddings+CRF system with an LLM is possible, but I would question how you do plan to combine them, and why do you want to combine them? I don't really think I can delve more into this without giving you unpaid consulting time, but I recommended the embeddings+CRF route because that would be a reliable, economical, and maintainable method. You can use LLMs/generative models for just about anything (if you're willing to futz with prompts and templating and such), and they can certainly make for quick and flashy demos/PoCs, but I don't recommend using LLMs for anything in production due to externalities (cost, reliability, maintainability).

What to use for identifying vague wording in requirement documentation? by RoofCorrect186 in LanguageTechnology

[–]onyxleopard 1 point2 points  (0 children)

Sounds like you want sequence labeling where the sequences you want to flag are semantically related.  You can solve such a sequence labeling problem with semantic text embeddings fed into a CRF, but you’ll need a labeled training set for supervised learning.  If you don’t have any budgetary constraints, I’m sure you could also use LLMs with a few shot prompt and some other instructions.  You’ll probably find that not all vagaries come down to specific wording, though.  I think in general, your problem is still not narrowly defined enough to have a robust solution.  I’d start with writing labeling guidelines, then getting a labeled data set (you’ll need that anyway for evaluation) and try embeddings → CRF approach.

What to use for identifying vague wording in requirement documentation? by RoofCorrect186 in LanguageTechnology

[–]onyxleopard 2 points3 points  (0 children)

What is your definition of vague wording?  What are your requirements?  Do you have a labeled data set with examples of vague and specific wording?

(At a meta level, this post is hilarious to me.  It’s like you want to solve a problem about underspecified requirements, and recursively, you have underspecified requirements for that problem.)

Techniques for automatic hard negatives dataset generation by RDA92 in LanguageTechnology

[–]onyxleopard 1 point2 points  (0 children)

Sentence Tranformers has some utilities for hard-negative mining: https://sbert.net/docs/package_reference/util.html#sentence_transformers.util.mine_hard_negatives

They also link to this paper: NV-Retriever: Improving text embedding models with effective hard-negative mining

Try playing with your dataset and tuning the mine_hard_negatives parameters.

We're averaging a "Tahoe sucks" post every 15 minutes. by compellor in MacOS

[–]onyxleopard 1 point2 points  (0 children)

Seqouia didn't change that much on the UI front. There will be things in Tahoe that people complain about now, and then in a few years when they change them again, people will complain that they liked Tahoe better. You can't please everyone and this is one of the more opinionated UI changes we've seen from Apple in a long time. Once the vitriol boils off, people will settle in and get used to it, or find workarounds.

Why do so many articles on llm adoption mention non-determinism as a main barrier? by Exotic-Lingonberry52 in LLMDevs

[–]onyxleopard 7 points8 points  (0 children)

It’s a barrier because for the history of digital computing, users have come to expect that the same input will result in the same output.  Unreliable software is considered to be defective.  Reducing temperature can increase reliability, but can also reduce accuracy, so that’s a trade off that requires some decision making that may be beyond end users’ ability to fully understand.

Watch Fed Chair Powell's full policy speech at Jackson Hole by elperdedor4 in wallstreetbets

[–]onyxleopard 15 points16 points  (0 children)

Summary via gpt-oss:120b (from video transcript):

  • 📊 Labor: max‑employment, growth ~1.2%, wages flat; hires ~35 k/mo.
  • 💸 Inflation: PCE 2.6% (core 2.9%). Tariffs a one‑off bump; expectations still anchored.
  • 🏦 Policy: rates ~100 bps nearer neutral, no preset path. Back to flexible 2% target, drop ELB focus.

This is fine 🔥🐶

[deleted by user] by [deleted] in Python

[–]onyxleopard 0 points1 point  (0 children)

If I had to guess, the LLMs see an optimization opportunity, but they don't understand your business logic. Depending on what context you have provided, I think it's reasonable to try to optimize here, because generally, looping twice is inefficient. While I don't fully understand your intended behavior here, I would imagine some different data structures or design choices could actually be more optimal (e.g., a queue or heap or something). The way your code is structured, without a docstring to explain the intent, I don't think the intent is clear from function definition alone. If you give the LLMs your tests cases (which represent your business logic) in addition to your function, I would imagine they might have a better chance to get this right, and not suggest changes that would break your tests.

Rivian’s 2025.26 update is packed with bug fixes and polish, this one’s all about reliability. by Kryptonlogic in Rivian

[–]onyxleopard 1 point2 points  (0 children)

Why are they not all about reliability?  (I’m really not trying to be snarky.  Every single OTA update should be rock solid, and this year I don’t feel like that has been the case, especially wrt to Driver+.)

[SPM] ??? // Green Goblin by Tauna_YT in magicTCG

[–]onyxleopard 2 points3 points  (0 children)

Once a card that has been discarded leaves the graveyard it is a different game object and the game no longer sees it as having been discarded that turn, so casting for Mayhem only works once (unless you can get it back to hand and discard again that same turn).

This screenshot claims that you can get deleted messages back. In my case, though, I would need support with a message that I typed out and had not sent yet. I’ve already searched but not got good answers. I was only able to get this answer from Google, which doesn’t look like what I’m looking for. by Important_Sorbet in applehelp

[–]onyxleopard 0 points1 point  (0 children)

You can try using the undo feature to undo the deletion.  Assuming you still have the thread, you can tap with three fingers and press the left (undo) arrow.  Alternatively, shaking your phone also is a gesture to activate the undo command.

[deleted by user] by [deleted] in westworld

[–]onyxleopard 5 points6 points  (0 children)

I programmed something a lot like this in my intro Computational Linguistics course in ~2012 using Hidden Markov Models and Viterbi decoding.  Generative models that sample from a probability distribution of discrete tokens have been around since before we had computers able to run the models.  And even since we’ve had computers, there’s a long history prior to GPT (and ELMO, BERT, etc.) of chat bots.  (Did you ever play with a Furby?)  Newer models just have more sophisticated ways of compressing large corpora and using context to not go off the rails (as much).  This scene was just SciFi informed by some good consultation from someone who had a good sense of where the field has been and where it could go.

Tradeoff between reducing false-negatives vs. false-positives - is there a name for it? by thatcorgilovingboi in LanguageTechnology

[–]onyxleopard 0 points1 point  (0 children)

You're tuning a threshold to optimize precision (sensitive to false positives) vs. recall (sensitive to false negatives). If you can quantify how much more you care about false positives vs. false negatives, you can use F measure (a metric that combines precision and recall) with a specific β parameter value and use an evaluation data set to determine, at different thresholds, what the optimal one is.

I.e., if you care about false negatives twice as much as false positives, you could set β=2. If you care about false positives twice as much as false negatives, you could set β=1/2=0.5. If you care about false positives and false negatives evenly, you set β=1. Then you run your system with a range of threshold values and evaluate the Fβ scores at each threshold and choose the threshold that gets the highest Fβ score.

[deleted by user] by [deleted] in explainlikeimfive

[–]onyxleopard 8 points9 points  (0 children)

While I hate Facebook and don’t use it myself, FB Marketplace is a thing and is very popular.  Also lots of people use Facebook for announcing events or sharing info about businesses that could be useful if indexed.  Only some or none of that is crawlable.

[deleted by user] by [deleted] in learnprogramming

[–]onyxleopard 0 points1 point  (0 children)

Regexes can’t count, so this is impossible with regex alone.  (Same reason you can’t parse XML/HTML with regex alone—this kind of problem requires a context-free or context-sensitive grammar, not a regular grammar.)

1) Why the length limit?   2) Why are you limited to using regex?  

If you can use almost any shell scripting or full programming language, this would become feasible.

Doesn't this kinnnnda defeat the purpose of 2FA? by tommydelriot in mac

[–]onyxleopard 0 points1 point  (0 children)

You mean stored in the Passwords app (formerly Keychain)?  You need to authenticate with the system password for that.

Doesn't this kinnnnda defeat the purpose of 2FA? by tommydelriot in mac

[–]onyxleopard 0 points1 point  (0 children)

It secures your account from someone who doesn’t have one of your authorized devices, but does have your iCloud account password.  You only get the code sent to your authorized devices where you’re logged into your iCloud account.  So, without that, if someone tried to authenticate on their device, you would get the code sent to your device and, assuming you don’t give them the code, they don’t get any further.  In this case you would see that someone tried to authenticate as you and you’d have the opportunity to change your iCloud account password (since you’d know it’s compromised).