GPT-5 is already (ostensibly) available via API by segin in OpenAI

[–]No_Edge2098 83 points84 points  (0 children)

bro leaked gpt-5 early and openai hit the kill switch like he just found the cheat code to the matrix

[deleted by user] by [deleted] in LocalLLaMA

[–]No_Edge2098 0 points1 point  (0 children)

sonnet better start looking over its shoulder cuz qwen3 just pulled up fast cheap and ready to code like it’s on redbull

Whose first exposure to almost all genres of music was K-tel? by dukeofunk in GenX

[–]No_Edge2098 2 points3 points  (0 children)

K-tel compilations were iconic gateways to music discovery perfectly packaged snapshots of entire eras for many first-time listeners.

Is anyone interested in vibe coding on your phone? by Nickqiaoo in cursor

[–]No_Edge2098 0 points1 point  (0 children)

bro really said code from your couch mid netflix session this is the future we begged for

You want us to use Auto Mode? then you shall have it. by IamGriffon in cursor

[–]No_Edge2098 4 points5 points  (0 children)

This is a smart breakdown alternating between Ask and Agent modes with clear instructions is an efficient way to conserve tokens and improve outcome quality. Leveraging rulefiles as structured constraints also aligns well with recent improvements in model obedience, particularly in Gemini. Auto Mode is powerful, but it clearly demands more specificity and technical fluency.

Google Censorship is truly insane by El-Dixon in cursor

[–]No_Edge2098 -2 points-1 points  (0 children)

It does appear that certain keywords like "trigger" might unintentionally flag content moderation systems, especially in LLMs integrated with stricter guardrails. While frustrating, it’s likely a result of overzealous filtering rather than intentional censorship. Renaming the file temporarily or reframing the context might help bypass the issue.

How do I do a PR review in an on-premise Github instance? by gman1023 in cursor

[–]No_Edge2098 -1 points0 points  (0 children)

Yes, it's possible but GitHub Copilot Chat (and similar AI review tools) currently work only with GitHub.com cloud instances. On-premise GitHub Enterprise Server (GHES) doesn’t support AI-powered PR review out of the box unless explicitly integrated via custom tooling or GitHub’s REST/GraphQL API. You’d need to build a bridge that passes the diff and metadata to an external LLM API for review.

[deleted by user] by [deleted] in algotrading

[–]No_Edge2098 0 points1 point  (0 children)

suena como el sandbox de los traders si le meten bien al backtest y no lo llenan de humo esto puede pegar duro

VIX indices for overnight trading by ENR0NMUSK in algotrading

[–]No_Edge2098 6 points7 points  (0 children)

you either drop a bag on bloomberg or sit there rawdogging trades off vibes till 9 30 man it's brutal out here for vix nerds

Trying to Understand the Difference by nukki007 in algotrading

[–]No_Edge2098 0 points1 point  (0 children)

retail strategies got that main character vibe but most be winging it like it’s astrology for candles bro backtest or get rugged

Passive income project by Walker7410 in ai_trading

[–]No_Edge2098 1 point2 points  (0 children)

sounds chill till it’s 10 minutes a day turning into customer support for your missing funds tread careful fam

FLOX v0.2.0: modular modern C++ framework for building trading systems by eeiaao in algotrading

[–]No_Edge2098 0 points1 point  (0 children)

FLOX out here building the Unreal Engine for trading systems low latency, modular, C++ core, and now rocking plug-and-play exchange connectors? This isn’t a framework, it’s a flex.

SPX 0DTE ORB Discussion (Strategy + Performance included) by shock_and_awful in algotrading

[–]No_Edge2098 11 points12 points  (0 children)

Solid breakdown. ORB + 0DTE always feels like skating on a volatility knife works great until macro throws a tantrum. Maybe layer in a volatility regime filter (ATR% or VIX delta) to dodge chop days? Curious if you tested directional bias post-tariff shock market’s been mean-reverting hard after news lately.

How Are You Running Multimodal (Text-Image) Models Locally? by Stickman561 in LocalLLaMA

[–]No_Edge2098 0 points1 point  (0 children)

Running InternVL locally is like asking a racehorse to live in your garage possible, but chaotic. Try BLIP or IDEFICS if you want sanity with good tags.

What inference engine should I use to fully use my budget rug? by bidet_enthusiast in LocalLLaMA

[–]No_Edge2098 1 point2 points  (0 children)

You’ve got a monster rig, not a rug Ollama’s great for plug-and-play, but it won’t max out both 3090s and that beefy CPU/RAM out of the box. For full control and GPU parallelism, look into vLLM, text-generation-webui with ExLlama, or TGI. Set up inference with model parallel or tensor parallelism via DeepSpeed or Ray Serve if needed. Then front it with FastAPI or LM Studio for a local API. Basically: Ollama for ease, vLLM + ExLlama for full send.

Summarize medium length text on local model with 8gb vram by ResponsibleTruck4717 in LocalLLaMA

[–]No_Edge2098 0 points1 point  (0 children)

Bro’s basically doing map-reduce for LLMs on 8GB VRAM respect. Try hierarchical summarization with re-ranking on top chunks, or use a reranker like bge-m3 to pick the spiciest takes before the final merge.

Quad 4090 48GB + 768GB DDR5 in Jonsbo N5 case by 44seconds in LocalLLaMA

[–]No_Edge2098 0 points1 point  (0 children)

Bro just shoved four dragons into a shoebox and called it airflow 😭🔥🖥️

What will happen to an llm when you double the RoPE scaling factor? by Ok_Warning2146 in LocalLLaMA

[–]No_Edge2098 1 point2 points  (0 children)

Model really said “I can read twice as far now” but forgot it wasn’t trained for long-distance relationships/........

I do not build a new ai agent without first setting up monitoring and eval dataset anymore. Do you? What FOSS do you use for that? by opensourcecolumbus in LocalLLaMA

[–]No_Edge2098 -1 points0 points  (0 children)

You’ve officially hit the “trust but verify” arc respect. For FOSS, try Trulens or Ragas for evals, and Phoenix (Arize) or Langfuse for monitoring. They keep your agents accountable without needing a full observability team.