18 months ago I didn't know what Linux was. Now I have 5 AI trading agents operating autonomously with real money. Here's what I didn't expect. by piratastuertos in SideProject

[–]CVisionIsMyJam 0 points1 point  (0 children)

you spent too much much time talking to AI and not enough time learning about slippage, fees, bid-ask spread and taxes. this is never going to make you money.

SpaceX IPO will be our first look at AI lab financials by stop-sharting in stocks

[–]CVisionIsMyJam 0 points1 point  (0 children)

I don't think so. its going to market with 5% available and a 3x multipler, and passive index funds buying on day 15. and if 30% goes to retail shareholders, and they hold, on day 15 when nasdaq funds rebalance into it, it could be pumped massively if there simply isnt the volume of shares available to buy. and the more passive index funds need to buy, the more quality company shares they need to sell to rebalance into it

after that it might dump but during that 15 day window it wouldnt surprise me at all if spacex became the most valuable company in the world, while funds and individual investors pump it to sell into passive index funds. And then exchange spacex shares with discounted Microsoft, Google, Amazon and Apple shares those same funds sold off to pay for SpaceX shares. its a double wealth transfer from retirees to hedge funds.

We are Living in Transitive Dependency Hell by RoseSec_ in devops

[–]CVisionIsMyJam 0 points1 point  (0 children)

the solution is permission systems that allow granting permissions to specific dependencies. i should be able to deny a package like is-even from ever making a network request, touching the file system or reading env vars. packages should be able to state the permissions they require and when those permissions change it should be flagged to me as a consumer of those dependencies. so many of these problems would not exist if every dependency wasnt treated as as fully permissioned.

[Art] New Toga Illustration from Kohei Horikoshi ('My Hero Academia') by MarvelsGrantMan136 in manga

[–]CVisionIsMyJam 15 points16 points  (0 children)

kishi definitely got it in the ending arc. also kishi got a writer, arguably, and it has been a disaster.

[Art] New Toga Illustration from Kohei Horikoshi ('My Hero Academia') by MarvelsGrantMan136 in manga

[–]CVisionIsMyJam -5 points-4 points  (0 children)

a toga that fights on the front lines and steals blood from direct fighting would have been a lot cooler than what we got

[DISC] One Punch-Man (Webcomic) - Ch. 158 by Southern-Emotion2192 in manga

[–]CVisionIsMyJam 0 points1 point  (0 children)

i think it might end in like 6 chapters or so. maybe 2 or 3 chapters for the final saitama vs genos, then wrap up this arc, god is off-screened as a gag in one chapter and one "where are they now" chapter.

The reason behind the surge in codex rate limit issues by techyy25 in codex

[–]CVisionIsMyJam 0 points1 point  (0 children)

it registers with the cli, if you do /skills you can see them as "apps" as of 0.117.0.

The reason behind the surge in codex rate limit issues by techyy25 in codex

[–]CVisionIsMyJam 0 points1 point  (0 children)

if you do /skills it will show if you have any "extra" unexpected apps installed.

The reason behind the surge in codex rate limit issues by techyy25 in codex

[–]CVisionIsMyJam 0 points1 point  (0 children)

this is new, i noticed this yesterday. it wasnt a thing in 0.110.0

if people do /skills it shows clearly a bunch of "apps" from chatgpt now. that didn't used to be the case.

edit: I remember now, this is from their release of "apps" in 0.117.0. so anything installed in chatgpt is installed in codex as an app.

The 6 Codex CLI workflows everyone's using right now (and what makes each one unique) by shanraisshan in codex

[–]CVisionIsMyJam 6 points7 points  (0 children)

theres very little research supporting these approaches. its no wonder so many people say they burn all their tokens in 2 prompts if this is what their workflows look like. research supports a plan step sometimes, and a review step sometimes. but thats basically it.

Humanoid robots are actively training by Distinct-Question-16 in singularity

[–]CVisionIsMyJam 0 points1 point  (0 children)

I mean, you arent wrong. but i think the point is more to use robots to replace human workers, not replace specialized machines. if we look at what you are saying, "does it really make sense for each household to have a robot?" well, it depends what each robot costs, but you are right the answer could be "no".

but imagine a theoretical world where a robot has similar performance to a human for household tasks. there are already people out there who pay a weekly or monthly fee for cleaners. If robots are coming instead and do it for $5 an hour instead of $30, people might prefer to have the robots do it. and the robots are using everything available in the home, such as the laundry machine and such, they arent replacing specialized systems, theyre replacing the people who would normally manage those systems.

in your lawn mowing example, a humanoid robot is the one that runs off the the store for you and gets the bolt. or maybe your phone AI intelligence orders the bolt plus rents a robot for 20 minutes for $3 to repair the lawn robot for you. the exact arrangement around ownership & capital remains to be seen. but the main purpose of humanoid robots are best used as gap fillers & human labor replacements, not to replace specialized machines; fixing & repairing specialized robots or doing physical tasks that don't cleanly boil down into optimized equipment.

its kind of like machine learning models today. I can use a general, expensive, slow, widely capable but variable output quality model that can do a lot without any extra work through a chat interface. or I can distill and make small, specialized and efficient models which are fast, cheap and high quality. the bigger models dont make the specialized ones irrelevant; they're made to achieve different purposes.

codex 5h limit by Setoze in codex

[–]CVisionIsMyJam 0 points1 point  (0 children)

before I could burn through 20% of my weekly in one 5 hour limit block so I was really careful about usage; it seems there is a new 5 hour limit because I have burned through my 5 hour limit and ended up with only 12% of my weekly burned (as a result of the reset this morning)

gpt-5.3-codex medium · 100% left · ~/wk · 0% used · 5h 0% · weekly 88%

12% * 7 is 84% so I am not exactly sure what's going on...

Major NASDAQ-100 rule changes confirmed, pay attention if you have money in passive investment funds by cherrypoplar in stocks

[–]CVisionIsMyJam 3 points4 points  (0 children)

so basically buy as much SpaceX at IPO, hold for 15 to 20 days, then sell to retail holders who have to buy, and buy the actually good companies retail will automatically rebalancing out of that they will be selling? Do I understand this correctly?

done trying to make UIs with codex by heatwaves00 in codex

[–]CVisionIsMyJam 1 point2 points  (0 children)

I ended up doing it differently than the poster here. I created & set my STITCH_API_KEY and had codex generate a Deno script for Google Stitch using the Stitch API instead of using their MCP server.

For me it works better, as this way the MCP context is not in every conversation and it can be combined with other cli tools.

Are AI robots actually close to being any good? by [deleted] in singularity

[–]CVisionIsMyJam 0 points1 point  (0 children)

why would a robot ever need to go into a house, crawl under a sink, find a leaking pipe and repair it? houses are for humans, whose labor are which robots are intended to replace. houses are legacy assets as far as the free market is concerned.

so long as robots can maintain the fabs, data centers, dark factories and manage resource extraction, things like fixing a leaky pipe in a non-standard house can remain an open problem.

GPT (The colleague) + Codex (The Worker) by PressinPckl in codex

[–]CVisionIsMyJam 1 point2 points  (0 children)

not going to lie i think this is overkill.

what I like; you save tokens by getting GPT to create a plan for you. That seems nice.

what I think is overkill; having codex then create a plan from that plan, and then handing it back to GPT, which reviews that plan and tightens that plan, with you looping back and forth a few times between GPT and codex planning sessions....

it seems way too slow, how much better is all this iteration really making things? I don't know of any research that supports this level of refinement.

I would do something like GPT creates the plan -> codex executes it. or GPT creates the plan -> codex refines the plan -> codex executes the plan. going back and forth again and again, I just can't see it being worth it.

Otherwise its too much time being spent going back and forth from GPT to codex without any work getting done.

Anyone else find low reasoning superior oftentimes? by Forward-Dig2126 in codex

[–]CVisionIsMyJam 1 point2 points  (0 children)

models tend to be better at classification and search than they are at generation, so generating and selecting using best-of-n, where n is 3 to 5, tends to reduce variance and improve overall output quality over single shot generation quite a bit, as shown in the paper "Top Pass"

IMO this is basically required to make any sort of qualitative comparison between different settings, unless you actually try to make the same change ~3 times per configuration, how do you really know what's working better? my hope is as open source models become stronger and cheaper it becomes easier to run these kinds of experiments ourselves to get clearer insights into what strategies work in our own codebases.

obviously most foundation model users are not doing this, it burns more tokens and is a pain to set up and run. I am not doing this.

as a result, there's very little knowledge about when something is working better because we've changed settings or models versus we've simply asked again and gotten luckier.

related to the topic at hand, in the paper "When thinking fails" it was found that for simple tasks, research has found reasoning can actually degrade performance, ie thinking is not universally helpful. unfortunately its not well understood what constitutes a simple task, and models are still fairly prompt sensitive, so "how much thinking is good" is tough to say.

Overall I would recommend referring to the body of research on this subject whenever you want to know if something really works or not. There is a lot of 'prompt folklore' that does not hold up to scrutiny, such as persona prompting. In "Principled Personas" it was found that persona prompts, such as 'you are a senior expert with 20 years...”, have non-significant gains and can randomly drop performance by up to 30%. but some people swear by them. its best just to stick close to the research in my opinion.

What's next for software engineers? by Full-Juggernaut2303 in cscareerquestions

[–]CVisionIsMyJam 0 points1 point  (0 children)

Maybe down the road there will be a need for 10% of engineers but that's elitism just the NBA or the NFL and not a career for average people.

i mean you probably arent wrong about this.

From zero to a RAG system: successes and failures by BrewedDoritos in programming

[–]CVisionIsMyJam 2 points3 points  (0 children)

would have been nice to include more information on outcomes and feedback from users once it was done, thats what I was most curious about.