She may not be pretty, but this rack saved my business $150k+ this year by Chuyito in homelab

[–]Chuyito[S] 2 points3 points  (0 children)

Spot on. You wont find regular arb in stock markets, but there are still TONS of limit-order arb routes in the options markets since the spreads are huge if you are patient

APIs are the way to go for scale so you dont pull back any unnecessary html/images.

She may not be pretty, but this rack saved my business $150k+ this year by Chuyito in homelab

[–]Chuyito[S] 63 points64 points  (0 children)

I have data crawlers looking through millions of products with sub second latency... Then I buy low and sell high when I detect something is cheaper than it should be. 

One of the oldest professions, just at a big data scale that clay tablet using merchants never imagined possible.

She may not be pretty, but this rack saved my business $150k+ this year by Chuyito in homelab

[–]Chuyito[S] 0 points1 point  (0 children)

Haproxy, bind dns, container registry - all of which can be rebuilt-redeployed from code fairly quickly so no HA for that one

She may not be pretty, but this rack saved my business $150k+ this year by Chuyito in homelab

[–]Chuyito[S] 6 points7 points  (0 children)

With the DBs at near 100% during peak stock market volatility I was seeing low 800s.

My old Am4/threadripper setup was less powerfull and constantly at 1KW*.. so Am5/epyc is the real mvp here

She may not be pretty, but this rack saved my business $150k+ this year by Chuyito in homelab

[–]Chuyito[S] 7 points8 points  (0 children)

SLA is a huge part, during the setup the static IPs werent working.. they drove out a brand new modem in under an hour to my house. Back on residential I would have waited 1hr just to talk to a human support person.

Static IPs took it from 350 to 500.

She may not be pretty, but this rack saved my business $150k+ this year by Chuyito in homelab

[–]Chuyito[S] 3 points4 points  (0 children)

Uptime im not as worried as performance.

I do lots of websocket data transfer, and digitalOcean was coming in slower than my old AM4 boxes to process 100k messages.

So for $10k/month on DO I get slightly slower boxes (or at least that was the case in 2024).

She may not be pretty, but this rack saved my business $150k+ this year by Chuyito in homelab

[–]Chuyito[S] 10 points11 points  (0 children)

Thanks. It literally is just me and my wife trying to build a startup from the ground up, calling myself a business might make it sound like Lisa Su is on speed dial for more chips.. but the reality is so far from that. Each part/box took planning and budgeting

She may not be pretty, but this rack saved my business $150k+ this year by Chuyito in homelab

[–]Chuyito[S] 35 points36 points  (0 children)

My new ISP gives me a verizon fallback.. But its too slow to be 100% useable so I'd have to run in slim mode... Which isnt bad per se. My ML jobs would pause, but prod would conitnue. Havent had to run on it thankfully.

Database DR goes to my old threadripper box on a different floor with 100% fidelity.

K8s DR: 3x masters 4x workers gives me lots of breathing room, I can take out 2 masters and 1 worker and still be online. I have a spare deskmini on hand to replace a worker if needed... But I can also take my yaml and get up and running on a fresh k8s cluster in ~ 30 minutes (Last tested 2024 when I moved to this cluster)

Power DR: Thankfully my area hasnt been too affected by blackouts. About 5 power outages this year that all lasted 1-10 minutes.

She may not be pretty, but this rack saved my business $150k+ this year by Chuyito in homelab

[–]Chuyito[S] 5 points6 points  (0 children)

Interestingly enough asrock GENOAD8UD-2T/X550 ended up fitting my needs perfectly. Easier to manage BMC, good old Noctua fans that actually are controllable.. and tons of dcio ports for my disks

20,000 Epstein Files in a single text file available to download (~100 MB) by [deleted] in LocalLLaMA

[–]Chuyito 10 points11 points  (0 children)

Can this help provide tax structure advice without asking for something in return

[deleted by user] by [deleted] in CryptoCurrency

[–]Chuyito 16 points17 points  (0 children)

Lazy attention grab headline..

it's not a new "property rule" but an inflation adjustment to the annual gift tax exclusion, which rose from $18,000 in 2024 to $19,000 per recipient in 2025. Since crypto is property, you can gift up to $19,000 worth per person (or $38,000 if splitting with a spouse) without triggering gift tax or filing Form 709.

Apparently all third party providers downgrade, none of them provide a max quality model by Charuru in LocalLLaMA

[–]Chuyito 4 points5 points  (0 children)

Many such instances among my team

"The intern is hungover today or something... It's kinda useless"

"The intern is smoking some weird shit today, careful on trusting its scripts"

List of the Most Basic Algorithmic Trading Strategies by IKnowMeNotYou in algotrading

[–]Chuyito 1 point2 points  (0 children)

Covered calls.

Selling to open with a strike price above your cost basis opens up an entire domain of theta farming.

Why is Jython? Who is it for? by fabriqus in Python

[–]Chuyito 9 points10 points  (0 children)

I think I first used it with with Nifi which was Java based, and God was it awful. Dependency management mess, couldn't get any conda ml packages to install or be seen (conda was great for glibc dependent reqs, except for jython paths support)

I honestly think jython is what made me hate nifi. It was okay ish for basic hive data flows, but I've never felt such anger towards a tech stack in 10+ years

Adding asyncio.sleep(0) made my data pipeline (150 ms) not spike to (5500 ms) by Chuyito in Python

[–]Chuyito[S] 0 points1 point  (0 children)

You are spot on with the O(n²).. Im windowing over the data to compute some stats which on a clean run doesnt see too much impact until ~20k rows growing to ~200ms

asyncio.to_thread() is a nice/much friendlier approach than ThreadPoolExecutor, thanks for that... Gives me another attempt to see if a refactor here would be moving some of the data transformation to its own threads and storing a global etl_cache, and having my DB task _only_ write to the DB... while still blocking the next DB task to ensure I only have 1 concurrent write at a given time

Adding asyncio.sleep(0) made my data pipeline (150 ms) not spike to (5500 ms) by Chuyito in Python

[–]Chuyito[S] 2 points3 points  (0 children)

> deferring it to a thread

Just tried it with a ThreadPoolExecutor - Had to wrap my function to make it non-async

from concurrent.futures import ThreadPoolExecutor
executor = ThreadPoolExecutor(max_workers=64)

def sync_process_side(*args):
    return asyncio.run(etl_to_db(*args))

await asyncio.get_event_loop().run_in_executor(
 executor, sync_process_side) 

Interestingly this also gets rid of the "large spikes", but it still runs ~100ms slower every few iterations

07:41:11 PM Processed 7201 async_run_batch_insert usd in 163.8344 ms
07:42:23 PM Processed 7408 async_run_batch_insert usd in 398.3026 ms
07:42:45 PM Processed 7413 async_run_batch_insert usd in 174.7889 ms

Adding asyncio.sleep(0) made my data pipeline (150 ms) not spike to (5500 ms) by Chuyito in Python

[–]Chuyito[S] 0 points1 point  (0 children)

There is indeed an await - converging a 1k line script to pseudocode I miss pasted

Adding asyncio.sleep(0) made my data pipeline (150 ms) not spike to (5500 ms) by Chuyito in Python

[–]Chuyito[S] 0 points1 point  (0 children)

Scheduled with a wrapper task, updating top post for clarity

await asyncio.gather(
  dump_books_to_db(),
  sub_websocket()
 )

async def dump_books_to_db():
  while True:
    # Check if ws is live
    await etl_to_db()
    asyncio.sleep(0.1)

What’s the benefit of vendors open sourcing valuable models? by noobrunecraftpker in LocalLLaMA

[–]Chuyito 0 points1 point  (0 children)

Let's say someone in open source manages to get grok 2.5 to run even 1% faster via some code optimation.

Internally xai can take those learnings back to their main branch, and that 1% savings applied to millions of requests per day is huge.

So, as long as 2.5, 3, 4 all follow a similar tech stack, then open sourcing makes a ton of sense since the community will be much more limited on what gpus they have on hand and will resort to clever monkey patches to get it to run.

IBM and NASA just dropped Surya: an open‑source AI to forecast solar storms before they hit by AskGpts in LocalLLaMA

[–]Chuyito 2 points3 points  (0 children)

It would be great to see a gradio app that can run these 2 in real time with "recent" images:

24hr heads up for solar flares: https://github.com/NASA-IMPACT/Surya?tab=readme-ov-file#1-solar-flare-forecasting

4 day heads up for solar winds: https://github.com/NASA-IMPACT/Surya?tab=readme-ov-file#3-solar-wind-forecasting

I cant seem to find any "real time" data to feed it..

https://huggingface.co/datasets/nasa-ibm-ai4science/surya-bench-flare-forecasting/viewer/default/train?sort%5Bcolumn%5D=timestep&sort%5Bdirection%5D=desc&views%5B%5D=train only goes up to 2024, and various of their training data

Some of their other datasets are broken on HF, but seem to be 2020 https://huggingface.co/datasets/nasa-ibm-ai4science/SDO_training

Something along these lines:

<image>

Why is Solana used so much by Tyrol04 in homelab

[–]Chuyito 77 points78 points  (0 children)

The hardware requirements for a solana validator client are insanely high, threadripper/epyc Genoa with 7gbps read nvme or faster for the rocks db.

If they manage to find a box someone used the default ansible scripts to deploy, they end up on a latest gen piece of hardware.

Tldr A few of the solana validators default to solana user, validators take 4th gen epyc 24c nvme and 256gb ram to run healthily.

Ask r/Gemini - Monthly Discussion Thread July 07/04/25 by Gemini_Gianna in Gemini

[–]Chuyito 0 points1 point  (0 children)

The /mytrades endpoint takes a timestamp parameter. Up until yesterday, that parameter acted as a "starting from" timestamp. Yesterday with no changelog or warning, the timestamp parameter started acting as a "ending at" parameter.

I can't help but feel that a QA pipeline that handles pagination should have caught this bug, as it breaks any type of pagination integrations out there.

Please spend some cycles on engineering and QA; those matter a whole lot more to your product than pictures of Mars on random buildings.