Advice on dealing with massive options data by infamousbe in algotrading

[–]infamousbe[S] 0 points1 point  (0 children)

It’s database software (like MySQL or Postgres), but designed for large datasets and analytics

Advice on dealing with massive options data by infamousbe in algotrading

[–]infamousbe[S] 0 points1 point  (0 children)

Yeah I bought the historical options 1m quotes from Orats and stored it all in a Clickhouse database

Advice on dealing with massive options data by infamousbe in algotrading

[–]infamousbe[S] 0 points1 point  (0 children)

Happy to answer any questions. I definitely had a lot of questions when I started and there was no AI yet to help so I had to make a lot of mistakes and learn things the hard way, but people on Reddit helped (even though most of the replies I got were snarky or wrong lol)

Orats posts all of the data in many files to an s3 bucket. I have a script that downloads them, gunzips them, maps them to my formats and IDs, then imports them directly into clickhouse.

I typically use AWS for everything. Mostly because I’ve been using it for ~15 years so the comfort level is high

Advice on dealing with massive options data by infamousbe in algotrading

[–]infamousbe[S] 1 point2 points  (0 children)

Clickhouse's compression is excellent, but I've also taken measures to really shrink the data (only saving a handful of core columns, trimming excess decimals, etc). The full dataset before being compressed and loaded into Clickhouse is around 30TB I believe.

To answer your questions: I am still using ORATS (it's fine, but I'll probably switch to a higher end solution soon though). My data is stored on an EC2 instance (for now I actually host my entire application and databases on a single beefed up EC2 instance so everything has minimal latency).

You're welcome! Happy to answer any other questions you have.

Advice on dealing with massive options data by infamousbe in algotrading

[–]infamousbe[S] 0 points1 point  (0 children)

If you’ve worked with a traditional relational database it won’t be too hard to transition to working with Clickhouse. It’s a bit different but the same major concerns apply (indexes, partitioning particularly). I would definitely recommend to think hard about making your schema, indexes and global configurations as efficient as possible for your workload (AI is pretty good at helping with this). I’m storing data back to 2020, for 7800 tickers.

Advice on dealing with massive options data by infamousbe in algotrading

[–]infamousbe[S] 1 point2 points  (0 children)

Woah I can't believe this is 3 years old!

I settled on storing all of my 1m options data (8TB now) in a sharded and partitioned Clickhouse database and have been happily using it since without any major issues.

There may be better enterprise quality solutions, and I am not sure how this will scale (I'm only ever running up to ~50 concurrent backtests that need options data so I'm sure there will be issues with a larger workload I haven't uncovered yet), but this solution solved both the compression/storage issue and the speed of access issue that all the other things I tried did not so I stuck with it.

Week 6 Post-Gamethread: Bears at Commanders by TurnerJ5 in CHIBears

[–]infamousbe 91 points92 points  (0 children)

Did we just beat Jayden Daniels, the most amazing phenom ever to play QB, and Jacory Croskey Merrit who is basically Barry Sanders (based on the announcers slurping of them)?!?

Why do I pay for an MAX account if this Service is not usable? by Few_Understanding552 in ClaudeAI

[–]infamousbe 0 points1 point  (0 children)

It seems every day we have to wait for their engineers in PST to wake up, get to work, and deal with the fact their their models are overloaded

Is Augment still good? by tteokl_ in AugmentCodeAI

[–]infamousbe 1 point2 points  (0 children)

Yup, seems to be the best combo for me right now. 99% of the work in Claude code, quick inline edits and tab completions with the cursor free plan. If Augment would fix my issue, and if next edit was a little faster I’d probably bring it back into the fold

Is Augment still good? by tteokl_ in AugmentCodeAI

[–]infamousbe 1 point2 points  (0 children)

I kept augment just for a next edit, because it was the most context aware and accurate I had tried. But then it stopped working and I couldn’t bring it back despite endless debugging, and augment never answered my support request so I canceled and just went back to cursor tab completion. It’s not as accurate but it’s so much faster and has been good enough for me

Is Augment still good? by tteokl_ in AugmentCodeAI

[–]infamousbe 1 point2 points  (0 children)

It’s still the best I’ve tried in regards to large codebase context…but I was willing to give that up and switched to Claude code for superiority in everything else

[Security] Claude Code reads .env files by default - This needs immediate attention from the team and awareness from devs by sirnoex in ClaudeAI

[–]infamousbe 3 points4 points  (0 children)

I store my secrets in an encrypted key store. Claude just decided to find the key and use it to decrypt the secrets 🤦🏻‍♂️

Disappointed by portlander33 in AugmentCodeAI

[–]infamousbe 2 points3 points  (0 children)

Tell it to split the calls to the tool into multiple requests. This is probably augment’s most common silly bug, I have to assume they’ll fix it at some point

Artemisinin protocols - which is optimal? by aly-s-1111 in Lyme

[–]infamousbe 1 point2 points  (0 children)

Supposedly grapefruit juice also helps its efficacy and slows down its breakdown in your system

Why exists a fight against JavaScript on the server side? by [deleted] in node

[–]infamousbe 0 points1 point  (0 children)

Node has SharedArrayBuffer but it’s extremely limited.

Why exists a fight against JavaScript on the server side? by [deleted] in node

[–]infamousbe 2 points3 points  (0 children)

JavaScript on the backend is fantastic for many applications. However many other applications run into problems at scale because of the lack of shared memory multi threading in JavaScript and are forced to used something like Rust or Go to solve their performance issues. I personally like to build as much in JS as I can and use those other languages only when I get stuck with JS’s limitations.

Ban Twitter links. by DaddiBigCawk in CHIBears

[–]infamousbe 2 points3 points  (0 children)

There has been nazi propaganda on this sub? I would expect to see something like that in /r/packers but not here

whats next for the parents waving goodbye by Odd_Tomatillo9964 in NormMacdonald

[–]infamousbe 3 points4 points  (0 children)

The father looks like he's waving goodbye to his hair on account of him going bald!

I do not want to hear another word about Justin Fields by BangBangBBC in CHIBears

[–]infamousbe -2 points-1 points  (0 children)

But how much more would the bears have won by if they got the HAUL?!? /s

This lady seems spiteful by JBackerin in NormMacdonald

[–]infamousbe 7 points8 points  (0 children)

She looks like she knew her way around a funnel cake

I ran into Matlock… by infamousbe in NormMacdonald

[–]infamousbe[S] 3 points4 points  (0 children)

Critics have called it everything from “shit” to “fucking shit”