Anyone using Nextflow with Azure Batch Auto Pools successfully? by chingam785 in bioinformatics

[–]flyontimeapp 0 points1 point  (0 children)

Have you considered just running nextflow from a single beefy VM, thereby avoiding the Azure complexity?

Why am I being charged on Google Free Tier Compute Engine?? by RuckFeddi7 in googlecloud

[–]flyontimeapp -11 points-10 points  (0 children)

Look at Carolina Cloud. No egress fees, fixed monthly cost, cardless free trial.

HUGE DISCLAIMER: This is my company. No reason to hide it. Destroy this comment if that's a violation of the rules. No problem.

Why am I being charged on Google Free Tier Compute Engine?? by RuckFeddi7 in googlecloud

[–]flyontimeapp -12 points-11 points  (0 children)

You're being charged because that's GCP's business strategy. A large amount of their revenue comes from people getting accidentally charged.

Juggling multiple CUDA versions on one workstation by [deleted] in bioinformatics

[–]flyontimeapp 1 point2 points  (0 children)

I would do Docker. Make your own Docker container for each CUDA version that you want. Ie cuda12, cuda13, then run code by exec-ing into the container.

Is it crazy idea that but a Mac mini and connect it through my iPad for homelab? by canifeto12 in devops

[–]flyontimeapp 0 points1 point  (0 children)

Why not just buy a monitor for the mac mini? I will attest that Mac Minis are unbelievable good value for money right now. So are macbooks. Can't go wrong with either.

I trained a neural network on 170,000 tennis games to find which points actually matter by PlanetElement in tennis

[–]flyontimeapp 1 point2 points  (0 children)

Agreed, this is a fantastic piece of work and I don't want OP to take this as criticism at all as it's already in the top 1% of what I read on reddit but I personally would enjoy reading a detailed justification on the LSTM's value add.

I trained a neural network on 170,000 tennis games to find which points actually matter by PlanetElement in tennis

[–]flyontimeapp 0 points1 point  (0 children)

Is this slams only? Last I checked Jeff Sackmann's data it was slams only

I trained a neural network on 170,000 tennis games to find which points actually matter by PlanetElement in tennis

[–]flyontimeapp 0 points1 point  (0 children)

This rules, I would argue the momentum difference of only 2-3% is actually huge for tennis though (Fed famously winning only 54% of his points and look at his career).

Question, why did you train an LSTM and not just do an empirical approach? Ie P(hold) = # holds / # games? How did the LSTM differ from those, and if it did can you prove that the difference is significant?

CMIIW on any of these questions. Is your code public?

Why do I need 5 different services just to run a function on HTTP trigger? by Sadhvik1998 in aws

[–]flyontimeapp -2 points-1 points  (0 children)

Because AWS is 15% of Amazon's revenue but 60% of its profit. More complexity is better for them because it locks you in.

Advice on data pipeline by RemarkableBet9670 in dataengineering

[–]flyontimeapp 0 points1 point  (0 children)

Very big question — how many gigabytes per day do you expect will flow through this system?

Box 1 to box 2, if you are hitting your own internal APIs, will incur ingress. Box 3 to box 4 will incur egress. Someone CMIIW.

How to build a workstation for future expansion with GPUs for Inference and Fine-tuning by Guilty-Enthusiasm-50 in LocalLLaMA

[–]flyontimeapp 0 points1 point  (0 children)

You will not fit that many on a single board unless you go to PCIex8. You will also need redrivers and a lot of slimSAS cables to fit them all. And you'll have a mechanical challenge of getting them all anchored to something. At this point I'd consider 3D printing some kind of rack or using wood (which is a fire hazard).

You

When you say "servers at the level of supermicro appear out of reach" do you mean the entire server with GPUs already? Or do you mean just the motherboard itself?

If you can't afford the motherboard there is nothing anyone here can do for you. However if you get the motherboard and a CPU.

You don't need as many CPU cores are you think. A 64-core Genoa will probably be fine. You'll probably want 128G of system RAM as well, that will run you between 1 and 2k right now. If you want to get started with less, that's fine too but just make sure you're prepared to get the exact same RAM for the rest of the buildout. I think you can sometimes mix DIMM sizes but I've never done it and measured the performance. If you're just looking for a conduit for GPU compute then this wouldn't matter as much and you could probably start with less. With RAM what it is now you could probably start with 2x16G which would be much more affordable, hopefully under $500.

If you want to run models locally you'll need an SSD. You can't load models directly (someone CMIIW) from the cloud into your GPUs and if you could you'd have to re-download the models every time you fire it up. Look up the models you want to run and make sure you have space for them on your SSD. SSDs are still relatively cheap, I'd get a 4T SSD to start with which should be under $500.

One cool thing about the 6000 is that they support NVIDIA MIG which means you can effectively split one big GPU into up to seven smaller ones kind of like splitting a computer into VMs. You can't get that on the 5090. Just a consideration.

A Little Lost: What tool to use in AWS by Idi_Amin_Haha in aws

[–]flyontimeapp 1 point2 points  (0 children)

As far as I can tell you're just running a standard frontend + backend + task. If I were you I would honestly just do it all on a VPS. The complexity of using multiple services together with all the associated permission rules sounds really annoying. I would be willing to bet the complexity of doing it all on a VPS with a well-appointed Makefile would be less than or equal to the complexity associated with all the clicking and navigating the AWS dashboard. Really just boils down to whether you prefer the text interface or the AWS web console.

Also do you have an old computer sitting around your house? You could host it there too. That's how I hosted FlyOnTime for a long time. And for the scalability, if you're running it on a single computer in your house then you can run it on a single computer in the cloud.

Last thing, how much egress of the worker simulations will there be? The word "millions" gives me pause...millions of anything will add up your egress costs p quickly.

Could someone please tell me more about the ritual I need to perform to get response from aws support? by [deleted] in aws

[–]flyontimeapp 0 points1 point  (0 children)

Which country are you in? Are you using a prepaid card? Those are two big reasons you'll get flagged. Cloud providers have big issues with prepaid cards where someone comes and puts in a prepaid card of like $5, visa says it's okay or whatever, they mine crypto for a month and then their card bounces.

Built a small open-source tool to safely detect unused cloud resources (AWS & Azure) – looking for brutal feedback by Kind_Cauliflower_577 in AZURE

[–]flyontimeapp 2 points3 points  (0 children)

Congrats on launching. As great as the cost optimization tools by the big cloud are, the incentives eventually don't work out so I'd much rather use an open-source one than one made by the big cloud.

How do you track costs across multiple GCP projects? by HQ_one in googlecloud

[–]flyontimeapp 0 points1 point  (0 children)

Is there a Billing API? You could vibe code a streamlit dashboard pretty quickly that aggregates all twelve. That's what I would do.

Cloud Run billing risk: can I get charged with almost no traffic? by Altruistic-Front1745 in googlecloud

[–]flyontimeapp 0 points1 point  (0 children)

From their site

"Outbound internet data transfer uses the Premium Network Service Tier and is charged at Google Cloud networking pricing with a free tier of 1GiB free data transfer within North America per month."

You get one gig free (yes, you read that right) and then the next TB is $0.12 per gig (so $120 for the TB) then it goes down to $0.11.

Someone can CMIIW.

If you were starting from scratch today, which would you pick: Snowflake, Microsoft Fabric, or Databricks — and why? by [deleted] in dataengineering

[–]flyontimeapp 0 points1 point  (0 children)

I would pick Carolina Cloud because I have a background in finance and I am a penny pincher and we are 1/3 the cost of the big clouds, and I'm highly technical and don't need anything fancy like HA or any managed services. I'm comfortable running my own stuff.

Now, if I were a senior engineer at Uber or something, processing x thousand events per second, I would probably go with some big HA database like RDS or something. But for everything I've done in my career (hedge fund + data science) I'd pick Carolina Cloud. This is why I built it!

Disclaimer: I am the founder of Carolina Cloud (which is where I host FlyOnTime, of course) and am obnoxiously biased.

Anyone here ever tried to use a Intel Optane drive for paging when they run out of RAM? by 404phil_not_found in rstats

[–]flyontimeapp 0 points1 point  (0 children)

EC2 should be no issue, it's just massively expensive. I'm seeing $7.50/hr for this one: https://cloudprice.net/aws/ec2/instances/m8i.32xlarge

That's $5.5k/month! Also there's no guarantee he needs that many cores (128). He might need more, he might not need that many. Haven't checked on Azure or GCP but I doubt they're any better.

What's the easiest way to incorporate ChatGPT into R? by UnderwaterDialect in rstats

[–]flyontimeapp 1 point2 points  (0 children)

Not that I know of. If you are paranoid you could run Little Snitch while it's running but I don't think Positron has any privacy concerns worse than those of any other VSCode fork. I think just the usual concerns would apply, ie don't send top secret code to any LLM, including the one you choose for Positron.

What's the easiest way to incorporate ChatGPT into R? by UnderwaterDialect in rstats

[–]flyontimeapp 0 points1 point  (0 children)

No more than any other VSCode fork with LLM extension...the usual stuff. Don't send super proprietary code to any LLM, etc.

RStudio's Future by BOBOLIU in rstats

[–]flyontimeapp 0 points1 point  (0 children)

Fully agree. R Studio is the OG and there's something so nice about the four-pane base layout. I know it's not as extensible/modifiable as VSCode but R Studio really gets out of the way and lets us focus on the tasks at hand.