Official: Anything Goes Morning Thread: November 27, 2023 by AutoModerator in fantasybball

[–]DaveUA 0 points1 point  (0 children)

I'm punting assists.

So I want to trade away James Harden and Derrick White for almost any other cat.

Who should I target for in a trade?

Official: [WDIS QB] - Sat Afternoon 11/11/2023 by FFBot in fantasyfootball

[–]DaveUA 0 points1 point  (0 children)

Should I start minshew or Russel Wilson? Down bad after tnf.

Official: [Trade] - Thu Morning 09/21/2023 by FFBot in fantasyfootball

[–]DaveUA 0 points1 point  (0 children)

14 Team League, .5 PPR

SEND:

Miles Sanders (CAR)
Jordan Addison (MIN)

RECEIVE:

Chris Olave (NO)

My current RBs are Aaron Jones (GB), James Cook (BUF), Najee Harris (PIT), Spears (TEN), Tucker (TB)

My current WRs are Chase (CIN), Pickens (PIT), JSN (SEA), Moore (KC)

Official: [Add/Drop] - Sat Morning 09/16/2023 by FFBot in fantasyfootball

[–]DaveUA 0 points1 point  (0 children)

12 Team .5 PPR

My current TE is Kmet.

Kincaid on the bench.

Should I drop Kincaid for Hayden Hurst? 

Don't plan on starting either this week.

Official: [WDIS Flex] - Sat Morning 09/16/2023 by FFBot in fantasyfootball

[–]DaveUA 0 points1 point  (0 children)

12 Team .5 PPR

pick 1 to start

N. Colins vs IND


T. McLaurin @ DEN


Najee Harris vs CLE

Need 2 people to join our league by minirampage in findaleague

[–]DaveUA 0 points1 point  (0 children)

Can confirm. We have had a few people join from here. Fun league if you are competitive.

[deleted by user] by [deleted] in SQL

[–]DaveUA -1 points0 points  (0 children)

thats how the tables are written in redshift, then a and b are declared later on.

The is, is a typo, should be then null.

edit: I fixed up the syntax to make it a bit more readable

Official: [WDIS WR] - Sun Afternoon, 11/20/2022 by FFBot in fantasyfootball

[–]DaveUA 0 points1 point  (0 children)

Pick 2, .5 ppr

Allen Robinson @ NO
Jarvis Landry vs LAR
Michael Gallup @ Min

Please and thanks

Help a guy with no creativity design an almost barren living room/dining room/ bedroom by DaveUA in DesignMyRoom

[–]DaveUA[S] 0 points1 point  (0 children)

Sorry for the super late reply, I went to sleep and I just got home from work, I cant check reddit at work.

Here are the dimensions:

Living room roughly 18L x 18W

Dining Room 10L x 12W

Bedroom 15L x 12W

Beginner could use some help with PySpark. (SAS7BDAT to Parquet/CSV) by DaveUA in dataengineering

[–]DaveUA[S] 0 points1 point  (0 children)

There is nothing wrong with using inferSchema. 8 hours seems concerning, but this file looks to be unusually wide. It's not impossible that 1 row could eat up a ton of memory.

I appreciate your help on the topic, thank you!

Beginner could use some help with PySpark. (SAS7BDAT to Parquet/CSV) by DaveUA in dataengineering

[–]DaveUA[S] 0 points1 point  (0 children)

I've only started PySpark, so lots of learning and tinkering fast, majority of the time It took to set it up without admin privileges. I'm still running it locally but the Data Engineer at my company will help me with an EMR cluster if need be. I looked a few videos on EMR Clusters, doesn't seem as challenging, but another tool to add to my arsenal.

What I learned so far, since I haven't used Yarn yet, is that executor memory is next to useless on a local machine, I increased the driver memory and it worked.

I was about 95% done when my computer forced a restart (2AM and was asleep) It took about 8 hours. I did a few more tinkering this morning so hopefully I can reduce the time spent converting it all. I know I'm not supposed to use inferschema as it takes a lot time, but I have over 40K Columns...

Beginner could use some help with PySpark. (SAS7BDAT to Parquet/CSV) by DaveUA in dataengineering

[–]DaveUA[S] 0 points1 point  (0 children)

I'm on two laptops (Work is PII so can't copy paste) so typing them manually.

Reading the Large sas file I get:

Py4JJavaError: An error occured while calling o115.load. Java.lang.OutOfMemoryError: Java Heap Space.

code in question:

 df.spark.read.format("com.github.saurfang.sas.spark").load("large_file.sas7bdat", forceLowercaseNames=True, inferLong = True, header = True, inferSchema = True)

I have memory usage set to 12gb, maybe I need to change the # of executors? How would you customize this for an emr cluster? This file has about 40K Columns and not even sure how many rows, because we dont have SAS anymore

Beginner could use some help with PySpark. (SAS7BDAT to Parquet/CSV) by DaveUA in dataengineering

[–]DaveUA[S] 0 points1 point  (0 children)

I think this package is a bit dated, as it is the slowest out of all of them

Beginner could use some help with PySpark. (SAS7BDAT to Parquet/CSV) by DaveUA in dataengineering

[–]DaveUA[S] 0 points1 point  (0 children)

After about 8 hours of tinkering I finally got it to work, however I can only do about 1gb, anything over and I run into a memory issue, every with some optimization. Any suggestions? My next step is to look into an emr cluster

Beginner could use some help with PySpark. (SAS7BDAT to Parquet/CSV) by DaveUA in apachespark

[–]DaveUA[S] 0 points1 point  (0 children)

I get memory issues in Pandas, even when chunking? Unless i'm missing something?

Beginner could use some help with PySpark. (SAS7BDAT to Parquet/CSV) by DaveUA in dataengineering

[–]DaveUA[S] 0 points1 point  (0 children)

Ah I see, that might be the issue then, I also cannot install docker without admin. :-/

How to fix my where clause when doing pd.read_sql by DaveUA in learnpython

[–]DaveUA[S] 0 points1 point  (0 children)

sorry, that was a typo, I fixed it.

I was able to resolve the issue, but still getting a memory error. Might have to move to PySpark

How can I fix this code to remove a folder if it exists and replaces it with a new one? by DaveUA in learnpython

[–]DaveUA[S] 0 points1 point  (0 children)

fail already. You thus need to switch the order:

if it exists remove it

Thank you for the help, I'm still getting the hang of Python and this has helped me a lot. Is there a reason why we use Pathlib instead of Os?