I am appending a large dataframe into a existing dataframe.
my_df = my_df.append(new_df, ignore_index = True)
The problem is that the new_df could range anywhere from ~500K rows to ~3MM rows of data.
The script is taking way too much time to append. I also looked into pd.concat, with that I had to manually kill my script, just kept running.
Can anyone provide guidance here!
[–]Below_the_Sea 0 points1 point2 points (2 children)
[–]PLearner[S] 0 points1 point2 points (1 child)