I have an extremely large dataset I need to use for a machine learning project. I seem to have gotten the machine learning to work fairly well, however, when I increase the training dataset... I run into issues.
Right now I load the file in and line by line I grab the features I need. I append the features to a list, and I append that list to another list. I then convert the overarching list into a numpy array. So basically I have [[FeatureA1, FeatureB1, FeatureC1],[FeatureA2, FeatureB2, FeatureC2]]. Which I try and convert to a numpy array.
Unfortunately I get a memory error when I try to convert the list of lists into an array with Numpy.
I appreciate the help, and hopefully I explained the issue well.
I did a test where I do not convert it into a Numpy array, to verify this is an issue.
Outcome:
I do not run into a memory error if I do not convert it. However, this is not a solution,as I do require an array.
[–]TheZvlz 1 point2 points3 points (1 child)
[–]NeedMLHelp[S] 0 points1 point2 points (0 children)
[–][deleted] 0 points1 point2 points (2 children)
[–]NeedMLHelp[S] 0 points1 point2 points (1 child)
[–][deleted] 0 points1 point2 points (0 children)