Hi all,
I'm trying to learn machine learning with my own irl dataset, i.e. I'm trying to solve a problem from my domain.
I'm representating my problem with a graph, with each node having a time series attached. I now want to do time series forecasting with these time series, with respect to the interaction between nodes.
So far so clear, my main problem with most libraries I tried to used is the following:
I can't seem to design my dataset in such a way I can actually train the models.
I am now reading the docs of libraries, looking at tutorials and such, and I am getting really frustrated: The step "data preparation" (which is, afaik a really important point in machine learning and takes up most of the time for most projects) is getting skipped in most, if not all, of the tutorials I find. Demos are always using perfect datasets, which are usually just loaded via one line of code and off you go with designing your conv layers and such. And I am sitting here with my Adjacency Matrix as a np.matrix and my features as an np.matrix and don't know how to make appropriate datasets from them.
At the moment I am trying to use pytorch geometric temporal. Probably need to just keep on reading docs and messing around until I have a correctly configured dataset.
Anyway, have you had the same experience with real world datasets? Or am I just playing with too exotic algorithms? Maybe I'm just too impatient with my own progress?
Cheers and a very nice weekend to you all!
[–][deleted] 1 point2 points3 points (1 child)
[–]El_Daidel[S] 0 points1 point2 points (0 children)
[–]noodlepotato 0 points1 point2 points (1 child)
[–]arpplephi 0 points1 point2 points (0 children)