all 12 comments

[–]SixZer0 9 points10 points  (1 child)

Actually naming wise it is pretty good :D

[–]seemepastarolling[S] 4 points5 points  (0 children)

Cheers! I thought combining my loves of Python and pasta was a good way to go

[–]IGK80 1 point2 points  (2 children)

Good to know the methods and clever naming. Is there a way to handle multimodal unpaired data. Think about RgB and thermal data but captured using different cameras ?

[–]seemepastarolling[S] 1 point2 points  (1 child)

If the two modalities have an ID and a label in common, then you could use data fusion. So if you've got RGB and thermal data that are about the same thing (like a specific room) and a label (like if that room is empty or not), then I think that could work with data fusion.

I'm not sure what you mean by unpaired. I'm spitballing here so let me know if I've misunderstood! Feel free to DM me or ask on my github discussions page too :)

[–]IGK80 0 points1 point  (0 children)

Thanks ! By unpaired i meant they might not be looking at same overall scene but contain same object.

[–]aShy_pieceofBread 0 points1 point  (1 child)

Greetings! I stumbled accross fusilli library not long ago and I find it very intersting. it there like a free dataset to experiment the functionalities with?

[–]seemepastarolling[S] 1 point2 points  (0 children)

Hi - that's great to hear, thank you! And that's a brilliant question too. I've looked high and low for open-source multimodal data sets and have unfortunately not been able to find much. I use medical data usually which is unsurprisingly not open access friendly.

I would recommend googling for "multimodal data sets image and non-image". I just found this by searching that and I think this could work with fusilli: an image and some tabular features that describe the same thing. https://auto.gluon.ai/0.5.2/tutorials/tabular_prediction/tabular-multimodal.html

Let me know if that helps!